首页 > 最新文献

Journal of the Acoustical Society of Korea最新文献

英文 中文
Masked cross self-attentive encoding based speaker embedding for speaker verification 基于掩模交叉自关注编码的说话人嵌入方法
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.497
Soonshin Seo, Ji-Hwan Kim
Constructing speaker embeddings in speaker verification is an important issue. In general, a self-attention mechanism has been applied for speaker embedding encoding. Previous studies focused on training the self-attention in a high-level layer, such as the last pooling layer. In this case, the effect of low-level layers is not well represented in the speaker embedding encoding. In this study, we propose Masked Cross Self-Attentive Encoding (MCSAE) using ResNet. It focuses on training the features of both high-level and low-level layers. Based on multi-layer aggregation, the output features of each residual layer are used for the MCSAE. In the MCSAE, the interdependence of each input features is trained by cross self-attention module. A random masking regularization module is also applied to prevent overfitting problem. The MCSAE enhances the weight of frames representing the speaker information. Then, the output features are concatenated and encoded in the speaker embedding. Therefore, a more informative speaker embedding is encoded by using the MCSAE. The experimental results showed an equal error rate of 2.63 % using the VoxCeleb1 evaluation dataset. It improved performance compared with the previous self-attentive encoding and state-of-the-art methods.
说话人嵌入的构建是说话人验证中的一个重要问题。一般来说,自注意机制被用于说话人嵌入编码。以往的研究主要集中在高层次的自我注意训练,如最后一层池化。在这种情况下,低层次的效果不能很好地体现在说话人嵌入编码中。在这项研究中,我们提出了基于ResNet的掩蔽交叉自关注编码(MCSAE)。它侧重于训练高级层和低级层的特征。基于多层聚合,将各残差层的输出特征用于MCSAE。在MCSAE中,通过交叉自注意模块训练各输入特征的相互依赖性。随机屏蔽正则化模块也用于防止过拟合问题。MCSAE增强了代表说话人信息的帧的权重。然后,将输出特征串接并编码到说话人嵌入中。因此,使用MCSAE对更有信息量的说话人嵌入进行编码。实验结果表明,使用VoxCeleb1评价数据集,错误率为2.63%。与以前的自关注编码和最先进的方法相比,它提高了性能。
{"title":"Masked cross self-attentive encoding based speaker embedding for speaker verification","authors":"Soonshin Seo, Ji-Hwan Kim","doi":"10.7776/ASK.2020.39.5.497","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.497","url":null,"abstract":"Constructing speaker embeddings in speaker verification is an important issue. In general, a self-attention mechanism has been applied for speaker embedding encoding. Previous studies focused on training the self-attention in a high-level layer, such as the last pooling layer. In this case, the effect of low-level layers is not well represented in the speaker embedding encoding. In this study, we propose Masked Cross Self-Attentive Encoding (MCSAE) using ResNet. It focuses on training the features of both high-level and low-level layers. Based on multi-layer aggregation, the output features of each residual layer are used for the MCSAE. In the MCSAE, the interdependence of each input features is trained by cross self-attention module. A random masking regularization module is also applied to prevent overfitting problem. The MCSAE enhances the weight of frames representing the speaker information. Then, the output features are concatenated and encoded in the speaker embedding. Therefore, a more informative speaker embedding is encoded by using the MCSAE. The experimental results showed an equal error rate of 2.63 % using the VoxCeleb1 evaluation dataset. It improved performance compared with the previous self-attentive encoding and state-of-the-art methods.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"497-504"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47000827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Triplet loss based domain adversarial training for robust wake-up word detection in noisy environments 噪声环境下基于三联体损失的域对抗训练鲁棒唤醒词检测
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.468
Hyungjun Lim, Myunghun Jung, Hoirin Kim
A good acoustic word embedding that can well express the characteristics of word plays an important role in wake-up word detection (WWD). However, the representation ability of acoustic word embedding may be weakened due to various types of environmental noise occurred in the place where WWD works, causing performance degradation. In this paper, we proposed triplet loss based Domain Adversarial Training (tDAT) mitigating environmental factors that can affect acoustic word embedding. Through experiments in noisy environments, we verified that the proposed method effectively improves the conventional DAT approach, and checked its scalability by combining with other method proposed for robust WWD.
一个好的声学词嵌入能够很好地表达词的特征,在唤醒词检测(WWD)中起着重要作用。但是,由于WWD工作地点存在各种类型的环境噪声,可能会削弱声词嵌入的表示能力,导致性能下降。在本文中,我们提出了基于三联体损失的领域对抗训练(tDAT),以减轻可能影响声词嵌入的环境因素。通过在噪声环境下的实验,验证了该方法对传统的数据分解方法的有效改进,并结合其他鲁棒WWD方法验证了该方法的可扩展性。
{"title":"Triplet loss based domain adversarial training for robust wake-up word detection in noisy environments","authors":"Hyungjun Lim, Myunghun Jung, Hoirin Kim","doi":"10.7776/ASK.2020.39.5.468","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.468","url":null,"abstract":"A good acoustic word embedding that can well express the characteristics of word plays an important role in wake-up word detection (WWD). However, the representation ability of acoustic word embedding may be weakened due to various types of environmental noise occurred in the place where WWD works, causing performance degradation. In this paper, we proposed triplet loss based Domain Adversarial Training (tDAT) mitigating environmental factors that can affect acoustic word embedding. Through experiments in noisy environments, we verified that the proposed method effectively improves the conventional DAT approach, and checked its scalability by combining with other method proposed for robust WWD.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"468-475"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47175818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Performance comparison of wake-up-word detection on mobile devices using various convolutional neural networks 不同卷积神经网络在移动设备上唤醒词检测的性能比较
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.454
Sangho Lee
Artificial intelligence assistants that provide speech recognition operate through cloud-based voice recognition with high accuracy. In cloud-based speech recognition, Wake-Up-Word (WUW) detection plays an important role in activating devices on standby. In this paper, we compare the performance of Convolutional Neural Network (CNN)-based WUW detection models for mobile devices by using Google's speech commands dataset, using the spectrogram and mel-frequency cepstral coefficient features as inputs. The CNN models used in this paper are multi-layer perceptron, general convolutional neural network, VGG16, VGG19, ResNet50, ResNet101, ResNet152, MobileNet. We also propose network that reduces the model size to 1/25 while maintaining the performance of MobileNet is also proposed.
提供语音识别的人工智能助手通过基于云的高精度语音识别进行操作。在基于云的语音识别中,唤醒词(WUW)检测在激活待机设备方面发挥着重要作用。在本文中,我们使用谷歌的语音命令数据集,使用频谱图和mel频率倒谱系数特征作为输入,比较了基于卷积神经网络(CNN)的移动设备WUW检测模型的性能。本文使用的CNN模型有多层感知器、通用卷积神经网络、VGG16、VGG19、ResNet50、ResNet101、ResNet152、MobileNet。我们还提出了在保持MobileNet性能的同时将模型大小减小到1/25的网络。
{"title":"Performance comparison of wake-up-word detection on mobile devices using various convolutional neural networks","authors":"Sangho Lee","doi":"10.7776/ASK.2020.39.5.454","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.454","url":null,"abstract":"Artificial intelligence assistants that provide speech recognition operate through cloud-based voice recognition with high accuracy. In cloud-based speech recognition, Wake-Up-Word (WUW) detection plays an important role in activating devices on standby. In this paper, we compare the performance of Convolutional Neural Network (CNN)-based WUW detection models for mobile devices by using Google's speech commands dataset, using the spectrogram and mel-frequency cepstral coefficient features as inputs. The CNN models used in this paper are multi-layer perceptron, general convolutional neural network, VGG16, VGG19, ResNet50, ResNet101, ResNet152, MobileNet. We also propose network that reduces the model size to 1/25 while maintaining the performance of MobileNet is also proposed.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"454-460"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46347014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
α-feature map scaling for raw waveform speaker verification 原始波形扬声器验证的α-特征映射缩放
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.441
Jee-weon Jung, Hye-jin Shim, Ju-ho Kim, Ha-jin Yu
{"title":"α-feature map scaling for raw waveform speaker verification","authors":"Jee-weon Jung, Hye-jin Shim, Ju-ho Kim, Ha-jin Yu","doi":"10.7776/ASK.2020.39.5.441","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.441","url":null,"abstract":"","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"441-446"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46488187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Acoustic model training using self-attention for low-resource speech recognition 基于自注意的低资源语音识别声学模型训练
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.483
Hosung Kim
This paper proposes acoustic model training using self-attention for low-resource speech recognition. In low-resource speech recognition, it is difficult for acoustic model to distinguish certain phones. For example, plosive /d/ and /t/, plosive /g/ and /k/ and affricate /z/ and /ch/. In acoustic model training, the self-attention generates attention weights from the deep neural network model. In this study, these weights handle the similar pronunciation error for low-resource speech recognition. When the proposed method was applied to Time Delay Neural Network-Output gate Projected Gated Recurrent Unit (TNDD-OPGRU)-based acoustic model, the proposed model showed a 5.98 % word error rate. It shows absolute improvement of 0.74 % compared with TDNN-OPGRU model.
本文提出了一种基于自注意的声学模型训练方法,用于低资源语音识别。在低资源语音识别中,声学模型难以区分特定的电话。例如,爆破音/d/和/t/,爆破音/g/和/k/,不灭音/z/和/ch/。在声学模型训练中,自注意从深度神经网络模型中生成注意权值。在本研究中,这些权重处理了低资源语音识别的类似发音错误。将该方法应用于基于时延神经网络输出门投影门控循环单元(TNDD-OPGRU)的声学模型,该模型的单词错误率为5.98%。与TDNN-OPGRU模型相比,该模型的绝对改进率为0.74%。
{"title":"Acoustic model training using self-attention for low-resource speech recognition","authors":"Hosung Kim","doi":"10.7776/ASK.2020.39.5.483","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.483","url":null,"abstract":"This paper proposes acoustic model training using self-attention for low-resource speech recognition. In low-resource speech recognition, it is difficult for acoustic model to distinguish certain phones. For example, plosive /d/ and /t/, plosive /g/ and /k/ and affricate /z/ and /ch/. In acoustic model training, the self-attention generates attention weights from the deep neural network model. In this study, these weights handle the similar pronunciation error for low-resource speech recognition. When the proposed method was applied to Time Delay Neural Network-Output gate Projected Gated Recurrent Unit (TNDD-OPGRU)-based acoustic model, the proposed model showed a 5.98 % word error rate. It shows absolute improvement of 0.74 % compared with TDNN-OPGRU model.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"483-489"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42189624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Absolute sound level algorithm for contents platform 内容平台的绝对声级算法
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.424
Du-Heon Gyeon
This paper describes an algorithm that calculates Absolute Sound Level (ASL) for contents platform. ASL is a single volume representing individual sound sources and is a concept designed to integrate and utilize the sound level units in digital sound source and physical domain from a speaker in practical areas. For this concept to be used in content platforms and others, it is necessary to automatically derive the ASL without having to go through a hearing of mastering engineers. The key parameters of which a person recognizes the representative sound level of an individual single sound source are the areas of “frequency, maximum energy, energy variation coefficient, and perceived energy distribution,” and the ASL was calculated through the normalizing of the weights.
本文介绍了一种计算内容平台绝对声级(ASL)的算法。ASL是代表单个声源的单个音量,是在实际领域中整合和利用扬声器在数字声源和物理领域中的声级单位的概念。为了在内容平台和其他平台中使用这个概念,有必要自动派生ASL,而不必经过母版工程师的听证会。一个人识别单个声源的代表性声级的关键参数是“频率、最大能量、能量变化系数和感知能量分布”区域,通过权重的归一化计算出ASL。
{"title":"Absolute sound level algorithm for contents platform","authors":"Du-Heon Gyeon","doi":"10.7776/ASK.2020.39.5.424","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.424","url":null,"abstract":"This paper describes an algorithm that calculates Absolute Sound Level (ASL) for contents platform. ASL is a single volume representing individual sound sources and is a concept designed to integrate and utilize the sound level units in digital sound source and physical domain from a speaker in practical areas. For this concept to be used in content platforms and others, it is necessary to automatically derive the ASL without having to go through a hearing of mastering engineers. The key parameters of which a person recognizes the representative sound level of an individual single sound source are the areas of “frequency, maximum energy, energy variation coefficient, and perceived energy distribution,” and the ASL was calculated through the normalizing of the weights.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"424-434"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48592183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved speech enhancement of multi-channel Wiener filter using adjustment of principal subspace vector 基于主子空间矢量调整的改进多通道维纳滤波器语音增强
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.490
Gibak Kim
We present a method to improve the performance of the multi-channel Wiener filter in noisy environment. To build subspace-based multi-channel Wiener filter, in the case of single target source, the target speech component can be effectively estimated in the principal subspace of speech correlation matrix. The speech correlation matrix can be estimated by subtracting noise correlation matrix from signal correlation matrix based on the assumption that the cross-correlation between speech and interfering noise is negligible compared with speech correlation. However, this assumption is not valid in the presence of strong interfering noise and significant error can be induced in the principal subspace accordingly. In this paper, we propose to adjust the principal subspace vector using speech presence probability and the steering vector for the desired speech source. The multi-channel speech presence probability is derived in the principal subspace and applied to adjust the principal subspace vector. Simulation results show that the proposed method improves the performance of multi-channel Wiener filter in noisy environment.
提出了一种在噪声环境下提高多通道维纳滤波器性能的方法。为了建立基于子空间的多通道维纳滤波器,在单个目标源的情况下,可以在语音相关矩阵的主子空间中有效地估计目标语音分量。基于语音与干扰噪声之间的互相关与语音相关相比可忽略不计的假设,可以通过从信号相关矩阵中减去噪声相关矩阵来估计语音相关矩阵。然而,在存在强干扰噪声的情况下,这一假设是无效的,因此可能在主子空间中引起显著误差。在本文中,我们建议使用语音存在概率和所需语音源的引导向量来调整主子空间向量。在主子空间中导出多通道语音存在概率,并将其用于调整主子空间向量。仿真结果表明,该方法提高了多通道维纳滤波器在噪声环境中的性能。
{"title":"Improved speech enhancement of multi-channel Wiener filter using adjustment of principal subspace vector","authors":"Gibak Kim","doi":"10.7776/ASK.2020.39.5.490","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.490","url":null,"abstract":"We present a method to improve the performance of the multi-channel Wiener filter in noisy environment. To build subspace-based multi-channel Wiener filter, in the case of single target source, the target speech component can be effectively estimated in the principal subspace of speech correlation matrix. The speech correlation matrix can be estimated by subtracting noise correlation matrix from signal correlation matrix based on the assumption that the cross-correlation between speech and interfering noise is negligible compared with speech correlation. However, this assumption is not valid in the presence of strong interfering noise and significant error can be induced in the principal subspace accordingly. In this paper, we propose to adjust the principal subspace vector using speech presence probability and the steering vector for the desired speech source. The multi-channel speech presence probability is derived in the principal subspace and applied to adjust the principal subspace vector. Simulation results show that the proposed method improves the performance of multi-channel Wiener filter in noisy environment.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"490-496"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48121114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
I-vector similarity based speech segmentation for interested speaker to speaker diarization system 基于I向量相似度的感兴趣说话人对说话人二元化系统语音分割
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.461
Ara Bae, Ki‑mu Yoon, Jaehong Jung, Bokyung Chung, Wooil Kim
In noisy and multi-speaker environments, the performance of speech recognition is unavoidably lower than in a clean environment. To improve speech recognition, in this paper, the signal of the speaker of interest is extracted from the mixed speech signals with multiple speakers. The VoiceFilter model is used to effectively separate overlapped speech signals. In this work, clustering by Probabilistic Linear Discriminant Analysis (PLDA) similarity score was employed to detect the speech signal of the interested speaker, which is used as the reference speaker to VoiceFilter-based separation. Therefore, by utilizing the speaker feature extracted from the detected speech by the proposed clustering method, this paper propose a speaker diarization system using only the mixed speech without an explicit reference speaker signal. We use phone-dataset consisting of two speakers to evaluate the performance of the speaker diarization system. Source to Distortion Ratio (SDR) of the operator (Rx) speech and customer speech (Tx) are 5.22 dB and –5.22 dB respectively before separation, and the results of the proposed separation system show 11.26 dB and 8.53 dB respectively.
在嘈杂和多说话人的环境中,语音识别的性能不可避免地会低于干净的环境。为了提高语音识别能力,本文从多个说话人的混合语音信号中提取目标说话人的信号。voiceffilter模型用于有效分离重叠的语音信号。在这项工作中,采用概率线性判别分析(PLDA)相似性评分聚类来检测感兴趣的说话人的语音信号,并将其作为基于voicefilter的分离的参考说话人。因此,本文利用所提出的聚类方法从检测语音中提取的说话人特征,提出了一种只使用混合语音而不使用明确参考说话人信号的说话人拨号系统。我们使用由两个说话人组成的电话数据集来评估说话人拨号系统的性能。分离前,运营商语音(Rx)和客户语音(Tx)的源失真比(SDR)分别为5.22 dB和-5.22 dB,分离系统的结果分别为11.26 dB和8.53 dB。
{"title":"I-vector similarity based speech segmentation for interested speaker to speaker diarization system","authors":"Ara Bae, Ki‑mu Yoon, Jaehong Jung, Bokyung Chung, Wooil Kim","doi":"10.7776/ASK.2020.39.5.461","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.461","url":null,"abstract":"In noisy and multi-speaker environments, the performance of speech recognition is unavoidably lower than in a clean environment. To improve speech recognition, in this paper, the signal of the speaker of interest is extracted from the mixed speech signals with multiple speakers. The VoiceFilter model is used to effectively separate overlapped speech signals. In this work, clustering by Probabilistic Linear Discriminant Analysis (PLDA) similarity score was employed to detect the speech signal of the interested speaker, which is used as the reference speaker to VoiceFilter-based separation. Therefore, by utilizing the speaker feature extracted from the detected speech by the proposed clustering method, this paper propose a speaker diarization system using only the mixed speech without an explicit reference speaker signal. We use phone-dataset consisting of two speakers to evaluate the performance of the speaker diarization system. Source to Distortion Ratio (SDR) of the operator (Rx) speech and customer speech (Tx) are 5.22 dB and –5.22 dB respectively before separation, and the results of the proposed separation system show 11.26 dB and 8.53 dB respectively.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"461-467"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42501548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Development of portable single-beam acoustic tweezers for biomedical applications 生物医学用便携式单束声镊的研制
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.435
Junsu Lee, Yeon-Seong Park, Miji Kim, Changhan Yoon
Single-beam acoustic tweezers that are capable of manipulating micron-size particles in a non-contact manner have been used in many biological and biomedical applications. Current single-beam acoustic tweezer systems developed for in vitro experiments consist of a function generator and a power amplifier, thus the system is bulky and expensive. This configuration would not be suitable for in vivo and clinical applications. Thus, in this paper, we present a portable single-beam acoustic tweezer system and its performances of trapping and manipulating micron-size objects. The developed system consists of an Field Programmable Gate Array (FPGA) chip and two pulsers, and parameters such as center frequency and pulse duration were controlled by a Personal Computer (PC) via a USB (Universal Serial Bus) interface in real-time. It was shown that the system was capable of generating the transmitting pulse up to 20 MHz, and producing sufficient intensity to trap microparticles and cells. The performance of the system was evaluated by trapping and manipulating 40 μm and 90 μm in diameter polystyrene particles.
单束声镊能够以非接触方式操纵微米大小的颗粒,已用于许多生物和生物医学应用。目前用于体外实验的单波束声镊系统由一个函数发生器和一个功率放大器组成,因此系统体积大,价格昂贵。这种结构不适合体内和临床应用。因此,在本文中,我们提出了一种便携式单波束声镊系统及其捕获和操纵微米尺寸物体的性能。所开发的系统由一个现场可编程门阵列(FPGA)芯片和两个脉冲器组成,中心频率和脉冲持续时间等参数由PC机通过USB(通用串行总线)接口实时控制。结果表明,该系统能够产生高达20 MHz的发射脉冲,并产生足够的强度来捕获微粒和细胞。通过捕获和操纵直径为40 μm和90 μm的聚苯乙烯颗粒,对该系统的性能进行了评价。
{"title":"Development of portable single-beam acoustic tweezers for biomedical applications","authors":"Junsu Lee, Yeon-Seong Park, Miji Kim, Changhan Yoon","doi":"10.7776/ASK.2020.39.5.435","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.435","url":null,"abstract":"Single-beam acoustic tweezers that are capable of manipulating micron-size particles in a non-contact manner have been used in many biological and biomedical applications. Current single-beam acoustic tweezer systems developed for in vitro experiments consist of a function generator and a power amplifier, thus the system is bulky and expensive. This configuration would not be suitable for in vivo and clinical applications. Thus, in this paper, we present a portable single-beam acoustic tweezer system and its performances of trapping and manipulating micron-size objects. The developed system consists of an Field Programmable Gate Array (FPGA) chip and two pulsers, and parameters such as center frequency and pulse duration were controlled by a Personal Computer (PC) via a USB (Universal Serial Bus) interface in real-time. It was shown that the system was capable of generating the transmitting pulse up to 20 MHz, and producing sufficient intensity to trap microparticles and cells. The performance of the system was evaluated by trapping and manipulating 40 μm and 90 μm in diameter polystyrene particles.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"435-440"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44370393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal design of impeller in fan motor unit of cordless vacuum cleaner for improving flow performance and reducing aerodynamic noise 无绳吸尘器风机电机单元叶轮优化设计,提高流动性能,降低气动噪声
IF 0.4 Q4 ACOUSTICS Pub Date : 2020-09-01 DOI: 10.7776/ASK.2020.39.5.379
Kunwoo Kim, Seo-Yoon Ryu, C. Cheong, Seongjin Seo, Cheolmin Jang
In this study, the flow and noise performances of high-speed fan motor unit for cordless vacuum cleaner is improved by optimizing the impeller which drives the suction air through flow passage of the cordless vacuum cleaner. Firstly, the unsteady incompressible Reynolds averaged Navier-Stokes (RANS) equations are solved to investigate the flow through the fan motor unit using the computational fluid dynamics techniques. Based on flow field results, the Ffowcs-Williams and Hawkings (FW-H) integral equation is used to predict flow noise radiated from the impeller. Predicted results are compared to the measured ones, which confirms the validity of the numerical method used. It is found that the strong vortex is formed around the mid-chord region of the main blades where the blade curvature change rapidly. Given that vortex acts as a loss for flow and a noise source for noise, impeller blade is redesigned to suppress the identified vortex. The response surface method using two factors is employed to determine the optimum inlet and outlet sweep angles for maximum flow rate and minimum noise. Further analysis of finally selected design confirms the improved flow and noise performance.
本研究通过优化驱动吸入空气通过无绳吸尘器流道的叶轮,改善了无绳吸尘器高速风机电机单元的流动性能和噪声性能。首先,利用计算流体力学技术求解非定常不可压缩Reynolds平均Navier-Stokes (RANS)方程,研究风机电机单元的流动。基于流场结果,采用Ffowcs-Williams and hawkins (FW-H)积分方程对叶轮辐射的流动噪声进行了预测。将预测结果与实测结果进行了比较,验证了数值方法的有效性。结果表明,在叶片曲率变化较快的主叶片中弦区周围形成强涡。考虑到涡旋是流动的损失和噪声的噪声源,对叶轮叶片进行了重新设计,以抑制已识别的涡旋。采用双因素响应面法确定了最大流量和最小噪声条件下的最佳进出口掠角。对最终选择的设计进行进一步分析,证实了改进后的流动性能和噪声性能。
{"title":"Optimal design of impeller in fan motor unit of cordless vacuum cleaner for improving flow performance and reducing aerodynamic noise","authors":"Kunwoo Kim, Seo-Yoon Ryu, C. Cheong, Seongjin Seo, Cheolmin Jang","doi":"10.7776/ASK.2020.39.5.379","DOIUrl":"https://doi.org/10.7776/ASK.2020.39.5.379","url":null,"abstract":"In this study, the flow and noise performances of high-speed fan motor unit for cordless vacuum cleaner is improved by optimizing the impeller which drives the suction air through flow passage of the cordless vacuum cleaner. Firstly, the unsteady incompressible Reynolds averaged Navier-Stokes (RANS) equations are solved to investigate the flow through the fan motor unit using the computational fluid dynamics techniques. Based on flow field results, the Ffowcs-Williams and Hawkings (FW-H) integral equation is used to predict flow noise radiated from the impeller. Predicted results are compared to the measured ones, which confirms the validity of the numerical method used. It is found that the strong vortex is formed around the mid-chord region of the main blades where the blade curvature change rapidly. Given that vortex acts as a loss for flow and a noise source for noise, impeller blade is redesigned to suppress the identified vortex. The response surface method using two factors is employed to determine the optimum inlet and outlet sweep angles for maximum flow rate and minimum noise. Further analysis of finally selected design confirms the improved flow and noise performance.","PeriodicalId":42689,"journal":{"name":"Journal of the Acoustical Society of Korea","volume":"39 1","pages":"379-389"},"PeriodicalIF":0.4,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45624374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Journal of the Acoustical Society of Korea
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1