Pub Date : 2023-05-19DOI: https://dl.acm.org/doi/10.1145/3594538
Li Yang, Xi Li, Zhuoru Ma, Lu Li, Neal Xiong, Jianfeng Ma
Gait authentication as a technique that can continuously provide identity recognition on mobile devices for security has been investigated by academics in the community for decades. However, most of the existing work achieves insufficient generalization to complex real-world environments due to the complexity of the noisy real-world gait data. To address this limitation, we propose an intelligent Implicit Real-time Gait Authentication (IRGA) system based on Deep Neural Networks (DNNs) for enhancing the adaptability of gait authentication in practice. In the proposed system, the gait data (whether with complex interference signals) will first be processed sequentially by the imperceptible collection module and data preprocessing module for improving data quality. In order to illustrate and verify the suitability of our proposal, we provide analysis of the impact of individual gait changes on data feature distribution. Finally, a fusion neural network composed of a Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) is designed to perform feature extraction and user authentication. We evaluate the proposed IRGA system in heterogeneous complex scenarios and present start-of-the-art comparisons on three datasets. Extensive experiments demonstrate that the IRGA system achieves improved performance simultaneously in several different metrics.
{"title":"IRGA: An Intelligent Implicit Real-time Gait Authentication System in Heterogeneous Complex Scenarios","authors":"Li Yang, Xi Li, Zhuoru Ma, Lu Li, Neal Xiong, Jianfeng Ma","doi":"https://dl.acm.org/doi/10.1145/3594538","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3594538","url":null,"abstract":"<p>Gait authentication as a technique that can continuously provide identity recognition on mobile devices for security has been investigated by academics in the community for decades. However, most of the existing work achieves insufficient generalization to complex real-world environments due to the complexity of the noisy real-world gait data. To address this limitation, we propose an intelligent Implicit Real-time Gait Authentication (IRGA) system based on Deep Neural Networks (DNNs) for enhancing the adaptability of gait authentication in practice. In the proposed system, the gait data (whether with complex interference signals) will first be processed sequentially by the imperceptible collection module and data preprocessing module for improving data quality. In order to illustrate and verify the suitability of our proposal, we provide analysis of the impact of individual gait changes on data feature distribution. Finally, a fusion neural network composed of a Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) is designed to perform feature extraction and user authentication. We evaluate the proposed IRGA system in heterogeneous complex scenarios and present start-of-the-art comparisons on three datasets. Extensive experiments demonstrate that the IRGA system achieves improved performance simultaneously in several different metrics.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"68 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-19DOI: https://dl.acm.org/doi/10.1145/3593585
Luca Muscariello, Michele Papalini, Olivier Roques, Mauro Sardara, Arthur Tran Van
In this article, we consider security aspects of online meeting applications based on protocols such as WebRTC that leverage the Information-centric Networking (ICN) architecture to make the system fundamentally more scalable. If the scalability properties provided by ICN have been proved in recent literature, the security challenges and implications for real-time applications have not been reviewed. We show that this class of applications can benefit from strong security and scalability jointly without any major tradeoff and with significant performance improvements over traditional WebRTC systems. To achieve this goal, some modifications to the current ICN architecture must be implemented in the way integrity and authentication are verified. Extensive performance analysis of the architecture based on the open source implementation of Hybrid-ICN proves that real-time applications can greatly benefit from this novel network architecture in terms of strong security and scalable communications.
{"title":"Securing Scalable Real-time Multiparty Communications with Hybrid Information-centric Networking","authors":"Luca Muscariello, Michele Papalini, Olivier Roques, Mauro Sardara, Arthur Tran Van","doi":"https://dl.acm.org/doi/10.1145/3593585","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3593585","url":null,"abstract":"<p>In this article, we consider security aspects of online meeting applications based on protocols such as WebRTC that leverage the Information-centric Networking (ICN) architecture to make the system fundamentally more scalable. If the scalability properties provided by ICN have been proved in recent literature, the security challenges and implications for real-time applications have not been reviewed. We show that this class of applications can benefit from strong security and scalability jointly without any major tradeoff and with significant performance improvements over traditional WebRTC systems. To achieve this goal, some modifications to the current ICN architecture must be implemented in the way integrity and authentication are verified. Extensive performance analysis of the architecture based on the open source implementation of Hybrid-ICN proves that real-time applications can greatly benefit from this novel network architecture in terms of strong security and scalable communications.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"100 1 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-19DOI: https://dl.acm.org/doi/10.1145/3586010
Pedro Victor Borges, Chantal Taconet, Sophie Chabridon, Denis Conan, Everton Cavalcante, Thais Batista
In the last years, Internet of Things (IoT) platforms have been designed to provide IoT applications with various services such as device discovery, context management, and data filtering. The lack of standardization has led each IoT platform to propose its own abstractions, APIs, and data models. As a consequence, programming interactions between an IoT consuming application and an IoT platform is time-consuming, is error prone, and depends on the developers’ level of knowledge about the IoT platform. To address these issues, this article introduces IoTvar, a middleware library deployed on the IoT consumer application that manages all its interactions with IoT platforms. IoTvar relies on declaring variables automatically mapped to sensors whose values are transparently updated with sensor observations through proxies on the client side. This article presents the IoTvar architecture and shows how it has been integrated into the FIWARE, OM2M, and muDEBS platforms. We also report the results of experiments performed to evaluate IoTvar, showing that it reduces the effort required to declare and manage IoT variables and has no considerable impact on CPU, memory, and energy consumption.
{"title":"Taming Internet of Things Application Development with the IoTvar Middleware","authors":"Pedro Victor Borges, Chantal Taconet, Sophie Chabridon, Denis Conan, Everton Cavalcante, Thais Batista","doi":"https://dl.acm.org/doi/10.1145/3586010","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3586010","url":null,"abstract":"<p>In the last years, Internet of Things (IoT) platforms have been designed to provide IoT applications with various services such as device discovery, context management, and data filtering. The lack of standardization has led each IoT platform to propose its own abstractions, APIs, and data models. As a consequence, programming interactions between an IoT consuming application and an IoT platform is time-consuming, is error prone, and depends on the developers’ level of knowledge about the IoT platform. To address these issues, this article introduces <i>IoTvar</i>, a middleware library deployed on the IoT consumer application that manages all its interactions with IoT platforms. IoTvar relies on declaring variables automatically mapped to sensors whose values are transparently updated with sensor observations through proxies on the client side. This article presents the IoTvar architecture and shows how it has been integrated into the FIWARE, OM2M, and <span>muDEBS</span> platforms. We also report the results of experiments performed to evaluate IoTvar, showing that it reduces the effort required to declare and manage IoT variables and has no considerable impact on CPU, memory, and energy consumption.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"99 3 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-19DOI: https://dl.acm.org/doi/10.1145/3583687
Jing Chen, Wenjun Jiang, Jie Wu, Kenli Li, Keqin Li
The Point Of Interest (POI) sequence recommendation is the key task in itinerary and travel route planning. Existing works usually consider the temporal and spatial factors in travel planning. However, the external environment, such as the weather, is usually overlooked. In fact, the weather is an important factor because it can affect a user’s check-in behaviors. Furthermore, most of the existing research is based on a static environment for POI sequence recommendation. While the external environment (e.g., the weather) may change during travel, it is difficult for existing works to adjust the POI sequence in time. What’s more, people usually prefer the attractive routes when traveling. To address these issues, we first conduct comprehensive data analysis on two real-world check-in datasets to study the effects of weather and time, as well as the features of the POI sequence. Based on this, we propose a model of Dynamic Personalized POI Sequence Recommendation with fine-grained contexts (DPSR for short). It extracts user interest and POI popularity with fine-grained contexts and captures the attractiveness of the POI sequence. Next, we apply the Monte Carlo Tree Search model (MCTS for short) to simulate the process of recommending POI sequence in the dynamic environment, i.e., the weather and time change after visiting a POI. What’s more, we consider different speeds to reflect the fact that people may take different transportation to transfer between POIs. To validate the efficacy of DPSR, we conduct extensive experiments. The results show that our model can improve the accuracy of the recommendation significantly. Furthermore, it can better meet user preferences and enhance experiences.
{"title":"Dynamic Personalized POI Sequence Recommendation with Fine-Grained Contexts","authors":"Jing Chen, Wenjun Jiang, Jie Wu, Kenli Li, Keqin Li","doi":"https://dl.acm.org/doi/10.1145/3583687","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3583687","url":null,"abstract":"<p>The Point Of Interest (POI) sequence recommendation is the key task in itinerary and travel route planning. Existing works usually consider the temporal and spatial factors in travel planning. However, the external environment, such as the weather, is usually overlooked. In fact, the weather is an important factor because it can affect a user’s check-in behaviors. Furthermore, most of the existing research is based on a static environment for POI sequence recommendation. While the external environment (e.g., the weather) may change during travel, it is difficult for existing works to adjust the POI sequence in time. What’s more, people usually prefer the attractive routes when traveling. To address these issues, we first conduct comprehensive data analysis on two real-world check-in datasets to study the effects of weather and time, as well as the features of the POI sequence. Based on this, we propose a model of Dynamic Personalized POI Sequence Recommendation with fine-grained contexts (<i>DPSR</i> for short). It extracts user interest and POI popularity with fine-grained contexts and captures the attractiveness of the POI sequence. Next, we apply the Monte Carlo Tree Search model (MCTS for short) to simulate the process of recommending POI sequence in the dynamic environment, i.e., the weather and time change after visiting a POI. What’s more, we consider different speeds to reflect the fact that people may take different transportation to transfer between POIs. To validate the efficacy of <i>DPSR</i>, we conduct extensive experiments. The results show that our model can improve the accuracy of the recommendation significantly. Furthermore, it can better meet user preferences and enhance experiences.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"23 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
When making decisions, individuals often express their preferences linguistically. The computing with words methodology is a key basis for supporting linguistic decision making, and the words in that methodology may mean different things to different individuals. Thus, in this article, we propose a continual personalized individual semantics learning model to support a consensus-reaching process in large-scale linguistic group decision making. Specifically, we first derive personalized numerical scales from the data of linguistic preference relations. We then perform a clustering ensemble method to divide large-scale group and conduct consensus management. Finally, we present a case study of intelligent route optimization in shared mobility to illustrate the usability of our proposed model. We also demonstrate its effectiveness and feasibility through a comparative analysis.
{"title":"Personalized Individual Semantics Learning to Support a Large-Scale Linguistic Consensus Process","authors":"Yucheng Dong, Qin Ran, Xiangrui Chao, Congcong Li, Shui Yu","doi":"https://dl.acm.org/doi/10.1145/3533432","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3533432","url":null,"abstract":"<p>When making decisions, individuals often express their preferences linguistically. The computing with words methodology is a key basis for supporting linguistic decision making, and the words in that methodology may mean different things to different individuals. Thus, in this article, we propose a continual personalized individual semantics learning model to support a consensus-reaching process in large-scale linguistic group decision making. Specifically, we first derive personalized numerical scales from the data of linguistic preference relations. We then perform a clustering ensemble method to divide large-scale group and conduct consensus management. Finally, we present a case study of intelligent route optimization in shared mobility to illustrate the usability of our proposed model. We also demonstrate its effectiveness and feasibility through a comparative analysis.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"52 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-19DOI: https://dl.acm.org/doi/10.1145/3589342
Arvind Kumar Gangwar, Sandeep Kumar
Software Defect Prediction (SDP) is crucial towards software quality assurance in software engineering. SDP analyzes the software metrics data for timely prediction of defect prone software modules. Prediction process is automated by constructing defect prediction classification models using machine learning techniques. These models are trained using metrics data from historical projects of similar types. Based on the learned experience, models are used to predict defect prone modules in currently tested software. These models perform well if the concept is stationary in a dynamic software development environment. But their performance degrades unexpectedly in the presence of change in concept (Concept Drift). Therefore, concept drift (CD) detection is an important activity for improving the overall accuracy of the prediction model. Previous studies on SDP have shown that CD may occur in software defect data and the used defect prediction model may require to be updated to deal with CD. This phenomenon of handling the CD is known as CD adaptation. It is observed that still efforts need to be done in this direction in the SDP domain. In this article, we have proposed a pair of paired learners (PoPL) approach for handling CD in SDP. We combined the drift detection capabilities of two independent paired learners and used the paired learner (PL) with the best performance in recent time for next prediction. We experimented on various publicly available software defect datasets garnered from public data repositories. Experimentation results showed that our proposed approach performed better than the existing similar works and the base PL model based on various performance measures.
{"title":"Concept Drift in Software Defect Prediction: A Method for Detecting and Handling the Drift","authors":"Arvind Kumar Gangwar, Sandeep Kumar","doi":"https://dl.acm.org/doi/10.1145/3589342","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3589342","url":null,"abstract":"<p>Software Defect Prediction (SDP) is crucial towards software quality assurance in software engineering. SDP analyzes the software metrics data for timely prediction of defect prone software modules. Prediction process is automated by constructing defect prediction classification models using machine learning techniques. These models are trained using metrics data from historical projects of similar types. Based on the learned experience, models are used to predict defect prone modules in currently tested software. These models perform well if the concept is stationary in a dynamic software development environment. But their performance degrades unexpectedly in the presence of change in concept (Concept Drift). Therefore, concept drift (CD) detection is an important activity for improving the overall accuracy of the prediction model. Previous studies on SDP have shown that CD may occur in software defect data and the used defect prediction model may require to be updated to deal with CD. This phenomenon of handling the CD is known as CD adaptation. It is observed that still efforts need to be done in this direction in the SDP domain. In this article, we have proposed a pair of paired learners (PoPL) approach for handling CD in SDP. We combined the drift detection capabilities of two independent paired learners and used the paired learner (PL) with the best performance in recent time for next prediction. We experimented on various publicly available software defect datasets garnered from public data repositories. Experimentation results showed that our proposed approach performed better than the existing similar works and the base PL model based on various performance measures.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"281 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-19DOI: https://dl.acm.org/doi/10.1145/3589765
Hucheng Wang, Zhi Wang, Lei Zhang, Xiaonan Luo, Xinheng Wang
Fusion positioning technology requires stable and effective positioning data, but this is often challenging to achieve in complex Non-Line-of-Sight (NLoS) environments. This paper proposes a fusion positioning method that can achieve stable and no hop points by adjusting parameters and predicting trends, even with a one-sided lack of fusion data. The method combines acoustic signal and Inertial Measurement Unit (IMU) data, exploiting their respective advantages. The fusion is achieved using the Kalman filter and Bayesian parameter estimation is performed for tuning IMU parameters and predicting motion trends. The proposed method overcomes the problem of fusion failure caused by long-term unilateral data loss in traditional fusion positioning. The positioning trajectory and error distribution analysis show that the proposed method performs optimally in severe NLoS experiments.
{"title":"A Highly Stable Fusion Positioning System of Smartphone under NLoS Acoustic Indoor Environment","authors":"Hucheng Wang, Zhi Wang, Lei Zhang, Xiaonan Luo, Xinheng Wang","doi":"https://dl.acm.org/doi/10.1145/3589765","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3589765","url":null,"abstract":"<p>Fusion positioning technology requires stable and effective positioning data, but this is often challenging to achieve in complex <b>Non-Line-of-Sight (NLoS)</b> environments. This paper proposes a fusion positioning method that can achieve stable and no hop points by adjusting parameters and predicting trends, even with a one-sided lack of fusion data. The method combines acoustic signal and <b>Inertial Measurement Unit (IMU)</b> data, exploiting their respective advantages. The fusion is achieved using the Kalman filter and Bayesian parameter estimation is performed for tuning IMU parameters and predicting motion trends. The proposed method overcomes the problem of fusion failure caused by long-term unilateral data loss in traditional fusion positioning. The positioning trajectory and error distribution analysis show that the proposed method performs optimally in severe NLoS experiments.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"23 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-18DOI: https://dl.acm.org/doi/10.1145/3543854
Yi-Bing Lin, Yuan-Fu Liao, Sin-Horng Chen, Shaw-Hwa Hwang, Yih-Ru Wang
The voice-based Internet of Multimedia Things (IoMT) is the combination of IoT interfaces and protocols with associated voice-related information, which enables advanced applications based on human-to-device interactions. An example is Automatic Speech Recognition (ASR) for live captioning and voice translation. Three major issues of ASR for IoMT are IoT development cost, speech recognition accuracy, and execution time complexity. For the first issue, most non-voice IoT applications are upgraded with the ASR feature through hard coding, which are error prone. For the second issue, recognition accuracy must be improved for ASR. For the third issue, many multimedia IoT services are real-time applications and, therefore, the ASR delay must be short.
This article elaborates on the above issues based on an IoT platform called VoiceTalk. We built the largest Taiwanese spoken corpus to train VoiceTalk ASR (VT-ASR) and show how the VT-ASR mechanism can be transparently integrated with existing IoT applications. We consider two performance measures for VoiceTalk: speech recognition accuracy and VT-ASR delay. For the acoustic tests of PAL-Labs, VT-ASR's accuracy is 96.47%, while Google's accuracy is 94.28%. We are the first to develop an analytic model to investigate the probability that the VT-ASR delay for the first speaker is complete before the second speaker starts talking. From the measurements and analytic modeling, we show that the VT-ASR delay is short enough to result in a very good user experience. Our solution has won several important government and commercial TV contracts in Taiwan. VT-ASR has demonstrated better Taiwanese Mandarin speech recognition accuracy than famous commercial products (including Google and Iflytek) in Formosa Speech Recognition Challenge 2018 (FSR-2018) and was the best among all participating ASR systems for Taiwanese recognition accuracy in FSR-2020.
基于语音的多媒体物联网(IoMT)是物联网接口和协议与相关语音相关信息的结合,它使基于人与设备交互的高级应用成为可能。一个例子是用于实时字幕和语音翻译的自动语音识别(ASR)。物联网ASR的三个主要问题是物联网开发成本、语音识别准确性和执行时间复杂性。对于第一个问题,大多数非语音物联网应用都是通过硬编码升级ASR功能的,这很容易出错。对于第二个问题,必须提高ASR的识别精度。对于第三个问题,许多多媒体物联网服务是实时应用,因此ASR延迟必须短。本文基于一个名为VoiceTalk的物联网平台详细阐述了上述问题。我们建立了最大的台湾口语语料库来训练VoiceTalk ASR (VT-ASR),并展示了VT-ASR机制如何与现有的物联网应用透明地集成。我们考虑了VoiceTalk的两个性能指标:语音识别精度和VT-ASR延迟。对于PAL-Labs的声学测试,VT-ASR的准确率为96.47%,而Google的准确率为94.28%。我们首先开发了一个分析模型来研究第一个说话者的VT-ASR延迟在第二个说话者开始说话之前完成的概率。从测量和分析建模中,我们表明VT-ASR延迟足够短,可以产生非常好的用户体验。我们的解决方案在台湾赢得了几个重要的政府和商业电视合同。在2018台塑语音识别挑战赛(FSR-2018)中,VT-ASR的台湾普通话识别准确率优于知名商用产品(包括Google和科大讯飞),在FSR-2020中,VT-ASR在所有参赛的ASR系统中台湾识别准确率最高。
{"title":"VoiceTalk: Multimedia-IoT Applications for Mixing Mandarin, Taiwanese, and English","authors":"Yi-Bing Lin, Yuan-Fu Liao, Sin-Horng Chen, Shaw-Hwa Hwang, Yih-Ru Wang","doi":"https://dl.acm.org/doi/10.1145/3543854","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3543854","url":null,"abstract":"<p>The voice-based Internet of Multimedia Things (IoMT) is the combination of IoT interfaces and protocols with associated voice-related information, which enables advanced applications based on human-to-device interactions. An example is Automatic Speech Recognition (ASR) for live captioning and voice translation. Three major issues of ASR for IoMT are IoT development cost, speech recognition accuracy, and execution time complexity. For the first issue, most non-voice IoT applications are upgraded with the ASR feature through hard coding, which are error prone. For the second issue, recognition accuracy must be improved for ASR. For the third issue, many multimedia IoT services are real-time applications and, therefore, the ASR delay must be short.</p><p>This article elaborates on the above issues based on an IoT platform called VoiceTalk. We built the largest Taiwanese spoken corpus to train <b>VoiceTalk ASR (VT-ASR)</b> and show how the VT-ASR mechanism can be transparently integrated with existing IoT applications. We consider two performance measures for VoiceTalk: speech recognition accuracy and VT-ASR delay. For the acoustic tests of PAL-Labs, VT-ASR's accuracy is 96.47%, while Google's accuracy is 94.28%. We are the first to develop an analytic model to investigate the probability that the VT-ASR delay for the first speaker is complete before the second speaker starts talking. From the measurements and analytic modeling, we show that the VT-ASR delay is short enough to result in a very good user experience. Our solution has won several important government and commercial TV contracts in Taiwan. VT-ASR has demonstrated better Taiwanese Mandarin speech recognition accuracy than famous commercial products (including Google and Iflytek) in Formosa Speech Recognition Challenge 2018 (FSR-2018) and was the best among all participating ASR systems for Taiwanese recognition accuracy in FSR-2020.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"19 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-18DOI: https://dl.acm.org/doi/10.1145/3584020
Yibin Xu, Jianhua Shao, Tijs Slaats, Boris Düdder
Blockchain sharding splits a blockchain into several shards where consensus is reached at the shard level rather than over the entire blockchain. It improves transaction throughput and reduces the computational resources required of individual nodes. But a derivation of trustworthy consensus within a shard becomes an issue as the longest chain based mechanisms used in conventional blockchains can no longer be used. Instead, a vote-based consensus mechanism must be employed. However, existing vote-based Byzantine fault tolerance consensus protocols do not offer sufficient security guarantees for sharded blockchains. First, when used to support consensus where only one block is allowed at a time (binary consensus), these protocols are susceptible to progress-hindering attacks (i.e., unable to reach a consensus). Second, when used to support a stronger type of consensus where multiple concurrent blocks are allowed (strong consensus), their tolerance of adversary nodes is low. This article proposes a new consensus protocol to address all these issues. We call the new protocol MWPoW+, as its basic framework is based on the existing Multiple Winners Proof of Work (MWPoW) protocol but includes new mechanisms to address the issues mentioned previously. MWPoW+ is a vote-based protocol for strong consensus, asynchronous in consensus derivation but synchronous in communication. We prove that it can tolerate up to f < n/2 adversary nodes in a n-node system as if using a binary consensus protocol and does not suffer from progress-hindering attacks.
{"title":"MWPoW+: A Strong Consensus Protocol for Intra-Shard Consensus in Blockchain Sharding","authors":"Yibin Xu, Jianhua Shao, Tijs Slaats, Boris Düdder","doi":"https://dl.acm.org/doi/10.1145/3584020","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3584020","url":null,"abstract":"<p>Blockchain sharding splits a blockchain into several shards where consensus is reached at the shard level rather than over the entire blockchain. It improves transaction throughput and reduces the computational resources required of individual nodes. But a derivation of trustworthy consensus within a shard becomes an issue as the longest chain based mechanisms used in conventional blockchains can no longer be used. Instead, a vote-based consensus mechanism must be employed. However, existing vote-based Byzantine fault tolerance consensus protocols do not offer sufficient security guarantees for sharded blockchains. First, when used to support consensus where only one block is allowed at a time (binary consensus), these protocols are susceptible to progress-hindering attacks (i.e., unable to reach a consensus). Second, when used to support a stronger type of consensus where multiple concurrent blocks are allowed (strong consensus), their tolerance of adversary nodes is low. This article proposes a new consensus protocol to address all these issues. We call the new protocol <i>MWPoW</i>+, as its basic framework is based on the existing Multiple Winners Proof of Work (MWPoW) protocol but includes new mechanisms to address the issues mentioned previously. MWPoW+ is a vote-based protocol for strong consensus, asynchronous in consensus derivation but synchronous in communication. We prove that it can tolerate up to <i>f</i> < <i>n</i>/2 adversary nodes in a n-node system as if using a binary consensus protocol and does not suffer from progress-hindering attacks.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"11 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-18DOI: https://dl.acm.org/doi/10.1145/3533430
Yazhou Zhang, Dan Ma, Prayag Tiwari, Chen Zhang, Mehedi Masud, Mohammad Shorfuzzaman, Dawei Song
Computational Linguistics (CL) associated with the Internet of Multimedia Things (IoMT)-enabled multimedia computing applications brings several research challenges, such as real-time speech understanding, deep fake video detection, emotion recognition, home automation, and so on. Due to the emergence of machine translation, CL solutions have increased tremendously for different natural language processing (NLP) applications. Nowadays, NLP-enabled IoMT is essential for its success. Sarcasm detection, a recently emerging artificial intelligence (AI) and NLP task, aims at discovering sarcastic, ironic, and metaphoric information implied in texts that are generated in the IoMT. It has drawn much attention from the AI and IoMT research community. The advance of sarcasm detection and NLP techniques will provide a cost-effective, intelligent way to work together with machine devices and high-level human-to-device interactions. However, existing sarcasm detection approaches neglect the hidden stance behind texts, thus insufficient to exploit the full potential of the task. Indeed, the stance, i.e., whether the author of a text is in favor of, against, or neutral toward the proposition or target talked in the text, largely determines the text’s actual sarcasm orientation. To fill the gap, in this research, we propose a new task: stance-level sarcasm detection (SLSD), where the goal is to uncover the author’s latent stance and based on it to identify the sarcasm polarity expressed in the text. We then propose an integral framework, which consists of Bidirectional Encoder Representations from Transformers (BERT) and a novel stance-centered graph attention networks (SCGAT). Specifically, BERT is used to capture the sentence representation, and SCGAT is designed to capture the stance information on specific target. Extensive experiments are conducted on a Chinese sarcasm sentiment dataset we created and the SemEval-2018 Task 3 English sarcasm dataset. The experimental results prove the effectiveness of the SCGAT framework over state-of-the-art baselines by a large margin.
{"title":"Stance-level Sarcasm Detection with BERT and Stance-centered Graph Attention Networks","authors":"Yazhou Zhang, Dan Ma, Prayag Tiwari, Chen Zhang, Mehedi Masud, Mohammad Shorfuzzaman, Dawei Song","doi":"https://dl.acm.org/doi/10.1145/3533430","DOIUrl":"https://doi.org/https://dl.acm.org/doi/10.1145/3533430","url":null,"abstract":"<p>Computational Linguistics (CL) associated with the Internet of Multimedia Things (IoMT)-enabled multimedia computing applications brings several research challenges, such as real-time speech understanding, deep fake video detection, emotion recognition, home automation, and so on. Due to the emergence of machine translation, CL solutions have increased tremendously for different natural language processing (NLP) applications. Nowadays, NLP-enabled IoMT is essential for its success. Sarcasm detection, a recently emerging artificial intelligence (AI) and NLP task, aims at discovering sarcastic, ironic, and metaphoric information implied in texts that are generated in the IoMT. It has drawn much attention from the AI and IoMT research community. The advance of sarcasm detection and NLP techniques will provide a cost-effective, intelligent way to work together with machine devices and high-level human-to-device interactions. However, existing sarcasm detection approaches neglect the hidden stance behind texts, thus insufficient to exploit the full potential of the task. Indeed, the stance, i.e., whether the author of a text is in favor of, against, or neutral toward the proposition or target talked in the text, largely determines the text’s actual sarcasm orientation. To fill the gap, in this research, we propose a new task: stance-level sarcasm detection (SLSD), where the goal is to uncover the author’s latent stance and based on it to identify the sarcasm polarity expressed in the text. We then propose an integral framework, which consists of Bidirectional Encoder Representations from Transformers (BERT) and a novel stance-centered graph attention networks (SCGAT). Specifically, BERT is used to capture the sentence representation, and SCGAT is designed to capture the stance information on specific target. Extensive experiments are conducted on a Chinese sarcasm sentiment dataset we created and the SemEval-2018 Task 3 English sarcasm dataset. The experimental results prove the effectiveness of the SCGAT framework over state-of-the-art baselines by a large margin.</p>","PeriodicalId":50911,"journal":{"name":"ACM Transactions on Internet Technology","volume":"16 1","pages":""},"PeriodicalIF":5.3,"publicationDate":"2023-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138533475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}