首页 > 最新文献

VNU Journal of Science: Computer Science and Communication Engineering最新文献

英文 中文
Aspect-Category based Sentiment Analysis with Unified Sequence-To-Sequence Transfer Transformers 基于方面分类的统一序对序传输变压器情感分析
Pub Date : 2023-08-07 DOI: 10.25073/2588-1086/vnucsce.662
D. Thin, N. Nguyen
In recent years, Aspect-based sentiment analysis (ABSA) has received increasing attention from the scientific community for Vietnamese language. However, most previous studies solved various subtasks in ABSA based on machine learning, deep learning and transformer-based architectures in a classification way. Recently, the release of pre-trained sequence-to-sequence brings a new approach to address the ABSA as a text generation problem for Vietnamese ABSA tasks. In this paper, we formulate the Aspect-category based sentiment analysis task as the conditional text generation task and investigate different unified generative transformer-based models. To represent the labels in a natural sentence, we apply a simple statistical method and observation of the commenting style. We conduct experiments on two benchmark datasets. As a result, our model achieved new state-of-the-art results with the micro F1-score of 75.53% and 86.60% for the two datasets with different levels for the restaurant domain. In addition, our experimental results achieved the best score for the smartphone domain with the macro F1-score of 81.10%.
近年来,基于方面的情感分析(ABSA)越来越受到越南语科学界的关注。然而,以往的研究大多以分类的方式解决了基于机器学习、深度学习和基于变压器的体系结构的ABSA中的各种子任务。最近,预训练序列到序列的发布带来了一种新的方法来解决越南ABSA任务的ABSA文本生成问题。本文提出了基于方面类别的情感分析任务作为条件文本生成任务,并研究了不同的统一的基于生成转换的模型。为了表示自然句子中的标签,我们采用了简单的统计方法和对评论风格的观察。我们在两个基准数据集上进行了实验。结果,我们的模型获得了新的最先进的结果,对于餐馆领域的两个不同级别的数据集,微观f1得分分别为75.53%和86.60%。此外,我们的实验结果在智能手机领域获得了最好的分数,宏f1得分为81.10%。
{"title":"Aspect-Category based Sentiment Analysis with Unified Sequence-To-Sequence Transfer Transformers","authors":"D. Thin, N. Nguyen","doi":"10.25073/2588-1086/vnucsce.662","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.662","url":null,"abstract":"In recent years, Aspect-based sentiment analysis (ABSA) has received increasing attention from the scientific community for Vietnamese language. However, most previous studies solved various subtasks in ABSA based on machine learning, deep learning and transformer-based architectures in a classification way. Recently, the release of pre-trained sequence-to-sequence brings a new approach to address the ABSA as a text generation problem for Vietnamese ABSA tasks. In this paper, we formulate the Aspect-category based sentiment analysis task as the conditional text generation task and investigate different unified generative transformer-based models. To represent the labels in a natural sentence, we apply a simple statistical method and observation of the commenting style. We conduct experiments on two benchmark datasets. As a result, our model achieved new state-of-the-art results with the micro F1-score of 75.53% and 86.60% for the two datasets with different levels for the restaurant domain. In addition, our experimental results achieved the best score for the smartphone domain with the macro F1-score of 81.10%.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121241789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Bandwidth-Efficient High-Performance RTL-Microarchitecture of 2D-Convolution for Deep Neural Networks 一种带宽高效的深度神经网络二维卷积高性能rtl微架构
Pub Date : 2023-08-07 DOI: 10.25073/2588-1086/vnucsce.596
Hung K. Nguyen, Long Quoc Tran
The computation complexity and huge memory access bandwidth of the convolutional layers in convolutional neural networks (CNNs) require specialized hardware architectures to accelerate CNN’s computations while keeping hardware costs reasonable for area-constrained embedded applications. This paper presents an RTL (Register Transfer Logic) level microarchitecture of hardware- and bandwidth-efficient high-performance 2D convolution unit for CNN in deep learning. The 2D convolution unit is made up of three main components including a dedicated Loader, a Circle Buffer, and a MAC (Multiplier-Accumulator) unit. The 2D convolution unit has a 2-stage pipeline structure that reduces latency, increases processing throughput, and reduces power consumption. The architecture proposed in the paper eliminates the reloading of both the weights as well as the input image data. The 2D convolution unit is configurable to support 2D convolution operations with different sizes of input image matrix and kernel filter. The architecture can reduce memory access time and power as well as execution time thanks to the efficient reuse of the preloaded input data, while simplifying hardware implementation. The 2D convolution unit has been simulated and implemented on Xilinx's FPGA platform to evaluate its superiority. Experimental results show that our design is 1.54× and 13.6× faster in performance than the design in [7] and [8], respectively, at lower hardware cost without using any FPGA’s dedicated hardware blocks. By reusing preloaded data, our design achieves a bandwidth reduction ratio between 66.4% and 90.5%.
卷积神经网络(CNN)中卷积层的计算复杂性和巨大的内存访问带宽需要专门的硬件架构来加速CNN的计算,同时在限制面积的嵌入式应用中保持合理的硬件成本。本文提出了一种RTL(寄存器传输逻辑)级微架构,用于CNN深度学习中具有硬件和带宽效率的高性能二维卷积单元。二维卷积单元由三个主要组件组成,包括专用的Loader、Circle Buffer和MAC (Multiplier-Accumulator)单元。2D卷积单元具有两级管道结构,可减少延迟,提高处理吞吐量并降低功耗。本文提出的结构消除了权重和输入图像数据的重新加载。二维卷积单元可配置,支持不同大小的输入图像矩阵和核滤波器的二维卷积操作。由于有效地重用预加载的输入数据,该架构可以减少内存访问时间和功耗以及执行时间,同时简化硬件实现。在Xilinx的FPGA平台上对二维卷积单元进行了仿真和实现,以评估其优越性。实验结果表明,在不使用FPGA专用硬件模块的情况下,本设计的性能比文献[7]和[8]中的设计分别提高了1.54倍和13.6倍,硬件成本更低。通过重用预加载数据,我们的设计实现了66.4%到90.5%的带宽减少率。
{"title":"A Bandwidth-Efficient High-Performance RTL-Microarchitecture of 2D-Convolution for Deep Neural Networks","authors":"Hung K. Nguyen, Long Quoc Tran","doi":"10.25073/2588-1086/vnucsce.596","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.596","url":null,"abstract":"The computation complexity and huge memory access bandwidth of the convolutional layers in convolutional neural networks (CNNs) require specialized hardware architectures to accelerate CNN’s computations while keeping hardware costs reasonable for area-constrained embedded applications. This paper presents an RTL (Register Transfer Logic) level microarchitecture of hardware- and bandwidth-efficient high-performance 2D convolution unit for CNN in deep learning. The 2D convolution unit is made up of three main components including a dedicated Loader, a Circle Buffer, and a MAC (Multiplier-Accumulator) unit. The 2D convolution unit has a 2-stage pipeline structure that reduces latency, increases processing throughput, and reduces power consumption. The architecture proposed in the paper eliminates the reloading of both the weights as well as the input image data. The 2D convolution unit is configurable to support 2D convolution operations with different sizes of input image matrix and kernel filter. The architecture can reduce memory access time and power as well as execution time thanks to the efficient reuse of the preloaded input data, while simplifying hardware implementation. The 2D convolution unit has been simulated and implemented on Xilinx's FPGA platform to evaluate its superiority. Experimental results show that our design is 1.54× and 13.6× faster in performance than the design in [7] and [8], respectively, at lower hardware cost without using any FPGA’s dedicated hardware blocks. By reusing preloaded data, our design achieves a bandwidth reduction ratio between 66.4% and 90.5%.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125919106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Noisy-label propagation for Video Anomaly Detection with Graph Transformer Network 基于图变换网络的视频异常检测中的噪声标签传播
Pub Date : 2023-06-16 DOI: 10.25073/2588-1086/vnucsce.659
Viet-Cuong Ta, Thu Uyen Do
In this paper, we study the efficiency of Graph Transformer Network for noisy label propagation in the task of classifying video anomaly actions. Given a weak supervised dataset, our methods focus on improving the quality of generated labels and use the labels for training a video classifier with deep network. From a full-length video, the anomaly properties of each segmented video can be decided through their relationship with other video. Therefore, we employ a label propagation mechanism with Graph Transformer Network. Our network combines both the feature-based relationship and temporal-based relationship to project the output features of the anomaly video to a hidden dimension. By learning in the new dimension, the video classifier can improve the quality of noisy, generated labels. Our experiments on three benchmark dataset show that the accuracy of our methods are better and more stable than other tested baselines.
本文研究了图变换网络在视频异常动作分类任务中的噪声标签传播效率。给定一个弱监督数据集,我们的方法侧重于提高生成标签的质量,并使用这些标签训练深度网络视频分类器。从一个完整的视频中,可以通过每个分段视频与其他视频的关系来判断其异常属性。因此,我们在图转换网络中采用了一种标签传播机制。我们的网络结合了基于特征的关系和基于时间的关系,将异常视频的输出特征投影到隐藏维度。通过在新的维度上学习,视频分类器可以提高生成的有噪声标签的质量。我们在三个基准数据集上的实验表明,我们的方法比其他测试基线的准确性更好,更稳定。
{"title":"Noisy-label propagation for Video Anomaly Detection with Graph Transformer Network","authors":"Viet-Cuong Ta, Thu Uyen Do","doi":"10.25073/2588-1086/vnucsce.659","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.659","url":null,"abstract":"In this paper, we study the efficiency of Graph Transformer Network for noisy label propagation in the task of classifying video anomaly actions. Given a weak supervised dataset, our methods focus on improving the quality of generated labels and use the labels for training a video classifier with deep network. From a full-length video, the anomaly properties of each segmented video can be decided through their relationship with other video. Therefore, we employ a label propagation mechanism with Graph Transformer Network. Our network combines both the feature-based relationship and temporal-based relationship to project the output features of the anomaly video to a hidden dimension. By learning in the new dimension, the video classifier can improve the quality of noisy, generated labels. Our experiments on three benchmark dataset show that the accuracy of our methods are better and more stable than other tested baselines.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129418158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FRSL: A Domain Specific Language to Specify Functional Requirements FRSL:用于指定功能需求的领域特定语言
Pub Date : 2023-06-06 DOI: 10.25073/2588-1086/vnucsce.803
Duc-Hanh Dang
In software development, to obtain a precise specification of the software system's functional requirements is significant to ensure the software quality as well as to automate the development. Use cases are an effective way to capture functional requirements, however, the use of ambiguous or vague language in the use case can lead to imprecision. It is essential to ensure that use case specifications are clear, concise, and complete to avoid imprecision in requirements. This paper aims to develop a domain specific language called FRSL to precisely specify use cases and to provide a basis for transformations to generate software artifacts from the use case specification. We define a FRSL metamodel to capture the technical domain of use cases for FRSL's abstract syntax, and then provides a textual concrete syntax for this language. We also define a formal operational semantics for FRSL by characterizing the execution of a FRSL specification as sequences of system snapshot transitions. This formal semantics on the one hand allows us to precisely explain the meaning of use cases and their relationships, on the other hand provides a basis for transformations from the use case specification. We implement a tool support for this language and perform an evaluation of its expressiveness in comparison with current use case specification languages. This work brings out (1)~a DSL to specify use cases that is defined based on a formal semantics of use cases; and (2)~a tool support realized as an Eclipse plugin for this DSL. The use case specification language FRSL would help precisely specify the system's functional requirements and bring more automation in the software development.
在软件开发中,准确地描述软件系统的功能需求,对保证软件质量和实现软件开发的自动化具有重要意义。用例是捕获功能需求的有效方法,然而,在用例中使用模棱两可或模糊的语言可能导致不精确。确保用例说明是清晰、简洁和完整的,以避免需求中的不精确,这是必不可少的。本文旨在开发一种称为FRSL的领域特定语言,以精确地指定用例,并为从用例规范生成软件工件的转换提供基础。我们定义一个FRSL元模型来捕获FRSL抽象语法用例的技术领域,然后为该语言提供文本的具体语法。我们还通过将FRSL规范的执行描述为系统快照转换序列来定义FRSL的形式化操作语义。这种形式化语义一方面允许我们精确地解释用例的含义和它们之间的关系,另一方面为用例规范的转换提供了基础。我们实现了对这种语言的工具支持,并通过与当前用例规范语言的比较,对其表达性进行了评估。这项工作带来了(1)~一个DSL来指定基于用例的形式化语义定义的用例;(2)为这个DSL实现一个Eclipse插件的工具支持。用例规范语言FRSL将有助于精确地指定系统的功能需求,并在软件开发中带来更多的自动化。
{"title":"FRSL: A Domain Specific Language to Specify Functional Requirements","authors":"Duc-Hanh Dang","doi":"10.25073/2588-1086/vnucsce.803","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.803","url":null,"abstract":"In software development, to obtain a precise specification of the software system's functional requirements is significant to ensure the software quality as well as to automate the development. Use cases are an effective way to capture functional requirements, however, the use of ambiguous or vague language in the use case can lead to imprecision. It is essential to ensure that use case specifications are clear, concise, and complete to avoid imprecision in requirements. This paper aims to develop a domain specific language called FRSL to precisely specify use cases and to provide a basis for transformations to generate software artifacts from the use case specification. We define a FRSL metamodel to capture the technical domain of use cases for FRSL's abstract syntax, and then provides a textual concrete syntax for this language. We also define a formal operational semantics for FRSL by characterizing the execution of a FRSL specification as sequences of system snapshot transitions. This formal semantics on the one hand allows us to precisely explain the meaning of use cases and their relationships, on the other hand provides a basis for transformations from the use case specification. We implement a tool support for this language and perform an evaluation of its expressiveness in comparison with current use case specification languages. This work brings out (1)~a DSL to specify use cases that is defined based on a formal semantics of use cases; and (2)~a tool support realized as an Eclipse plugin for this DSL. The use case specification language FRSL would help precisely specify the system's functional requirements and bring more automation in the software development.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132392629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Contract-Based Specification Method for Model Transformations 基于契约的模型转换规范方法
Pub Date : 2023-04-01 DOI: 10.25073/2588-1086/vnucsce.657
Duc-Hanh Dang, Thi-Hanh Nguyen
Model transformations play an essential role in model-driven engineering. However, model transformations are often complex to develop, maintain, and ensure quality. Platform-independent specification languages for transformations are required to fully and accurately express requirements of transformation systems and to offer support for realization and verification tasks. Several specification languages have been proposed, but it still lacks a strong one based on a solid formal foundation for both high expressiveness and usability. This paper introduces a language called TC4MT to precisely specify requirements of transformations. The language is designed based on a combination of a contract-based approach and the graph theory foundation of triple graph grammar. Specifically, we consider graph patterns as core elements of our language and provide a concrete syntax in the form of UML class diagrams together with OCL conditions to visually and intuitively represent such pattern-based specifications. We develop a support tool and evaluate our proposed method by comparing it with current methods in literature.
模型转换在模型驱动工程中起着重要的作用。然而,模型转换在开发、维护和确保质量方面通常是复杂的。需要独立于平台的转换规范语言来全面准确地表达转换系统的需求,并为实现和验证任务提供支持。已经提出了几种规范语言,但它仍然缺乏一种基于高表达性和可用性的坚实形式基础的强大规范语言。本文介绍了一种名为TC4MT的语言来精确地指定转换的需求。该语言是基于基于契约的方法和三图语法的图论基础相结合而设计的。具体来说,我们将图形模式视为我们语言的核心元素,并以UML类图的形式提供具体语法,以及OCL条件,以直观地表示这种基于模式的规范。我们开发了一个支持工具,并通过与文献中现有方法的比较来评估我们提出的方法。
{"title":"A Contract-Based Specification Method for Model Transformations","authors":"Duc-Hanh Dang, Thi-Hanh Nguyen","doi":"10.25073/2588-1086/vnucsce.657","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.657","url":null,"abstract":"Model transformations play an essential role in model-driven engineering. However, model transformations are often complex to develop, maintain, and ensure quality. Platform-independent specification languages for transformations are required to fully and accurately express requirements of transformation systems and to offer support for realization and verification tasks. Several specification languages have been proposed, but it still lacks a strong one based on a solid formal foundation for both high expressiveness and usability. This paper introduces a language called TC4MT to precisely specify requirements of transformations. The language is designed based on a combination of a contract-based approach and the graph theory foundation of triple graph grammar. Specifically, we consider graph patterns as core elements of our language and provide a concrete syntax in the form of UML class diagrams together with OCL conditions to visually and intuitively represent such pattern-based specifications. We develop a support tool and evaluate our proposed method by comparing it with current methods in literature.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116570937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
vnNLI - VLSP 2021: Vietnamese and English-Vietnamese Textual Entailment Based on Pre-trained Multilingual Language Models vnli - VLSP 2021:基于预训练多语言模型的越南语和英越语文本蕴涵
Pub Date : 2022-12-16 DOI: 10.25073/2588-1086/vnucsce.329
Ngan Nguyen Luu Thuy, Đặng Văn Thìn, Hoàng Xuân Vũ, Nguyễn Văn Tài, Khoa Thi-Kim Phan
Natural Language Inference (NLI) is a high-level semantic task in Natural Language Processing - NLP, and it extends further challenges if it is in the cross-lingual scenario. In recent years, pre-trained multilingual language models (e.g., mBERT ,XLM-R, InfoXLM) have greatly contributed to the success of dealing with these challenges. Based on the motivation behind these achievements, this paper describes our approach based on fine-tuning pretrained multilingual language models (XLM-R, InfoXLM) to tackle the shared task ``Vietnamese and English-Vietnamese Textual Entailment'' at the 8th International Workshop on Vietnamese Language and Speech Processing (VLSP 2021footnote{https://vlsp.org.vn/vlsp2021}). We investigate other techniques to improve the performance of our work: Cross-validation, Pseudo-labeling (PL), Learning rate adjustment, and Postagging. All experimental results demonstrated that our approach based on the InfoXLM model achieved competitive results, ranking 2nd for the task evaluation in VLSP 2021 with 0.89 in terms of F1-score on the private test set.
自然语言推理(NLI)是自然语言处理(NLP)中的高级语义任务,如果是在跨语言场景中,它将进一步扩展挑战。近年来,预训练的多语言模型(例如,mBERT、XLM-R、InfoXLM)为成功应对这些挑战做出了巨大贡献。基于这些成就背后的动机,本文描述了我们在第八届越南语言和语音处理国际研讨会(VLSP 2021 footnote{https://vlsp.org.vn/vlsp2021})上基于微调预训练的多语言语言模型(XLM-R, InfoXLM)来解决共享任务“越南语和英语/越南语文本蕴因”的方法。我们研究了其他技术来提高我们的工作性能:交叉验证,伪标记(PL),学习率调整和Postagging。所有实验结果都表明,我们基于InfoXLM模型的方法取得了有竞争力的结果,在VLSP 2021中以0.89的F1-score在私有测试集中排名第二。
{"title":"vnNLI - VLSP 2021: Vietnamese and English-Vietnamese Textual Entailment Based on Pre-trained Multilingual Language Models","authors":"Ngan Nguyen Luu Thuy, Đặng Văn Thìn, Hoàng Xuân Vũ, Nguyễn Văn Tài, Khoa Thi-Kim Phan","doi":"10.25073/2588-1086/vnucsce.329","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.329","url":null,"abstract":"Natural Language Inference (NLI) is a high-level semantic task in Natural Language Processing - NLP, and it extends further challenges if it is in the cross-lingual scenario. In recent years, pre-trained multilingual language models (e.g., mBERT ,XLM-R, InfoXLM) have greatly contributed to the success of dealing with these challenges. Based on the motivation behind these achievements, this paper describes our approach based on fine-tuning pretrained multilingual language models (XLM-R, InfoXLM) to tackle the shared task ``Vietnamese and English-Vietnamese Textual Entailment'' at the 8th International Workshop on Vietnamese Language and Speech Processing (VLSP 2021footnote{https://vlsp.org.vn/vlsp2021}). We investigate other techniques to improve the performance of our work: Cross-validation, Pseudo-labeling (PL), Learning rate adjustment, and Postagging. All experimental results demonstrated that our approach based on the InfoXLM model achieved competitive results, ranking 2nd for the task evaluation in VLSP 2021 with 0.89 in terms of F1-score on the private test set.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124389717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Early CTU Termination and Three-steps Mode Decision Method for Fast Versatile Video Coding 快速通用视频编码的早期CTU终止和三步模式判定方法
Pub Date : 2022-12-16 DOI: 10.25073/2588-1086/vnucsce.375
S. Q. Nguyen, Tien Huu Vu, Duong Trieu Dinh, Minh Bao Dinh, Minh N. Do, X. Hoang
Versatile Video Coding (VVC) has been recently becoming popular in coding videos due to its compression efficiency. To reach this performance, Joint Video Experts Team (JVET) has introduced a number of improvement techniques to VVC coding model. Among them, VVC Intra coding introduces a new concept of quad-tree nested multi-type tree (QTMT) and extends the predicted modes with up to 67 options. As a result, the complexity of the VVC Intra encoding also greatly increases. To make VVC Intra coding more feasible in real-time applications, we propose in this paper a novel deep learning based fast QTMT and an early mode prediction method. At the first stage, we use a learned convolutional neural network (CNN) model to predict the coding unit map and then fed into the VVC encoder to early terminate the block partitioning process. After that, we design a statistical model to predict a list of most probable modes (MPM) for each selected Coding using (CU) size. Finally, we employ a so-called three-steps mode decision algorithm to estimate the optimal directional mode without sacrificing the compression performance. The proposed early CU splitting and fast intra prediction are integrated into the latest VTM reference software. Experimental results show that the proposed method can save 50.2% of encoding time with a negligible BD-Rate increase.
多功能视频编码(VVC)由于其压缩效率高,近年来在视频编码中得到了广泛的应用。为了达到这一性能,联合视频专家小组(JVET)对VVC编码模型引入了许多改进技术。其中,VVC Intra编码引入了四叉树嵌套多类型树(QTMT)的新概念,将预测模式扩展到67个选项。因此,VVC Intra编码的复杂性也大大增加。为了使VVC Intra编码在实时应用中更加可行,本文提出了一种基于深度学习的快速QTMT和早期模式预测方法。在第一阶段,我们使用学习卷积神经网络(CNN)模型预测编码单元映射,然后将其输入到VVC编码器中,以提前终止块划分过程。之后,我们设计了一个统计模型来预测每个选择的编码使用(CU)大小的最可能模式(MPM)列表。最后,我们采用所谓的三步模式决策算法来估计最优的方向模式,而不牺牲压缩性能。在最新的VTM参考软件中集成了提出的早期CU分割和快速内预测。实验结果表明,该方法可以节省50.2%的编码时间,而BD-Rate的提高可以忽略不计。
{"title":"Early CTU Termination and Three-steps Mode Decision Method for Fast Versatile Video Coding","authors":"S. Q. Nguyen, Tien Huu Vu, Duong Trieu Dinh, Minh Bao Dinh, Minh N. Do, X. Hoang","doi":"10.25073/2588-1086/vnucsce.375","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.375","url":null,"abstract":"Versatile Video Coding (VVC) has been recently becoming popular in coding videos due to its compression efficiency. To reach this performance, Joint Video Experts Team (JVET) has introduced a number of improvement techniques to VVC coding model. Among them, VVC Intra coding introduces a new concept of quad-tree nested multi-type tree (QTMT) and extends the predicted modes with up to 67 options. As a result, the complexity of the VVC Intra encoding also greatly increases. To make VVC Intra coding more feasible in real-time applications, we propose in this paper a novel deep learning based fast QTMT and an early mode prediction method. At the first stage, we use a learned convolutional neural network (CNN) model to predict the coding unit map and then fed into the VVC encoder to early terminate the block partitioning process. After that, we design a statistical model to predict a list of most probable modes (MPM) for each selected Coding using (CU) size. Finally, we employ a so-called three-steps mode decision algorithm to estimate the optimal directional mode without sacrificing the compression performance. The proposed early CU splitting and fast intra prediction are integrated into the latest VTM reference software. Experimental results show that the proposed method can save 50.2% of encoding time with a negligible BD-Rate increase.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131282797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
vnNLI - VLSP2021: An Empirical Study on Vietnamese-English Natural Language Inference Based on Pretrained Language Models with Data Augmentation vnli - VLSP2021:基于数据增强的预训练语言模型的越英自然语言推理实证研究
Pub Date : 2022-12-16 DOI: 10.25073/2588-1086/vnucsce.330
Thin Dang Van, D. Hao, N. Nguyen, Luân Đình Ngô, Kiến Lê Hiếu Ngô
In this paper, we describe an empirical study of data augmentation techniques with various pre-trained language models on the bilingual dataset which was presented at the VLSP 2021 - Vietnamese and English-Vietnamese Textual Entailment. We apply the machine translation tool to generate new training set from original training data and then  investigate and compare the effectiveness of a monolingual and multilingual model on the new data set. Our experimental results show that fine-tuning a pre-trained multilingual language XLM-R model with an augmented training set gives the best performance. Our system was ranked third in the shared-task VLSP 2021 with the  F1-score of about 0.88.
在本文中,我们描述了在VLSP 2021 -越南语和英语-越南语文本蕴意上发表的双语数据集上使用各种预训练语言模型进行数据增强技术的实证研究。我们使用机器翻译工具从原始训练数据生成新的训练集,然后研究和比较单语言和多语言模型在新数据集上的有效性。实验结果表明,使用增强训练集对预训练的多语言XLM-R模型进行微调可以获得最佳性能。我们的系统在共享任务VLSP 2021中排名第三,f1得分约为0.88。
{"title":"vnNLI - VLSP2021: An Empirical Study on Vietnamese-English Natural Language Inference Based on Pretrained Language Models with Data Augmentation","authors":"Thin Dang Van, D. Hao, N. Nguyen, Luân Đình Ngô, Kiến Lê Hiếu Ngô","doi":"10.25073/2588-1086/vnucsce.330","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.330","url":null,"abstract":"In this paper, we describe an empirical study of data augmentation techniques with various pre-trained language models on the bilingual dataset which was presented at the VLSP 2021 - Vietnamese and English-Vietnamese Textual Entailment. We apply the machine translation tool to generate new training set from original training data and then  investigate and compare the effectiveness of a monolingual and multilingual model on the new data set. Our experimental results show that fine-tuning a pre-trained multilingual language XLM-R model with an augmented training set gives the best performance. Our system was ranked third in the shared-task VLSP 2021 with the  F1-score of about 0.88.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126472102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Method for Test Data Generation for Unit Testing of C/C++ Projects C/ c++项目单元测试中测试数据生成的混合方法
Pub Date : 2022-12-16 DOI: 10.25073/2588-1086/vnucsce.354
Tran Nguyen Huong
In recent years, automated test data generation from source code has gained a significantpopularity in software testing. This paper proposes a method, named Hybrid, to generate test datafor unit testing C/C++ projects. The method is a combination of two test data generation methodsnamed IBVGT and WCFT. In IBVGT method, the source code is analyzed to find simple conditions.Then, bases on these conditions, IBVGT generates test data for boundary values without having tosolve test paths constraints. This makes the method faster than BVTG method when generating testdata. In Hybrid method, while generating test data using WCFT, simple conditions are collected forboundary values test data generation. Test data generated by Hybrid are able to ensure both highsource code coverage and error detection ability. In addition, Hybrid is capable of finding infeasibleexecution paths and dead code. Experimental results with some popular unit functions show thatHybrid outperforms STCFG method in terms of test data generation time and boundary values relatederror detection. IBVGT is superior to BVTG in term of test data generation time whilst its boundaryvalues related error detection ability depends on the number of simple conditions inside each unitfunction.
近年来,从源代码自动生成测试数据在软件测试中获得了显著的普及。本文提出了一种为C/ c++项目的单元测试生成测试数据的方法——Hybrid。该方法结合了IBVGT和WCFT两种测试数据生成方法。在IBVGT方法中,对源代码进行分析,找到简单的条件。然后,基于这些条件,IBVGT生成边界值的测试数据,而无需解决测试路径约束。这使得该方法在生成测试数据时比BVTG方法更快。在混合方法中,在使用WCFT生成测试数据的同时,收集简单的条件来生成边界值测试数据。Hybrid生成的测试数据能够保证高源代码覆盖率和错误检测能力。此外,Hybrid能够找到不可行的执行路径和死代码。使用一些常用的单元函数进行的实验结果表明,athybrid在测试数据生成时间和边界值相关误差检测方面优于STCFG方法。IBVGT在测试数据生成时间上优于BVTG,而其边界值相关的错误检测能力取决于每个单元函数内部简单条件的个数。
{"title":"A Hybrid Method for Test Data Generation for Unit Testing of C/C++ Projects","authors":"Tran Nguyen Huong","doi":"10.25073/2588-1086/vnucsce.354","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.354","url":null,"abstract":"In recent years, automated test data generation from source code has gained a significantpopularity in software testing. This paper proposes a method, named Hybrid, to generate test datafor unit testing C/C++ projects. The method is a combination of two test data generation methodsnamed IBVGT and WCFT. In IBVGT method, the source code is analyzed to find simple conditions.Then, bases on these conditions, IBVGT generates test data for boundary values without having tosolve test paths constraints. This makes the method faster than BVTG method when generating testdata. In Hybrid method, while generating test data using WCFT, simple conditions are collected forboundary values test data generation. Test data generated by Hybrid are able to ensure both highsource code coverage and error detection ability. In addition, Hybrid is capable of finding infeasibleexecution paths and dead code. Experimental results with some popular unit functions show thatHybrid outperforms STCFG method in terms of test data generation time and boundary values relatederror detection. IBVGT is superior to BVTG in term of test data generation time whilst its boundaryvalues related error detection ability depends on the number of simple conditions inside each unitfunction.","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129205211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
vieCap4H Challenge 2021: A transformer-based method for Healthcare Image Captioning in Vietnamese vieCap4H挑战2021:越南医疗保健图像字幕的基于转换器的方法
Pub Date : 2022-12-16 DOI: 10.25073/2588-1086/vnucsce.371
Doanh Bui Cao, Truc Thi Thanh Trinh, Trong-Thuan Nguyen, V. Nguyen, Nguyen D. Vo
The automatic image caption generation is attractive to both Computer Vision and Natural Language Processing research community because it lies in the gap between these two fields. Within the vieCap4H contest organized by VLSP 2021, we participate and present a Transformer-based solution for image captioning in the healthcare domain. In detail, we use grid features as visual presentation and pre-training a BERT-based language model from PhoBERT-base pre-trained model to obtain language presentation used in the Adaptive Decoder module in the RSTNet model. Besides, we indicate a suitable schedule with the self-critical training sequence (SCST) technique to achieve the best results. Through experiments, we achieve an average of 30.3% BLEU score on the public-test round and 28.9% on the private-test round, which ranks 3rd and 4th, respectively. Source code is available at https://github.com/caodoanh2001/uit-vlsp-viecap4h-solution.  
由于图像标题的自动生成处于计算机视觉和自然语言处理两个领域的空白地带,因此受到了计算机视觉和自然语言处理研究领域的广泛关注。在VLSP 2021组织的vieCap4H竞赛中,我们参与并展示了一个基于transformer的医疗保健领域图像字幕解决方案。具体来说,我们使用网格特征作为视觉表示,并从基于phobert的预训练模型中预训练一个基于bert的语言模型,以获得RSTNet模型中Adaptive Decoder模块使用的语言表示。此外,我们提出了一个合适的时间表与自我批判训练序列(SCST)技术,以达到最佳效果。通过实验,我们在公测轮和私测轮的BLEU平均分分别达到30.3%和28.9%,分别排名第3和第4。源代码可从https://github.com/caodoanh2001/uit-vlsp-viecap4h-solution获得。
{"title":"vieCap4H Challenge 2021: A transformer-based method for Healthcare Image Captioning in Vietnamese","authors":"Doanh Bui Cao, Truc Thi Thanh Trinh, Trong-Thuan Nguyen, V. Nguyen, Nguyen D. Vo","doi":"10.25073/2588-1086/vnucsce.371","DOIUrl":"https://doi.org/10.25073/2588-1086/vnucsce.371","url":null,"abstract":"The automatic image caption generation is attractive to both Computer Vision and Natural Language Processing research community because it lies in the gap between these two fields. Within the vieCap4H contest organized by VLSP 2021, we participate and present a Transformer-based solution for image captioning in the healthcare domain. In detail, we use grid features as visual presentation and pre-training a BERT-based language model from PhoBERT-base pre-trained model to obtain language presentation used in the Adaptive Decoder module in the RSTNet model. Besides, we indicate a suitable schedule with the self-critical training sequence (SCST) technique to achieve the best results. Through experiments, we achieve an average of 30.3% BLEU score on the public-test round and 28.9% on the private-test round, which ranks 3rd and 4th, respectively. Source code is available at https://github.com/caodoanh2001/uit-vlsp-viecap4h-solution. \u0000 ","PeriodicalId":416488,"journal":{"name":"VNU Journal of Science: Computer Science and Communication Engineering","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127272648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
VNU Journal of Science: Computer Science and Communication Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1