首页 > 最新文献

Machine Intelligence Research最新文献

英文 中文
Rolling Shutter Camera: Modeling, Optimization and Learning 卷帘式相机:建模、优化和学习
4区 计算机科学 Pub Date : 2023-11-09 DOI: 10.1007/s11633-022-1399-z
Bin Fan, Yuchao Dai, Mingyi He
{"title":"Rolling Shutter Camera: Modeling, Optimization and Learning","authors":"Bin Fan, Yuchao Dai, Mingyi He","doi":"10.1007/s11633-022-1399-z","DOIUrl":"https://doi.org/10.1007/s11633-022-1399-z","url":null,"abstract":"","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135291213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
State of the Art on Deep Learning-enhanced Rendering Methods 深度学习增强渲染方法的最新进展
4区 计算机科学 Pub Date : 2023-11-09 DOI: 10.1007/s11633-022-1400-x
Qi Wang, Zhihua Zhong, Yuchi Huo, Hujun Bao, Rui Wang
{"title":"State of the Art on Deep Learning-enhanced Rendering Methods","authors":"Qi Wang, Zhihua Zhong, Yuchi Huo, Hujun Bao, Rui Wang","doi":"10.1007/s11633-022-1400-x","DOIUrl":"https://doi.org/10.1007/s11633-022-1400-x","url":null,"abstract":"","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135290693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transmission Line Insulator Defect Detection Based on Swin Transformer and Context 基于Swin变压器和上下文的输电线路绝缘子缺陷检测
4区 计算机科学 Pub Date : 2023-09-15 DOI: 10.1007/s11633-022-1355-y
Yu Xi, Ke Zhou, Ling-Wen Meng, Bo Chen, Hao-Min Chen, Jing-Yi Zhang
Insulators are important components of power transmission lines. Once a failure occurs, it may cause a large-scale blackout and other hidden dangers. Due to the large image size and complex background, detecting small defect objects is a challenge. We make improvements based on the two-stage network Faster R-convolutional neural networks (CNN). First, we use a hierarchical Swin Transformer with shifted windows as the feature extraction network, instead of ResNet, to extract more discriminative features, and then design the deformable receptive field block to encode global and local context information, which is utilized to capture key clues for detecting objects in complex backgrounds. Finally, the filling data augmentation method is proposed for the problem of insufficient defects and more images of insulator defects under different backgrounds are added to the training set to improve the robustness of the model. As a result, the recall increases from 89.5% to 92.1%, and the average precision increases from 81.0% to 87.1%. To further prove the superiority of the proposed algorithm, we also tested the model on the public data set Pascal visual object classes (VOC), which also yields outstanding results.
绝缘子是输电线路的重要组成部分。一旦发生故障,可能造成大面积停电等隐患。由于图像尺寸大,背景复杂,小缺陷物体的检测是一个挑战。我们在两阶段网络的基础上改进了更快的r -卷积神经网络。首先,我们使用带移位窗口的分层Swin Transformer作为特征提取网络,代替ResNet提取更多的判别特征,然后设计可变形的接受场块,对全局和局部上下文信息进行编码,用于捕获复杂背景下目标检测的关键线索。最后,针对缺陷不足的问题,提出了填充数据增强方法,并在训练集中加入更多不同背景下的绝缘子缺陷图像,提高了模型的鲁棒性。召回率从89.5%提高到92.1%,平均准确率从81.0%提高到87.1%。为了进一步证明该算法的优越性,我们还在公共数据集Pascal visual object classes (VOC)上对该模型进行了测试,同样取得了显著的效果。
{"title":"Transmission Line Insulator Defect Detection Based on Swin Transformer and Context","authors":"Yu Xi, Ke Zhou, Ling-Wen Meng, Bo Chen, Hao-Min Chen, Jing-Yi Zhang","doi":"10.1007/s11633-022-1355-y","DOIUrl":"https://doi.org/10.1007/s11633-022-1355-y","url":null,"abstract":"Insulators are important components of power transmission lines. Once a failure occurs, it may cause a large-scale blackout and other hidden dangers. Due to the large image size and complex background, detecting small defect objects is a challenge. We make improvements based on the two-stage network Faster R-convolutional neural networks (CNN). First, we use a hierarchical Swin Transformer with shifted windows as the feature extraction network, instead of ResNet, to extract more discriminative features, and then design the deformable receptive field block to encode global and local context information, which is utilized to capture key clues for detecting objects in complex backgrounds. Finally, the filling data augmentation method is proposed for the problem of insufficient defects and more images of insulator defects under different backgrounds are added to the training set to improve the robustness of the model. As a result, the recall increases from 89.5% to 92.1%, and the average precision increases from 81.0% to 87.1%. To further prove the superiority of the proposed algorithm, we also tested the model on the public data set Pascal visual object classes (VOC), which also yields outstanding results.","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135396705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
YOLO-CORE: Contour Regression for Efficient Instance Segmentation YOLO-CORE:高效实例分割的轮廓回归
4区 计算机科学 Pub Date : 2023-09-15 DOI: 10.1007/s11633-022-1379-3
Haoliang Liu, Wei Xiong, Yu Zhang
Instance segmentation has drawn mounting attention due to its significant utility. However, high computational costs have been widely acknowledged in this domain, as the instance mask is generally achieved by pixel-level labeling. In this paper, we present a conceptually efficient contour regression network based on the you only look once (YOLO) architecture named YOLO-CORE for instance segmentation. The mask of the instance is efficiently acquired by explicit and direct contour regression using our designed multi-order constraint consisting of a polar distance loss and a sector loss. Our proposed YOLO-CORE yields impressive segmentation performance in terms of both accuracy and speed. It achieves 57.9% AP@0.5 with 47 FPS (frames per second) on the semantic boundaries dataset (SBD) and 51.1% AP@0.5 with 46 FPS on the COCO dataset. The superior performance achieved by our method with explicit contour regression suggests a new technique line in the YOLO-based image understanding field. Moreover, our instance segmentation design can be flexibly integrated into existing deep detectors with negligible computation cost (65.86 BFLOPs (billion float operations per second) to 66.15 BFLOPs with the YOLOv3 detector).
实例分割由于其重要的实用性而引起了越来越多的关注。然而,由于实例掩码通常是通过像素级标记来实现的,因此该领域的计算成本很高。在本文中,我们提出了一个概念上高效的轮廓回归网络,基于你只看一次(YOLO)架构,命名为YOLO- core,用于实例分割。利用我们设计的由极距离损失和扇形损失组成的多阶约束,通过显式和直接的轮廓回归有效地获取实例的掩码。我们提出的YOLO-CORE在准确性和速度方面都具有令人印象深刻的分割性能。它在语义边界数据集(SBD)上达到57.9% AP@0.5和47 FPS(帧/秒),在COCO数据集上达到51.1% AP@0.5和46 FPS。显式轮廓回归方法所取得的优异性能为基于yolo的图像理解领域开辟了一条新的技术路线。此外,我们的实例分割设计可以灵活地集成到现有的深度检测器中,计算成本可以忽略(65.86 BFLOPs(每秒十亿次浮点运算)到使用YOLOv3检测器的66.15 BFLOPs)。
{"title":"YOLO-CORE: Contour Regression for Efficient Instance Segmentation","authors":"Haoliang Liu, Wei Xiong, Yu Zhang","doi":"10.1007/s11633-022-1379-3","DOIUrl":"https://doi.org/10.1007/s11633-022-1379-3","url":null,"abstract":"Instance segmentation has drawn mounting attention due to its significant utility. However, high computational costs have been widely acknowledged in this domain, as the instance mask is generally achieved by pixel-level labeling. In this paper, we present a conceptually efficient contour regression network based on the you only look once (YOLO) architecture named YOLO-CORE for instance segmentation. The mask of the instance is efficiently acquired by explicit and direct contour regression using our designed multi-order constraint consisting of a polar distance loss and a sector loss. Our proposed YOLO-CORE yields impressive segmentation performance in terms of both accuracy and speed. It achieves 57.9% AP@0.5 with 47 FPS (frames per second) on the semantic boundaries dataset (SBD) and 51.1% AP@0.5 with 46 FPS on the COCO dataset. The superior performance achieved by our method with explicit contour regression suggests a new technique line in the YOLO-based image understanding field. Moreover, our instance segmentation design can be flexibly integrated into existing deep detectors with negligible computation cost (65.86 BFLOPs (billion float operations per second) to 66.15 BFLOPs with the YOLOv3 detector).","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135396713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Practical Blind Image Denoising via Swin-Conv-UNet and Data Synthesis 基于swn - convn - unet和数据综合的实用盲图像去噪
4区 计算机科学 Pub Date : 2023-09-15 DOI: 10.1007/s11633-023-1466-0
Zhang, Kai, Li, Yawei, Liang, Jingyun, Cao, Jiezhang, Zhang, Yulun, Tang, Hao, Timofte, Radu, Van Gool, Luc
While recent years have witnessed a dramatic upsurge of exploiting deep neural networks toward solving image denoising, existing methods mostly rely on simple noise assumptions, such as additive white Gaussian noise (AWGN), JPEG compression noise and camera sensor noise, and a general-purpose blind denoising method for real images remains unsolved. In this paper, we attempt to solve this problem from the perspective of network architecture design and training data synthesis. Specifically, for the network architecture design, we propose a swin-conv block to incorporate the local modeling ability of residual convolutional layer and non-local modeling ability of swin transformer block, and then plug it as the main building block into the widely-used image-to-image translation UNet architecture. For the training data synthesis, we design a practical noise degradation model which takes into consideration different kinds of noise (including Gaussian, Poisson, speckle, JPEG compression, and processed camera sensor noises) and resizing, and also involves a random shuffle strategy and a double degradation strategy. Extensive experiments on AGWN removal and real image denoising demonstrate that the new network architecture design achieves state-of-the-art performance and the new degradation model can help to significantly improve the practicability. We believe our work can provide useful insights into current denoising research.
近年来,利用深度神经网络解决图像去噪问题的研究蓬勃发展,但现有的方法大多依赖于简单的噪声假设,如加性高斯白噪声(AWGN)、JPEG压缩噪声和相机传感器噪声,尚未解决实际图像的通用盲去噪方法。本文试图从网络架构设计和训练数据综合的角度来解决这一问题。具体来说,在网络架构设计上,我们提出了一种结合残差卷积层局部建模能力和swin变压器块非局部建模能力的swan -conv块,并将其作为主要构建块插入到广泛使用的图像到图像转换UNet体系结构中。对于训练数据的合成,我们设计了一个实用的噪声退化模型,该模型考虑了不同类型的噪声(包括高斯噪声、泊松噪声、散斑噪声、JPEG压缩噪声和处理后的相机传感器噪声)和大小调整,并涉及随机洗牌策略和双重退化策略。大量的AGWN去除和真实图像去噪实验表明,新的网络架构设计达到了最先进的性能,新的退化模型可以显著提高实用性。我们相信我们的工作可以为当前的去噪研究提供有用的见解。
{"title":"Practical Blind Image Denoising via Swin-Conv-UNet and Data Synthesis","authors":"Zhang, Kai, Li, Yawei, Liang, Jingyun, Cao, Jiezhang, Zhang, Yulun, Tang, Hao, Timofte, Radu, Van Gool, Luc","doi":"10.1007/s11633-023-1466-0","DOIUrl":"https://doi.org/10.1007/s11633-023-1466-0","url":null,"abstract":"While recent years have witnessed a dramatic upsurge of exploiting deep neural networks toward solving image denoising, existing methods mostly rely on simple noise assumptions, such as additive white Gaussian noise (AWGN), JPEG compression noise and camera sensor noise, and a general-purpose blind denoising method for real images remains unsolved. In this paper, we attempt to solve this problem from the perspective of network architecture design and training data synthesis. Specifically, for the network architecture design, we propose a swin-conv block to incorporate the local modeling ability of residual convolutional layer and non-local modeling ability of swin transformer block, and then plug it as the main building block into the widely-used image-to-image translation UNet architecture. For the training data synthesis, we design a practical noise degradation model which takes into consideration different kinds of noise (including Gaussian, Poisson, speckle, JPEG compression, and processed camera sensor noises) and resizing, and also involves a random shuffle strategy and a double degradation strategy. Extensive experiments on AGWN removal and real image denoising demonstrate that the new network architecture design achieves state-of-the-art performance and the new degradation model can help to significantly improve the practicability. We believe our work can provide useful insights into current denoising research.","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135353684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
DepthFormer: Exploiting Long-range Correlation and Local Information for Accurate Monocular Depth Estimation DepthFormer:利用远程相关和局部信息进行精确的单目深度估计
4区 计算机科学 Pub Date : 2023-09-13 DOI: 10.1007/s11633-023-1458-0
Zhenyu Li, Zehui Chen, Xianming Liu, Junjun Jiang
Abstract This paper aims to address the problem of supervised monocular depth estimation. We start with a meticulous pilot study to demonstrate that the long-range correlation is essential for accurate depth estimation. Moreover, the Transformer and convolution are good at long-range and close-range depth estimation, respectively. Therefore, we propose to adopt a parallel encoder architecture consisting of a Transformer branch and a convolution branch. The former can model global context with the effective attention mechanism and the latter aims to preserve the local information as the Transformer lacks the spatial inductive bias in modeling such contents. However, independent branches lead to a shortage of connections between features. To bridge this gap, we design a hierarchical aggregation and heterogeneous interaction module to enhance the Transformer features and model the affinity between the heterogeneous features in a set-to-set translation manner. Due to the unbearable memory cost introduced by the global attention on high-resolution feature maps, we adopt the deformable scheme to reduce the complexity. Extensive experiments on the KITTI, NYU, and SUN RGB-D datasets demonstrate that our proposed model, termed DepthFormer, surpasses state-of-the-art monocular depth estimation methods with prominent margins. The effectiveness of each proposed module is elaborately evaluated through meticulous and intensive ablation studies.
摘要本文旨在解决有监督的单目深度估计问题。我们从细致的试点研究开始,以证明远程相关性对于准确的深度估计是必不可少的。此外,Transformer和convolution分别擅长远程和近距离深度估计。因此,我们建议采用由变压器分支和卷积分支组成的并行编码器架构。前者可以利用有效的注意机制对全局上下文进行建模,而后者的目的是保留局部信息,因为Transformer在对这些内容建模时缺乏空间归纳偏差。但是,独立的分支导致功能之间缺乏连接。为了弥补这一差距,我们设计了一个分层聚合和异构交互模块来增强Transformer特征,并以集合到集合的转换方式对异构特征之间的亲和力进行建模。由于全球对高分辨率特征图的关注带来了难以承受的存储成本,我们采用了可变形方案来降低复杂性。在KITTI、NYU和SUN RGB-D数据集上进行的大量实验表明,我们提出的模型(称为DepthFormer)超越了最先进的单目深度估计方法,具有显著的边际。每个提出的模块的有效性是通过细致和密集的消融研究精心评估。
{"title":"DepthFormer: Exploiting Long-range Correlation and Local Information for Accurate Monocular Depth Estimation","authors":"Zhenyu Li, Zehui Chen, Xianming Liu, Junjun Jiang","doi":"10.1007/s11633-023-1458-0","DOIUrl":"https://doi.org/10.1007/s11633-023-1458-0","url":null,"abstract":"Abstract This paper aims to address the problem of supervised monocular depth estimation. We start with a meticulous pilot study to demonstrate that the long-range correlation is essential for accurate depth estimation. Moreover, the Transformer and convolution are good at long-range and close-range depth estimation, respectively. Therefore, we propose to adopt a parallel encoder architecture consisting of a Transformer branch and a convolution branch. The former can model global context with the effective attention mechanism and the latter aims to preserve the local information as the Transformer lacks the spatial inductive bias in modeling such contents. However, independent branches lead to a shortage of connections between features. To bridge this gap, we design a hierarchical aggregation and heterogeneous interaction module to enhance the Transformer features and model the affinity between the heterogeneous features in a set-to-set translation manner. Due to the unbearable memory cost introduced by the global attention on high-resolution feature maps, we adopt the deformable scheme to reduce the complexity. Extensive experiments on the KITTI, NYU, and SUN RGB-D datasets demonstrate that our proposed model, termed DepthFormer, surpasses state-of-the-art monocular depth estimation methods with prominent margins. The effectiveness of each proposed module is elaborately evaluated through meticulous and intensive ablation studies.","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134990420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 63
How Good is Google Bard’s Visual Understanding? An Empirical Study on Open Challenges b谷歌Bard的视觉理解有多好?开放性挑战的实证研究
4区 计算机科学 Pub Date : 2023-07-27 DOI: 10.1007/s11633-023-1469-x
Haotong Qin, Ge-Peng Ji, Salman Khan, Deng-Ping Fan, F. Khan, L. Gool
{"title":"How Good is Google Bard’s Visual Understanding? An Empirical Study on Open Challenges","authors":"Haotong Qin, Ge-Peng Ji, Salman Khan, Deng-Ping Fan, F. Khan, L. Gool","doi":"10.1007/s11633-023-1469-x","DOIUrl":"https://doi.org/10.1007/s11633-023-1469-x","url":null,"abstract":"","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44455389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Transformer: A General Framework from Machine Translation to Others Transformer:从机器翻译到其他翻译的通用框架
4区 计算机科学 Pub Date : 2023-06-02 DOI: 10.1007/s11633-022-1393-5
Yang Zhao, Jiajun Zhang, Chengqing Zong
{"title":"Transformer: A General Framework from Machine Translation to Others","authors":"Yang Zhao, Jiajun Zhang, Chengqing Zong","doi":"10.1007/s11633-022-1393-5","DOIUrl":"https://doi.org/10.1007/s11633-022-1393-5","url":null,"abstract":"","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41718075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Machine Learning Methods in Solving the Boolean Satisfiability Problem 解决布尔可满足性问题的机器学习方法
4区 计算机科学 Pub Date : 2023-06-01 DOI: 10.1007/s11633-022-1396-2
Wenxuan Guo, Hui-Ling Zhen, Xijun Li, Wanqian Luo, Mingxuan Yuan, Yaohui Jin, Junchi Yan
This paper reviews the recent literature on solving the Boolean satisfiability problem (SAT), an archetypal $$cal{N}cal{P}$$ -complete problem, with the aid of machine learning (ML) techniques. Over the last decade, the machine learning society advances rapidly and surpasses human performance on several tasks. This trend also inspires a number of works that apply machine learning methods for SAT solving. In this survey, we examine the evolving ML SAT solvers from naive classifiers with handcrafted features to emerging end-to-end SAT solvers, as well as recent progress on combinations of existing conflict-driven clause learning (CDCL) and local search solvers with machine learning methods. Overall, solving SAT with machine learning is a promising yet challenging research topic. We conclude the limitations of current works and suggest possible future directions. The collected paper list is available at https://github.com/Thinklab-SJTU/awesome-ml4co .
本文综述了利用机器学习技术解决布尔可满足性问题(SAT)的最新文献,这是一个典型的$$cal{N}cal{P}$$完全问题。在过去的十年里,机器学习社会发展迅速,在一些任务上超过了人类的表现。这一趋势也激发了许多将机器学习方法应用于SAT求解的作品。在本调查中,我们研究了不断发展的ML SAT解算器,从具有手工制作特征的朴素分类器到新兴的端到端SAT解算器,以及现有冲突驱动子句学习(CDCL)和局部搜索解算器与机器学习方法相结合的最新进展。总的来说,用机器学习解决SAT是一个有前途但具有挑战性的研究课题。我们总结了当前工作的局限性,并提出了可能的未来方向。收集的论文清单可在https://github.com/Thinklab-SJTU/awesome-ml4co上获得。
{"title":"Machine Learning Methods in Solving the Boolean Satisfiability Problem","authors":"Wenxuan Guo, Hui-Ling Zhen, Xijun Li, Wanqian Luo, Mingxuan Yuan, Yaohui Jin, Junchi Yan","doi":"10.1007/s11633-022-1396-2","DOIUrl":"https://doi.org/10.1007/s11633-022-1396-2","url":null,"abstract":"This paper reviews the recent literature on solving the Boolean satisfiability problem (SAT), an archetypal $$cal{N}cal{P}$$ -complete problem, with the aid of machine learning (ML) techniques. Over the last decade, the machine learning society advances rapidly and surpasses human performance on several tasks. This trend also inspires a number of works that apply machine learning methods for SAT solving. In this survey, we examine the evolving ML SAT solvers from naive classifiers with handcrafted features to emerging end-to-end SAT solvers, as well as recent progress on combinations of existing conflict-driven clause learning (CDCL) and local search solvers with machine learning methods. Overall, solving SAT with machine learning is a promising yet challenging research topic. We conclude the limitations of current works and suggest possible future directions. The collected paper list is available at https://github.com/Thinklab-SJTU/awesome-ml4co .","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135727200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Speech Emotion Recognition Using Cascaded Attention Network with Joint Loss for Discrimination of Confusions 基于联合损失级联注意网络的语音情绪识别
4区 计算机科学 Pub Date : 2023-06-01 DOI: 10.1007/s11633-022-1356-x
Yang Liu, Haoqin Sun, Wenbo Guan, Yu-xin Xia, Zhen Zhao
{"title":"Speech Emotion Recognition Using Cascaded Attention Network with Joint Loss for Discrimination of Confusions","authors":"Yang Liu, Haoqin Sun, Wenbo Guan, Yu-xin Xia, Zhen Zhao","doi":"10.1007/s11633-022-1356-x","DOIUrl":"https://doi.org/10.1007/s11633-022-1356-x","url":null,"abstract":"","PeriodicalId":29727,"journal":{"name":"Machine Intelligence Research","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42414786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Machine Intelligence Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1