首页 > 最新文献

Asian Conference on Machine Learning最新文献

英文 中文
Learning Practical Communication Strategies in Cooperative Multi-Agent Reinforcement Learning 协同多智能体强化学习中的实用沟通策略学习
Pub Date : 2022-09-02 DOI: 10.48550/arXiv.2209.01288
Diyi Hu, Chi Zhang, V. Prasanna, Bhaskar, Krishnamachari
In Multi-Agent Reinforcement Learning, communication is critical to encourage cooperation among agents. Communication in realistic wireless networks can be highly unreliable due to network conditions varying with agents' mobility, and stochasticity in the transmission process. We propose a framework to learn practical communication strategies by addressing three fundamental questions: (1) When: Agents learn the timing of communication based on not only message importance but also wireless channel conditions. (2) What: Agents augment message contents with wireless network measurements to better select the game and communication actions. (3) How: Agents use a novel neural message encoder to preserve all information from received messages, regardless of the number and order of messages. Simulating standard benchmarks under realistic wireless network settings, we show significant improvements in game performance, convergence speed and communication efficiency compared with state-of-the-art.
在多智能体强化学习中,沟通是促进智能体之间合作的关键。在现实无线网络中,由于网络条件随智能体的移动性和传输过程的随机性而变化,通信存在高度不可靠的问题。我们提出了一个框架,通过解决三个基本问题来学习实际的通信策略:(1)何时:智能体不仅根据消息的重要性,而且根据无线信道条件学习通信的时机。(2) What: agent通过无线网络测量来增强消息内容,以更好地选择游戏和通信动作。(3)方法:智能体使用一种新的神经信息编码器来保存接收到的信息中的所有信息,而不考虑消息的数量和顺序。在真实的无线网络设置下模拟标准基准,我们展示了与最先进的游戏性能,收敛速度和通信效率相比的显着改进。
{"title":"Learning Practical Communication Strategies in Cooperative Multi-Agent Reinforcement Learning","authors":"Diyi Hu, Chi Zhang, V. Prasanna, Bhaskar, Krishnamachari","doi":"10.48550/arXiv.2209.01288","DOIUrl":"https://doi.org/10.48550/arXiv.2209.01288","url":null,"abstract":"In Multi-Agent Reinforcement Learning, communication is critical to encourage cooperation among agents. Communication in realistic wireless networks can be highly unreliable due to network conditions varying with agents' mobility, and stochasticity in the transmission process. We propose a framework to learn practical communication strategies by addressing three fundamental questions: (1) When: Agents learn the timing of communication based on not only message importance but also wireless channel conditions. (2) What: Agents augment message contents with wireless network measurements to better select the game and communication actions. (3) How: Agents use a novel neural message encoder to preserve all information from received messages, regardless of the number and order of messages. Simulating standard benchmarks under realistic wireless network settings, we show significant improvements in game performance, convergence speed and communication efficiency compared with state-of-the-art.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133008170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sliced Wasserstein Variational Inference 切片Wasserstein变分推理
Pub Date : 2022-07-26 DOI: 10.48550/arXiv.2207.13177
Mingxuan Yi, Song Liu
Variational Inference approximates an unnormalized distribution via the minimization of Kullback-Leibler (KL) divergence. Although this divergence is efficient for computation and has been widely used in applications, it suffers from some unreasonable properties. For example, it is not a proper metric, i.e., it is non-symmetric and does not preserve the triangle inequality. On the other hand, optimal transport distances recently have shown some advantages over KL divergence. With the help of these advantages, we propose a new variational inference method by minimizing sliced Wasserstein distance, a valid metric arising from optimal transport. This sliced Wasserstein distance can be approximated simply by running MCMC but without solving any optimization problem. Our approximation also does not require a tractable density function of variational distributions so that approximating families can be amortized by generators like neural networks. Furthermore, we provide an analysis of the theoretical properties of our method. Experiments on synthetic and real data are illustrated to show the performance of the proposed method.
变分推理通过最小化Kullback-Leibler (KL)散度来近似非标准化分布。虽然这种散度计算效率高,在实际应用中得到了广泛的应用,但也存在一些不合理的性质。例如,它不是一个适当的度规,也就是说,它是非对称的,不保持三角形不等式。另一方面,最近最优运输距离比KL散度显示出一些优势。利用这些优点,我们提出了一种新的变分推理方法,通过最小化切片沃瑟斯坦距离,这是一个由最优传输产生的有效度量。这个Wasserstein距离可以简单地通过运行MCMC来近似,但不需要解决任何优化问题。我们的近似也不需要易处理的变分分布密度函数,因此近似族可以由神经网络之类的生成器平摊。此外,我们还对该方法的理论性质进行了分析。仿真实验和实际数据验证了该方法的有效性。
{"title":"Sliced Wasserstein Variational Inference","authors":"Mingxuan Yi, Song Liu","doi":"10.48550/arXiv.2207.13177","DOIUrl":"https://doi.org/10.48550/arXiv.2207.13177","url":null,"abstract":"Variational Inference approximates an unnormalized distribution via the minimization of Kullback-Leibler (KL) divergence. Although this divergence is efficient for computation and has been widely used in applications, it suffers from some unreasonable properties. For example, it is not a proper metric, i.e., it is non-symmetric and does not preserve the triangle inequality. On the other hand, optimal transport distances recently have shown some advantages over KL divergence. With the help of these advantages, we propose a new variational inference method by minimizing sliced Wasserstein distance, a valid metric arising from optimal transport. This sliced Wasserstein distance can be approximated simply by running MCMC but without solving any optimization problem. Our approximation also does not require a tractable density function of variational distributions so that approximating families can be amortized by generators like neural networks. Furthermore, we provide an analysis of the theoretical properties of our method. Experiments on synthetic and real data are illustrated to show the performance of the proposed method.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114565771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Domain Alignment Meets Fully Test-Time Adaptation 域对齐满足完全的测试时间适应
Pub Date : 2022-07-09 DOI: 10.48550/arXiv.2207.04185
Kowshik Thopalli, P. Turaga, Jayaraman J. Thiagarajan
A foundational requirement of a deployed ML model is to generalize to data drawn from a testing distribution that is different from training. A popular solution to this problem is to adapt a pre-trained model to novel domains using only unlabeled data. In this paper, we focus on a challenging variant of this problem, where access to the original source data is restricted. While fully test-time adaptation (FTTA) and unsupervised domain adaptation (UDA) are closely related, the advances in UDA are not readily applicable to TTA, since most UDA methods require access to the source data. Hence, we propose a new approach, CATTAn, that bridges UDA and FTTA, by relaxing the need to access entire source data, through a novel deep subspace alignment strategy. With a minimal overhead of storing the subspace basis set for the source data, CATTAn enables unsupervised alignment between source and target data during adaptation. Through extensive experimental evaluation on multiple 2D and 3D vision benchmarks (ImageNet-C, Office-31, OfficeHome, DomainNet, PointDA-10) and model architectures, we demonstrate significant gains in FTTA performance. Furthermore, we make a number of crucial findings on the utility of the alignment objective even with inherently robust models, pre-trained ViT representations and under low sample availability in the target domain.
部署的ML模型的一个基本要求是泛化到从不同于训练的测试分布中提取的数据。这个问题的一个流行的解决方案是只使用未标记的数据使预训练的模型适应新的领域。在本文中,我们将重点关注该问题的一个具有挑战性的变体,即对原始源数据的访问受到限制。虽然完全测试时自适应(FTTA)和无监督域自适应(UDA)密切相关,但UDA的进展并不容易适用于TTA,因为大多数UDA方法需要访问源数据。因此,我们提出了一种新的方法,CATTAn,它通过一种新的深子空间对齐策略,通过放松对整个源数据的访问需求,架起了UDA和FTTA的桥梁。由于存储源数据的子空间基集的开销最小,CATTAn可以在自适应期间实现源数据和目标数据之间的无监督对齐。通过对多种2D和3D视觉基准(ImageNet-C、Office-31、OfficeHome、DomainNet、PointDA-10)和模型架构进行广泛的实验评估,我们证明了FTTA性能的显著提高。此外,我们对对齐目标的效用做出了许多重要的发现,即使在固有的鲁棒模型,预训练的ViT表示和低样本可用性的目标域中。
{"title":"Domain Alignment Meets Fully Test-Time Adaptation","authors":"Kowshik Thopalli, P. Turaga, Jayaraman J. Thiagarajan","doi":"10.48550/arXiv.2207.04185","DOIUrl":"https://doi.org/10.48550/arXiv.2207.04185","url":null,"abstract":"A foundational requirement of a deployed ML model is to generalize to data drawn from a testing distribution that is different from training. A popular solution to this problem is to adapt a pre-trained model to novel domains using only unlabeled data. In this paper, we focus on a challenging variant of this problem, where access to the original source data is restricted. While fully test-time adaptation (FTTA) and unsupervised domain adaptation (UDA) are closely related, the advances in UDA are not readily applicable to TTA, since most UDA methods require access to the source data. Hence, we propose a new approach, CATTAn, that bridges UDA and FTTA, by relaxing the need to access entire source data, through a novel deep subspace alignment strategy. With a minimal overhead of storing the subspace basis set for the source data, CATTAn enables unsupervised alignment between source and target data during adaptation. Through extensive experimental evaluation on multiple 2D and 3D vision benchmarks (ImageNet-C, Office-31, OfficeHome, DomainNet, PointDA-10) and model architectures, we demonstrate significant gains in FTTA performance. Furthermore, we make a number of crucial findings on the utility of the alignment objective even with inherently robust models, pre-trained ViT representations and under low sample availability in the target domain.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123102536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
AS-IntroVAE: Adversarial Similarity Distance Makes Robust IntroVAE AS-IntroVAE:对抗性相似距离产生稳健的IntroVAE
Pub Date : 2022-06-28 DOI: 10.48550/arXiv.2206.13903
Chang-Tien Lu, Shen Zheng, Zirui Wang, O. Dib, Gaurav Gupta
Recently, introspective models like IntroVAE and S-IntroVAE have excelled in image generation and reconstruction tasks. The principal characteristic of introspective models is the adversarial learning of VAE, where the encoder attempts to distinguish between the real and the fake (i.e., synthesized) images. However, due to the unavailability of an effective metric to evaluate the difference between the real and the fake images, the posterior collapse and the vanishing gradient problem still exist, reducing the fidelity of the synthesized images. In this paper, we propose a new variation of IntroVAE called Adversarial Similarity Distance Introspective Variational Autoencoder (AS-IntroVAE). We theoretically analyze the vanishing gradient problem and construct a new Adversarial Similarity Distance (AS-Distance) using the 2-Wasserstein distance and the kernel trick. With weight annealing on AS-Distance and KL-Divergence, the AS-IntroVAE are able to generate stable and high-quality images. The posterior collapse problem is addressed by making per-batch attempts to transform the image so that it better fits the prior distribution in the latent space. Compared with the per-image approach, this strategy fosters more diverse distributions in the latent space, allowing our model to produce images of great diversity. Comprehensive experiments on benchmark datasets demonstrate the effectiveness of AS-IntroVAE on image generation and reconstruction tasks.
最近,IntroVAE和S-IntroVAE等内省模型在图像生成和重建任务中表现出色。内省模型的主要特征是VAE的对抗性学习,其中编码器试图区分真实和虚假(即合成)图像。然而,由于缺乏一种有效的度量来评估真假图像之间的差异,后验崩溃和梯度消失问题仍然存在,降低了合成图像的保真度。本文提出了一种新的内省变分自编码器——对抗相似距离内省变分自编码器(AS-IntroVAE)。我们从理论上分析了梯度消失问题,并利用2-Wasserstein距离和核技巧构造了一个新的对抗相似距离(AS-Distance)。通过对AS-Distance和KL-Divergence进行加权退火,AS-IntroVAE能够生成稳定的高质量图像。后验崩溃问题是通过逐批尝试变换图像,使其更好地适应潜在空间中的先验分布来解决的。与单图像方法相比,该策略在潜在空间中培养了更多样化的分布,使我们的模型能够产生具有极大多样性的图像。在基准数据集上的综合实验证明了AS-IntroVAE在图像生成和重建任务上的有效性。
{"title":"AS-IntroVAE: Adversarial Similarity Distance Makes Robust IntroVAE","authors":"Chang-Tien Lu, Shen Zheng, Zirui Wang, O. Dib, Gaurav Gupta","doi":"10.48550/arXiv.2206.13903","DOIUrl":"https://doi.org/10.48550/arXiv.2206.13903","url":null,"abstract":"Recently, introspective models like IntroVAE and S-IntroVAE have excelled in image generation and reconstruction tasks. The principal characteristic of introspective models is the adversarial learning of VAE, where the encoder attempts to distinguish between the real and the fake (i.e., synthesized) images. However, due to the unavailability of an effective metric to evaluate the difference between the real and the fake images, the posterior collapse and the vanishing gradient problem still exist, reducing the fidelity of the synthesized images. In this paper, we propose a new variation of IntroVAE called Adversarial Similarity Distance Introspective Variational Autoencoder (AS-IntroVAE). We theoretically analyze the vanishing gradient problem and construct a new Adversarial Similarity Distance (AS-Distance) using the 2-Wasserstein distance and the kernel trick. With weight annealing on AS-Distance and KL-Divergence, the AS-IntroVAE are able to generate stable and high-quality images. The posterior collapse problem is addressed by making per-batch attempts to transform the image so that it better fits the prior distribution in the latent space. Compared with the per-image approach, this strategy fosters more diverse distributions in the latent space, allowing our model to produce images of great diversity. Comprehensive experiments on benchmark datasets demonstrate the effectiveness of AS-IntroVAE on image generation and reconstruction tasks.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126557518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
FLVoogd: Robust And Privacy Preserving Federated Learning FLVoogd:鲁棒和隐私保护联邦学习
Pub Date : 2022-06-24 DOI: 10.48550/arXiv.2207.00428
Yuhang Tian, Rui Wang, Yan Qiao, E. Panaousis, K. Liang
In this work, we propose FLVoogd, an updated federated learning method in which servers and clients collaboratively eliminate Byzantine attacks while preserving privacy. In particular, servers use automatic Density-based Spatial Clustering of Applications with Noise (DBSCAN) combined with S2PC to cluster the benign majority without acquiring sensitive personal information. Meanwhile, clients build dual models and perform test-based distance controlling to adjust their local models toward the global one to achieve personalizing. Our framework is automatic and adaptive that servers/clients don't need to tune the parameters during the training. In addition, our framework leverages Secure Multi-party Computation (SMPC) operations, including multiplications, additions, and comparison, where costly operations, like division and square root, are not required. Evaluations are carried out on some conventional datasets from the image classification field. The result shows that FLVoogd can effectively reject malicious uploads in most scenarios; meanwhile, it avoids data leakage from the server-side.
在这项工作中,我们提出了FLVoogd,这是一种更新的联邦学习方法,其中服务器和客户端协同消除拜占庭攻击,同时保护隐私。特别是,服务器使用自动基于密度的噪声应用空间聚类(DBSCAN)与S2PC相结合,在不获取敏感个人信息的情况下对良性多数进行聚类。同时,客户建立双重模型,并进行基于测试的距离控制,将本地模型向全球模型调整,实现个性化。我们的框架是自动和自适应的,服务器/客户端不需要在训练期间调整参数。此外,我们的框架利用安全多方计算(SMPC)操作,包括乘法、加法和比较,这些操作不需要昂贵的操作,如除法和平方根。对图像分类领域的一些常规数据集进行了评价。结果表明,FLVoogd在大多数场景下都能有效地拒绝恶意上传;同时,避免了服务器端的数据泄露。
{"title":"FLVoogd: Robust And Privacy Preserving Federated Learning","authors":"Yuhang Tian, Rui Wang, Yan Qiao, E. Panaousis, K. Liang","doi":"10.48550/arXiv.2207.00428","DOIUrl":"https://doi.org/10.48550/arXiv.2207.00428","url":null,"abstract":"In this work, we propose FLVoogd, an updated federated learning method in which servers and clients collaboratively eliminate Byzantine attacks while preserving privacy. In particular, servers use automatic Density-based Spatial Clustering of Applications with Noise (DBSCAN) combined with S2PC to cluster the benign majority without acquiring sensitive personal information. Meanwhile, clients build dual models and perform test-based distance controlling to adjust their local models toward the global one to achieve personalizing. Our framework is automatic and adaptive that servers/clients don't need to tune the parameters during the training. In addition, our framework leverages Secure Multi-party Computation (SMPC) operations, including multiplications, additions, and comparison, where costly operations, like division and square root, are not required. Evaluations are carried out on some conventional datasets from the image classification field. The result shows that FLVoogd can effectively reject malicious uploads in most scenarios; meanwhile, it avoids data leakage from the server-side.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124371782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An online semi-definite programming with a generalised log-determinant regularizer and its applications 具有广义对数行列式正则化器的在线半确定规划及其应用
Pub Date : 2022-03-25 DOI: 10.3390/math10071055
Yaxiong Liu, Ken-ichiro Moridomi, Kohei Hatano, Eiji Takimoto
We consider a variant of the online semi-definite programming problem (OSDP). Specifically, in our problem, the setting of the decision space is a set of positive semi-definite matrices constrained by two norms in parallel: the L∞ norm to the diagonal entries and the Γ-trace norm, which is a generalized trace norm with a positive definite matrix Γ. Our setting recovers the original one when Γ is an identity matrix. To solve this problem, we design a follow-the-regularized-leader algorithm with a Γ-dependent regularizer, which also generalizes the log-determinant function. Next, we focus on online binary matrix completion (OBMC) with side information and online similarity prediction with side information. By reducing to the OSDP framework and applying our proposed algorithm, we remove the logarithmic factors in the previous mistake bound of the above two problems. In particular, for OBMC, our bound is optimal. Furthermore, our result implies a better offline generalization bound for the algorithm, which is similar to those of SVMs with the best kernel, if the side information is involved in advance.
我们考虑了在线半确定规划问题(OSDP)的一个变体。具体来说,在我们的问题中,决策空间的设置是由两个并行范数约束的正半定矩阵的集合:对角线项的L∞范数和Γ-trace范数,这是一个具有正定矩阵Γ的广义迹范数。当Γ是单位矩阵时,我们的设置恢复原来的设置。为了解决这个问题,我们设计了一个带Γ-dependent正则化器的跟随正则化领导者算法,该算法也推广了对数行列式函数。接下来,我们重点研究了基于边信息的在线二值矩阵补全(OBMC)和基于边信息的在线相似度预测。通过简化到OSDP框架并应用我们提出的算法,我们消除了上述两个问题的前一个错误界中的对数因子。特别是对于OBMC,我们的界是最优的。此外,我们的结果表明,该算法具有更好的离线泛化边界,类似于具有最佳核的支持向量机,如果提前涉及侧信息。
{"title":"An online semi-definite programming with a generalised log-determinant regularizer and its applications","authors":"Yaxiong Liu, Ken-ichiro Moridomi, Kohei Hatano, Eiji Takimoto","doi":"10.3390/math10071055","DOIUrl":"https://doi.org/10.3390/math10071055","url":null,"abstract":"We consider a variant of the online semi-definite programming problem (OSDP). Specifically, in our problem, the setting of the decision space is a set of positive semi-definite matrices constrained by two norms in parallel: the L∞ norm to the diagonal entries and the Γ-trace norm, which is a generalized trace norm with a positive definite matrix Γ. Our setting recovers the original one when Γ is an identity matrix. To solve this problem, we design a follow-the-regularized-leader algorithm with a Γ-dependent regularizer, which also generalizes the log-determinant function. Next, we focus on online binary matrix completion (OBMC) with side information and online similarity prediction with side information. By reducing to the OSDP framework and applying our proposed algorithm, we remove the logarithmic factors in the previous mistake bound of the above two problems. In particular, for OBMC, our bound is optimal. Furthermore, our result implies a better offline generalization bound for the algorithm, which is similar to those of SVMs with the best kernel, if the side information is involved in advance.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127411502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Detecting Accounting Frauds in Publicly Traded U.S. Firms: A Machine Learning Approach 美国上市公司的会计欺诈检测:机器学习方法
Pub Date : 2020-03-01 DOI: 10.2139/SSRN.2670703
Bin Li, Julia Yu, Jie Zhang, B. Ke
This paper studies how machine learning techniques can facilitate the detection of accounting fraud in publicly traded US rms. Existing studies often mimic human experts and employ the nancial or nonnancial
本文研究了机器学习技术如何促进美国上市公司会计欺诈的检测。现有的研究经常模仿人类专家,并采用金融或非金融手段
{"title":"Detecting Accounting Frauds in Publicly Traded U.S. Firms: A Machine Learning Approach","authors":"Bin Li, Julia Yu, Jie Zhang, B. Ke","doi":"10.2139/SSRN.2670703","DOIUrl":"https://doi.org/10.2139/SSRN.2670703","url":null,"abstract":"This paper studies how machine learning techniques can facilitate the detection of accounting fraud in publicly traded US rms. Existing studies often mimic human experts and employ the nancial or nonnancial","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130746043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Active Change-Point Detection 主动变更点检测
Pub Date : 2019-10-15 DOI: 10.1527/tjsai.35-5_e-ja10
S. Hayashi, Yoshinobu Kawahara, H. Kashima
We introduce Active Change-Point Detection (ACPD), a novel active learning problem for efficient change-point detection in situations where the cost of data acquisition is expensive. At each round of ACPD, the task is to adaptively determine the next input, in order to detect the change-point in a black-box expensive-to-evaluate function, with as few evaluations as possible. We propose a novel framework that can be generalized for different types of data and change-points, by utilizing an existing change-point detection method to compute change scores and a Bayesian optimization method to determine the next input. We demonstrate the efficiency of our proposed framework in different settings of datasets and change-points, using synthetic data and real-world data, such as material science data and seafloor depth data.
我们引入了主动变化点检测(ACPD),这是一种新的主动学习问题,用于在数据采集成本昂贵的情况下进行有效的变化点检测。在每一轮ACPD中,任务是自适应地确定下一个输入,以便用尽可能少的评估来检测黑盒中难以评估的函数中的变化点。我们提出了一个新的框架,可以推广到不同类型的数据和变化点,利用现有的变化点检测方法来计算变化分数和贝叶斯优化方法来确定下一个输入。我们使用合成数据和真实世界的数据,如材料科学数据和海底深度数据,证明了我们提出的框架在不同数据集和变化点设置下的效率。
{"title":"Active Change-Point Detection","authors":"S. Hayashi, Yoshinobu Kawahara, H. Kashima","doi":"10.1527/tjsai.35-5_e-ja10","DOIUrl":"https://doi.org/10.1527/tjsai.35-5_e-ja10","url":null,"abstract":"We introduce Active Change-Point Detection (ACPD), a novel active learning problem for efficient change-point detection in situations where the cost of data acquisition is expensive. At each round of ACPD, the task is to adaptively determine the next input, in order to detect the change-point in a black-box expensive-to-evaluate function, with as few evaluations as possible. We propose a novel framework that can be generalized for different types of data and change-points, by utilizing an existing change-point detection method to compute change scores and a Bayesian optimization method to determine the next input. We demonstrate the efficiency of our proposed framework in different settings of datasets and change-points, using synthetic data and real-world data, such as material science data and seafloor depth data.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120970584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Inverse Visual Question Answering with Multi-Level Attentions 多层次关注的逆向视觉问答
Pub Date : 2019-09-17 DOI: 10.22215/etd/2019-13929
Yaser Alwatter, Yuhong Guo
In this paper, we propose a novel deep multi-level attention model to address inverse visual question answering. The proposed model generates regional visual and semantic features at the object level and then enhances them with the answer cue by using attention mechanisms. Two levels of multiple attentions are employed in the model, including the dual attention at the partial question encoding step and the dynamic attention at the next question word generation step. We evaluate the proposed model on the VQA V1 dataset. It demonstrates state-of-the-art performance in terms of multiple commonly used metrics.
在本文中,我们提出了一种新的深度多层次注意模型来解决逆向视觉问答问题。该模型首先在对象层面生成区域视觉和语义特征,然后利用注意力机制对答案线索进行增强。该模型采用了两个层次的多重注意,包括部分问题编码步骤的双重注意和下一个问题词生成步骤的动态注意。我们在VQA V1数据集上评估了所提出的模型。它根据多个常用指标展示了最先进的性能。
{"title":"Inverse Visual Question Answering with Multi-Level Attentions","authors":"Yaser Alwatter, Yuhong Guo","doi":"10.22215/etd/2019-13929","DOIUrl":"https://doi.org/10.22215/etd/2019-13929","url":null,"abstract":"In this paper, we propose a novel deep multi-level attention model to address inverse visual question answering. The proposed model generates regional visual and semantic features at the object level and then enhances them with the answer cue by using attention mechanisms. Two levels of multiple attentions are employed in the model, including the dual attention at the partial question encoding step and the dynamic attention at the next question word generation step. We evaluate the proposed model on the VQA V1 dataset. It demonstrates state-of-the-art performance in terms of multiple commonly used metrics.","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2019-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116948454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Query Selection via Weighted Entropy in Graph-Based Semi-supervised Classification 基于加权熵的图半监督分类查询选择
Pub Date : 2009-11-03 DOI: 10.1007/978-3-642-05224-8_22
Krikamol Muandet, S. Marukatat, C. Nattee
{"title":"Query Selection via Weighted Entropy in Graph-Based Semi-supervised Classification","authors":"Krikamol Muandet, S. Marukatat, C. Nattee","doi":"10.1007/978-3-642-05224-8_22","DOIUrl":"https://doi.org/10.1007/978-3-642-05224-8_22","url":null,"abstract":"","PeriodicalId":119756,"journal":{"name":"Asian Conference on Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127293371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Asian Conference on Machine Learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1