首页 > 最新文献

ArXiv最新文献

英文 中文
Generative Representational Instruction Tuning 生成表征指令调整
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09906
Niklas Muennighoff, Hongjin Su, Liang Wang, Nan Yang, Furu Wei, Tao Yu, Amanpreet Singh, Douwe Kiela
All text-based language problems can be reduced to either generation or embedding. Current models only perform well at one or the other. We introduce generative representational instruction tuning (GRIT) whereby a large language model is trained to handle both generative and embedding tasks by distinguishing between them through instructions. Compared to other open models, our resulting GritLM 7B sets a new state of the art on the Massive Text Embedding Benchmark (MTEB) and outperforms all models up to its size on a range of generative tasks. By scaling up further, GritLM 8x7B outperforms all open generative language models that we tried while still being among the best embedding models. Notably, we find that GRIT matches training on only generative or embedding data, thus we can unify both at no performance loss. Among other benefits, the unification via GRIT speeds up Retrieval-Augmented Generation (RAG) by>60% for long documents, by no longer requiring separate retrieval and generation models. Models, code, etc. are freely available at https://github.com/ContextualAI/gritlm.
所有基于文本的语言问题都可以归结为生成或嵌入。目前的模型只能很好地处理其中一个问题。我们引入了生成表征指令调整(GRIT),通过指令区分生成和嵌入任务,训练大型语言模型同时处理生成和嵌入任务。与其他开放模型相比,我们的 GritLM 7B 在大规模文本嵌入基准(MTEB)上创造了新的技术水平,并在一系列生成任务中优于其规模的所有模型。通过进一步扩展,GritLM 8x7B 超越了我们尝试过的所有开放式生成语言模型,同时仍然是最好的嵌入模型之一。值得注意的是,我们发现 GRIT 只匹配生成数据或嵌入数据的训练,因此我们可以在不损失性能的情况下统一这两种数据。除其他优点外,通过 GRIT 实现统一后,不再需要单独的检索和生成模型,长文档的检索-增强生成(RAG)速度提高了 60%以上。模型、代码等可在 https://github.com/ContextualAI/gritlm 免费获取。
{"title":"Generative Representational Instruction Tuning","authors":"Niklas Muennighoff, Hongjin Su, Liang Wang, Nan Yang, Furu Wei, Tao Yu, Amanpreet Singh, Douwe Kiela","doi":"10.48550/arXiv.2402.09906","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09906","url":null,"abstract":"All text-based language problems can be reduced to either generation or embedding. Current models only perform well at one or the other. We introduce generative representational instruction tuning (GRIT) whereby a large language model is trained to handle both generative and embedding tasks by distinguishing between them through instructions. Compared to other open models, our resulting GritLM 7B sets a new state of the art on the Massive Text Embedding Benchmark (MTEB) and outperforms all models up to its size on a range of generative tasks. By scaling up further, GritLM 8x7B outperforms all open generative language models that we tried while still being among the best embedding models. Notably, we find that GRIT matches training on only generative or embedding data, thus we can unify both at no performance loss. Among other benefits, the unification via GRIT speeds up Retrieval-Augmented Generation (RAG) by>60% for long documents, by no longer requiring separate retrieval and generation models. Models, code, etc. are freely available at https://github.com/ContextualAI/gritlm.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"25 20","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Stage Algorithm for Group Testing with Prior Statistics 利用先验统计数据进行分组测试的多阶段算法
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.10018
Ayelet C. Portnoy, Alejandro Cohen
In this paper, we propose an efficient multi-stage algorithm for non-adaptive Group Testing (GT) with general correlated prior statistics. The proposed solution can be applied to any correlated statistical prior represented in trellis, e.g., finite state machines and Markov processes. We introduce a variation of List Viterbi Algorithm (LVA) to enable accurate recovery using much fewer tests than objectives, which efficiently gains from the correlated prior statistics structure. Our numerical results demonstrate that the proposed Multi-Stage GT (MSGT) algorithm can obtain the optimal Maximum A Posteriori (MAP) performance with feasible complexity in practical regimes, such as with COVID-19 and sparse signal recovery applications, and reduce in the scenarios tested the number of pooled tests by at least $25%$ compared to existing classical low complexity GT algorithms. Moreover, we analytically characterize the complexity of the proposed MSGT algorithm that guarantees its efficiency.
在本文中,我们提出了一种高效的多阶段算法,用于具有一般相关先验统计量的非自适应分组检验(GT)。所提出的解决方案可应用于任何以树状结构表示的相关统计先验,如有限状态机和马尔可夫过程。我们引入了列表维特比算法(LVA)的一种变体,使用比目标少得多的测试来实现精确恢复,从而有效地从相关先验统计结构中获益。我们的数值结果表明,所提出的多阶段 GT(MSGT)算法在实际应用中,如 COVID-19 和稀疏信号恢复应用中,能以可行的复杂度获得最佳的最大后验(MAP)性能,并且与现有的经典低复杂度 GT 算法相比,在所测试的场景中至少减少了 25%$ 的集合测试次数。此外,我们还分析了所提出的 MSGT 算法的复杂度特征,从而保证了其效率。
{"title":"Multi-Stage Algorithm for Group Testing with Prior Statistics","authors":"Ayelet C. Portnoy, Alejandro Cohen","doi":"10.48550/arXiv.2402.10018","DOIUrl":"https://doi.org/10.48550/arXiv.2402.10018","url":null,"abstract":"In this paper, we propose an efficient multi-stage algorithm for non-adaptive Group Testing (GT) with general correlated prior statistics. The proposed solution can be applied to any correlated statistical prior represented in trellis, e.g., finite state machines and Markov processes. We introduce a variation of List Viterbi Algorithm (LVA) to enable accurate recovery using much fewer tests than objectives, which efficiently gains from the correlated prior statistics structure. Our numerical results demonstrate that the proposed Multi-Stage GT (MSGT) algorithm can obtain the optimal Maximum A Posteriori (MAP) performance with feasible complexity in practical regimes, such as with COVID-19 and sparse signal recovery applications, and reduce in the scenarios tested the number of pooled tests by at least $25%$ compared to existing classical low complexity GT algorithms. Moreover, we analytically characterize the complexity of the proposed MSGT algorithm that guarantees its efficiency.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"22 26","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Not Just Novelty: A Longitudinal Study on Utility and Customization of AI Workflows 不仅仅是新奇:关于人工智能工作流程的实用性和定制化的纵向研究
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09894
Tao Long, Katy Ilonka Gero, Lydia B. Chilton
Generative AI brings novel and impressive abilities to help people in everyday tasks. There are many AI workflows that solve real and complex problems by chaining AI outputs together with human interaction. Although there is an undeniable lure of AI, it's uncertain how useful generative AI workflows are after the novelty wears off. Additionally, tools built with generative AI have the potential to be personalized and adapted quickly and easily, but do users take advantage of the potential to customize? We conducted a three-week longitudinal study with 12 users to understand the familiarization and customization of generative AI tools for science communication. Our study revealed that the familiarization phase lasts for 4.3 sessions, where users explore the capabilities of the workflow and which aspects they find useful. After familiarization, the perceived utility of the system is rated higher than before, indicating that the perceived utility of AI is not just a novelty effect. The increase in benefits mainly comes from end-users' ability to customize prompts, and thus appropriate the system to their own needs. This points to a future where generative AI systems can allow us to design for appropriation.
生成式人工智能为帮助人们完成日常任务带来了新颖而令人印象深刻的能力。有许多人工智能工作流程通过将人工智能输出与人机交互结合在一起,解决了实际而复杂的问题。虽然人工智能具有不可否认的诱惑力,但在新鲜感消失后,生成式人工智能工作流程的实用性如何还不确定。此外,使用生成式人工智能构建的工具具有快速、轻松地进行个性化调整的潜力,但用户是否会利用这种潜力进行定制呢?我们对 12 名用户进行了为期三周的纵向研究,以了解他们对用于科学传播的生成式人工智能工具的熟悉和定制情况。我们的研究表明,熟悉阶段持续了 4.3 个疗程,用户在这一阶段探索工作流程的功能以及他们认为哪些方面有用。熟悉之后,系统的感知效用评分高于熟悉之前,这表明人工智能的感知效用不仅仅是新奇效应。好处的增加主要来自最终用户定制提示的能力,从而使系统符合他们自己的需求。这预示着在未来,生成式人工智能系统可以让我们设计出适合自己的系统。
{"title":"Not Just Novelty: A Longitudinal Study on Utility and Customization of AI Workflows","authors":"Tao Long, Katy Ilonka Gero, Lydia B. Chilton","doi":"10.48550/arXiv.2402.09894","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09894","url":null,"abstract":"Generative AI brings novel and impressive abilities to help people in everyday tasks. There are many AI workflows that solve real and complex problems by chaining AI outputs together with human interaction. Although there is an undeniable lure of AI, it's uncertain how useful generative AI workflows are after the novelty wears off. Additionally, tools built with generative AI have the potential to be personalized and adapted quickly and easily, but do users take advantage of the potential to customize? We conducted a three-week longitudinal study with 12 users to understand the familiarization and customization of generative AI tools for science communication. Our study revealed that the familiarization phase lasts for 4.3 sessions, where users explore the capabilities of the workflow and which aspects they find useful. After familiarization, the perceived utility of the system is rated higher than before, indicating that the perceived utility of AI is not just a novelty effect. The increase in benefits mainly comes from end-users' ability to customize prompts, and thus appropriate the system to their own needs. This points to a future where generative AI systems can allow us to design for appropriation.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"29 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
System-level Impact of Non-Ideal Program-Time of Charge Trap Flash (CTF) on Deep Neural Network 电荷陷阱闪存 (CTF) 非理想编程时间对深度神经网络的系统级影响
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09792
S. Shrivastava, A. Biswas, S. Chakrabarty, G. Dash, V. Saraswat, U. Ganguly
Learning of deep neural networks (DNN) using Resistive Processing Unit (RPU) architecture is energy-efficient as it utilizes dedicated neuromorphic hardware and stochastic computation of weight updates for in-memory computing. Charge Trap Flash (CTF) devices can implement RPU-based weight updates in DNNs. However, prior work has shown that the weight updates (V_T) in CTF-based RPU are impacted by the non-ideal program time of CTF. The non-ideal program time is affected by two factors of CTF. Firstly, the effects of the number of input pulses (N) or pulse width (pw), and secondly, the gap between successive update pulses (t_gap) used for the stochastic computation of weight updates. Therefore, the impact of this non-ideal program time must be studied for neural network training simulations. In this study, Firstly, we propose a pulse-train design compensation technique to reduce the total error caused by non-ideal program time of CTF and stochastic variance of a network. Secondly, we simulate RPU-based DNN with non-ideal program time of CTF on MNIST and Fashion-MNIST datasets. We find that for larger N (~1000), learning performance approaches the ideal (software-level) training level and, therefore, is not much impacted by the choice of t_gap used to implement RPU-based weight updates. However, for lower N (<500), learning performance depends on T_gap of the pulses. Finally, we also performed an ablation study to isolate the causal factor of the improved learning performance. We conclude that the lower noise level in the weight updates is the most likely significant factor to improve the learning performance of DNN. Thus, our study attempts to compensate for the error caused by non-ideal program time and standardize the pulse length (N) and pulse gap (t_gap) specifications for CTF-based RPUs for accurate system-level on-chip training.
使用电阻式处理单元(RPU)架构学习深度神经网络(DNN)非常节能,因为它利用了专用的神经形态硬件和随机计算权重更新的内存计算。电荷陷阱闪存(CTF)设备可以在 DNN 中实现基于 RPU 的权重更新。然而,先前的研究表明,基于 CTF 的 RPU 中的权重更新(V_T)会受到 CTF 非理想编程时间的影响。非理想程序时间受 CTF 的两个因素影响。首先是输入脉冲数(N)或脉冲宽度(pw)的影响,其次是用于随机计算权重更新的连续更新脉冲之间的间隙(t_gap)。因此,必须研究这种非理想程序时间对神经网络训练模拟的影响。在本研究中,首先,我们提出了一种脉冲-训练设计补偿技术,以减少 CTF 非理想程序时间和网络随机方差造成的总误差。其次,我们在 MNIST 和 Fashion-MNIST 数据集上模拟了基于 RPU 的 DNN 与 CTF 的非理想编程时间。我们发现,对于较大的 N(约 1000),学习性能接近理想的(软件级)训练水平,因此,用于实现基于 RPU 的权重更新的 t_gap 选择不会对学习性能产生太大影响。然而,对于较低的 N(<500),学习性能取决于脉冲的 T_gap。最后,我们还进行了一项消融研究,以找出学习性能提高的原因。我们得出的结论是,权值更新中较低的噪声水平最有可能是提高 DNN 学习性能的重要因素。因此,我们的研究试图弥补非理想程序时间造成的误差,并对基于 CTF 的 RPU 的脉冲长度(N)和脉冲间隙(t_gap)规格进行标准化,以实现精确的系统级片上训练。
{"title":"System-level Impact of Non-Ideal Program-Time of Charge Trap Flash (CTF) on Deep Neural Network","authors":"S. Shrivastava, A. Biswas, S. Chakrabarty, G. Dash, V. Saraswat, U. Ganguly","doi":"10.48550/arXiv.2402.09792","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09792","url":null,"abstract":"Learning of deep neural networks (DNN) using Resistive Processing Unit (RPU) architecture is energy-efficient as it utilizes dedicated neuromorphic hardware and stochastic computation of weight updates for in-memory computing. Charge Trap Flash (CTF) devices can implement RPU-based weight updates in DNNs. However, prior work has shown that the weight updates (V_T) in CTF-based RPU are impacted by the non-ideal program time of CTF. The non-ideal program time is affected by two factors of CTF. Firstly, the effects of the number of input pulses (N) or pulse width (pw), and secondly, the gap between successive update pulses (t_gap) used for the stochastic computation of weight updates. Therefore, the impact of this non-ideal program time must be studied for neural network training simulations. In this study, Firstly, we propose a pulse-train design compensation technique to reduce the total error caused by non-ideal program time of CTF and stochastic variance of a network. Secondly, we simulate RPU-based DNN with non-ideal program time of CTF on MNIST and Fashion-MNIST datasets. We find that for larger N (~1000), learning performance approaches the ideal (software-level) training level and, therefore, is not much impacted by the choice of t_gap used to implement RPU-based weight updates. However, for lower N (<500), learning performance depends on T_gap of the pulses. Finally, we also performed an ablation study to isolate the causal factor of the improved learning performance. We conclude that the lower noise level in the weight updates is the most likely significant factor to improve the learning performance of DNN. Thus, our study attempts to compensate for the error caused by non-ideal program time and standardize the pulse length (N) and pulse gap (t_gap) specifications for CTF-based RPUs for accurate system-level on-chip training.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"20 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TOAD: Task-Oriented Automatic Dialogs with Diverse Response Styles TOAD:以任务为导向、响应风格多样的自动对话框
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.10137
Yinhong Liu, Yimai Fang, David Vandyke, Nigel Collier
In light of recent advances in large language models (LLMs), the expectations for the next generation of virtual assistants include enhanced naturalness and adaptability across diverse usage scenarios. However, the creation of high-quality annotated data for Task-Oriented Dialog (TOD) is recognized to be slow and costly. To address these challenges, we introduce Task-Oriented Automatic Dialogs (TOAD), a novel and scalable TOD dataset along with its automatic generation pipeline. The TOAD dataset simulates realistic app context interaction and provide a variety of system response style options. Two aspects of system response styles are considered, verbosity level and users' expression mirroring. We benchmark TOAD on two response generation tasks and the results show that modelling more verbose or responses without user expression mirroring is more challenging.
鉴于最近在大型语言模型(LLMs)方面取得的进展,人们对下一代虚拟助手的期望包括在各种使用场景中增强自然性和适应性。然而,为面向任务的对话(TOD)创建高质量的注释数据被认为是缓慢而昂贵的。为了应对这些挑战,我们推出了任务导向自动对话(TOAD)--一种新颖且可扩展的 TOD 数据集及其自动生成管道。TOAD 数据集模拟了真实的应用程序上下文交互,并提供了多种系统响应风格选项。我们考虑了系统响应风格的两个方面,即冗长程度和用户表达镜像。我们在两个响应生成任务中对 TOAD 进行了基准测试,结果表明,在没有用户表情镜像的情况下模拟更多的冗长响应或响应更具挑战性。
{"title":"TOAD: Task-Oriented Automatic Dialogs with Diverse Response Styles","authors":"Yinhong Liu, Yimai Fang, David Vandyke, Nigel Collier","doi":"10.48550/arXiv.2402.10137","DOIUrl":"https://doi.org/10.48550/arXiv.2402.10137","url":null,"abstract":"In light of recent advances in large language models (LLMs), the expectations for the next generation of virtual assistants include enhanced naturalness and adaptability across diverse usage scenarios. However, the creation of high-quality annotated data for Task-Oriented Dialog (TOD) is recognized to be slow and costly. To address these challenges, we introduce Task-Oriented Automatic Dialogs (TOAD), a novel and scalable TOD dataset along with its automatic generation pipeline. The TOAD dataset simulates realistic app context interaction and provide a variety of system response style options. Two aspects of system response styles are considered, verbosity level and users' expression mirroring. We benchmark TOAD on two response generation tasks and the results show that modelling more verbose or responses without user expression mirroring is more challenging.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"16 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
User Privacy Harms and Risks in Conversational AI: A Proposed Framework 人工智能对话中的用户隐私危害与风险:一个拟议框架
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09716
Ece Gumusel, Kyrie Zhixuan Zhou, M. Sanfilippo
This study presents a unique framework that applies and extends Solove (2006)'s taxonomy to address privacy concerns in interactions with text-based AI chatbots. As chatbot prevalence grows, concerns about user privacy have heightened. While existing literature highlights design elements compromising privacy, a comprehensive framework is lacking. Through semi-structured interviews with 13 participants interacting with two AI chatbots, this study identifies 9 privacy harms and 9 privacy risks in text-based interactions. Using a grounded theory approach for interview and chatlog analysis, the framework examines privacy implications at various interaction stages. The aim is to offer developers, policymakers, and researchers a tool for responsible and secure implementation of conversational AI, filling the existing gap in addressing privacy issues associated with text-based AI chatbots.
本研究提出了一个独特的框架,应用并扩展了 Solove(2006 年)的分类法,以解决与基于文本的人工智能聊天机器人交互过程中的隐私问题。随着聊天机器人的普及,人们对用户隐私的担忧也随之增加。虽然现有文献强调了损害隐私的设计要素,但缺乏一个全面的框架。本研究通过对与两个人工智能聊天机器人互动的 13 名参与者进行半结构式访谈,确定了基于文本的互动中的 9 种隐私危害和 9 种隐私风险。该框架采用基础理论方法进行访谈和聊天记录分析,研究了不同交互阶段的隐私影响。其目的是为开发人员、决策者和研究人员提供一个负责任地、安全地实施人工智能对话的工具,填补在解决与基于文本的人工智能聊天机器人相关的隐私问题方面的现有空白。
{"title":"User Privacy Harms and Risks in Conversational AI: A Proposed Framework","authors":"Ece Gumusel, Kyrie Zhixuan Zhou, M. Sanfilippo","doi":"10.48550/arXiv.2402.09716","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09716","url":null,"abstract":"This study presents a unique framework that applies and extends Solove (2006)'s taxonomy to address privacy concerns in interactions with text-based AI chatbots. As chatbot prevalence grows, concerns about user privacy have heightened. While existing literature highlights design elements compromising privacy, a comprehensive framework is lacking. Through semi-structured interviews with 13 participants interacting with two AI chatbots, this study identifies 9 privacy harms and 9 privacy risks in text-based interactions. Using a grounded theory approach for interview and chatlog analysis, the framework examines privacy implications at various interaction stages. The aim is to offer developers, policymakers, and researchers a tool for responsible and secure implementation of conversational AI, filling the existing gap in addressing privacy issues associated with text-based AI chatbots.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"13 11","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Systematic Evaluation of Evolving Highly Nonlinear Boolean Functions in Odd Sizes 对奇数大小高度非线性布尔函数进化的系统评估
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09937
C. Carlet, Marko Ðurasevic, D. Jakobović, S. Picek, L. Mariot
Boolean functions are mathematical objects used in diverse applications. Different applications also have different requirements, making the research on Boolean functions very active. In the last 30 years, evolutionary algorithms have been shown to be a strong option for evolving Boolean functions in different sizes and with different properties. Still, most of those works consider similar settings and provide results that are mostly interesting from the evolutionary algorithm's perspective. This work considers the problem of evolving highly nonlinear Boolean functions in odd sizes. While the problem formulation sounds simple, the problem is remarkably difficult, and the related work is extremely scarce. We consider three solutions encodings and four Boolean function sizes and run a detailed experimental analysis. Our results show that the problem is challenging, and finding optimal solutions is impossible except for the smallest tested size. However, once we added local search to the evolutionary algorithm, we managed to find a Boolean function in nine inputs with nonlinearity 241, which, to our knowledge, had never been accomplished before with evolutionary algorithms.
布尔函数是用于各种应用的数学对象。不同的应用也有不同的要求,因此布尔函数的研究非常活跃。在过去的 30 年中,进化算法已被证明是进化不同大小和不同性质的布尔函数的有力选择。不过,这些研究大多考虑的是类似的设置,提供的结果大多是从进化算法的角度来看比较有趣的。这项工作考虑的是奇数大小的高度非线性布尔函数的进化问题。虽然问题的表述听起来很简单,但这个问题却非常困难,相关的工作也非常少。我们考虑了三种解决方案编码和四种布尔函数大小,并进行了详细的实验分析。我们的结果表明,这个问题极具挑战性,除了最小的测试规模外,找到最优解是不可能的。然而,一旦我们在进化算法中加入局部搜索,我们就能在九个输入中找到一个布尔函数,非线性241,据我们所知,这是进化算法从未实现过的。
{"title":"A Systematic Evaluation of Evolving Highly Nonlinear Boolean Functions in Odd Sizes","authors":"C. Carlet, Marko Ðurasevic, D. Jakobović, S. Picek, L. Mariot","doi":"10.48550/arXiv.2402.09937","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09937","url":null,"abstract":"Boolean functions are mathematical objects used in diverse applications. Different applications also have different requirements, making the research on Boolean functions very active. In the last 30 years, evolutionary algorithms have been shown to be a strong option for evolving Boolean functions in different sizes and with different properties. Still, most of those works consider similar settings and provide results that are mostly interesting from the evolutionary algorithm's perspective. This work considers the problem of evolving highly nonlinear Boolean functions in odd sizes. While the problem formulation sounds simple, the problem is remarkably difficult, and the related work is extremely scarce. We consider three solutions encodings and four Boolean function sizes and run a detailed experimental analysis. Our results show that the problem is challenging, and finding optimal solutions is impossible except for the smallest tested size. However, once we added local search to the evolutionary algorithm, we managed to find a Boolean function in nine inputs with nonlinearity 241, which, to our knowledge, had never been accomplished before with evolutionary algorithms.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"20 17","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploiting Alpha Transparency In Language And Vision-Based AI Systems 在基于语言和视觉的人工智能系统中利用阿尔法透明度
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09671
David A. Noever, Forrest McKee
This investigation reveals a novel exploit derived from PNG image file formats, specifically their alpha transparency layer, and its potential to fool multiple AI vision systems. Our method uses this alpha layer as a clandestine channel invisible to human observers but fully actionable by AI image processors. The scope tested for the vulnerability spans representative vision systems from Apple, Microsoft, Google, Salesforce, Nvidia, and Facebook, highlighting the attack's potential breadth. This vulnerability challenges the security protocols of existing and fielded vision systems, from medical imaging to autonomous driving technologies. Our experiments demonstrate that the affected systems, which rely on convolutional neural networks or the latest multimodal language models, cannot quickly mitigate these vulnerabilities through simple patches or updates. Instead, they require retraining and architectural changes, indicating a persistent hole in multimodal technologies without some future adversarial hardening against such vision-language exploits.
这项研究揭示了一种源自 PNG 图像文件格式(特别是其阿尔法透明层)的新型漏洞利用方法,及其欺骗多种人工智能视觉系统的潜力。我们的方法将阿尔法层用作人类观察者看不到、但人工智能图像处理器完全可以操作的秘密通道。该漏洞的测试范围涵盖苹果、微软、谷歌、Salesforce、Nvidia 和 Facebook 等公司的代表性视觉系统,凸显了攻击的潜在广度。从医疗成像到自动驾驶技术,该漏洞对现有和已投入使用的视觉系统的安全协议提出了挑战。我们的实验表明,依赖卷积神经网络或最新多模态语言模型的受影响系统无法通过简单的补丁或更新快速缓解这些漏洞。相反,它们需要重新训练和改变架构,这表明,如果未来不针对此类视觉语言漏洞进行对抗性加固,多模态技术中的漏洞将长期存在。
{"title":"Exploiting Alpha Transparency In Language And Vision-Based AI Systems","authors":"David A. Noever, Forrest McKee","doi":"10.48550/arXiv.2402.09671","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09671","url":null,"abstract":"This investigation reveals a novel exploit derived from PNG image file formats, specifically their alpha transparency layer, and its potential to fool multiple AI vision systems. Our method uses this alpha layer as a clandestine channel invisible to human observers but fully actionable by AI image processors. The scope tested for the vulnerability spans representative vision systems from Apple, Microsoft, Google, Salesforce, Nvidia, and Facebook, highlighting the attack's potential breadth. This vulnerability challenges the security protocols of existing and fielded vision systems, from medical imaging to autonomous driving technologies. Our experiments demonstrate that the affected systems, which rely on convolutional neural networks or the latest multimodal language models, cannot quickly mitigate these vulnerabilities through simple patches or updates. Instead, they require retraining and architectural changes, indicating a persistent hole in multimodal technologies without some future adversarial hardening against such vision-language exploits.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"17 15","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
User Modeling and User Profiling: A Comprehensive Survey 用户建模和用户分析:全面调查
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.09660
Erasmo Purificato, Ludovico Boratto, Ernesto William De Luca
The integration of artificial intelligence (AI) into daily life, particularly through information retrieval and recommender systems, has necessitated advanced user modeling and profiling techniques to deliver personalized experiences. These techniques aim to construct accurate user representations based on the rich amounts of data generated through interactions with these systems. This paper presents a comprehensive survey of the current state, evolution, and future directions of user modeling and profiling research. We provide a historical overview, tracing the development from early stereotype models to the latest deep learning techniques, and propose a novel taxonomy that encompasses all active topics in this research area, including recent trends. Our survey highlights the paradigm shifts towards more sophisticated user profiling methods, emphasizing implicit data collection, multi-behavior modeling, and the integration of graph data structures. We also address the critical need for privacy-preserving techniques and the push towards explainability and fairness in user modeling approaches. By examining the definitions of core terminology, we aim to clarify ambiguities and foster a clearer understanding of the field by proposing two novel encyclopedic definitions of the main terms. Furthermore, we explore the application of user modeling in various domains, such as fake news detection, cybersecurity, and personalized education. This survey serves as a comprehensive resource for researchers and practitioners, offering insights into the evolution of user modeling and profiling and guiding the development of more personalized, ethical, and effective AI systems.
人工智能(AI)与日常生活的融合,特别是通过信息检索和推荐系统,需要先进的用户建模和分析技术来提供个性化体验。这些技术旨在根据与这些系统交互时产生的大量数据构建准确的用户表征。本文全面介绍了用户建模和特征分析研究的现状、发展和未来方向。我们提供了一个历史概述,追溯了从早期定型模型到最新深度学习技术的发展历程,并提出了一个新颖的分类法,涵盖了该研究领域的所有活跃课题,包括最新趋势。我们的调查突出了向更复杂的用户剖析方法的范式转变,强调了隐式数据收集、多行为建模和图数据结构的整合。我们还讨论了对隐私保护技术的迫切需求,以及在用户建模方法中对可解释性和公平性的推动。通过研究核心术语的定义,我们提出了两个新颖的百科全书式的主要术语定义,旨在澄清歧义,促进对该领域更清晰的理解。此外,我们还探讨了用户建模在假新闻检测、网络安全和个性化教育等不同领域的应用。这份调查报告为研究人员和从业人员提供了全面的资源,让他们深入了解用户建模和用户画像的演变,并为开发更个性化、更道德、更有效的人工智能系统提供指导。
{"title":"User Modeling and User Profiling: A Comprehensive Survey","authors":"Erasmo Purificato, Ludovico Boratto, Ernesto William De Luca","doi":"10.48550/arXiv.2402.09660","DOIUrl":"https://doi.org/10.48550/arXiv.2402.09660","url":null,"abstract":"The integration of artificial intelligence (AI) into daily life, particularly through information retrieval and recommender systems, has necessitated advanced user modeling and profiling techniques to deliver personalized experiences. These techniques aim to construct accurate user representations based on the rich amounts of data generated through interactions with these systems. This paper presents a comprehensive survey of the current state, evolution, and future directions of user modeling and profiling research. We provide a historical overview, tracing the development from early stereotype models to the latest deep learning techniques, and propose a novel taxonomy that encompasses all active topics in this research area, including recent trends. Our survey highlights the paradigm shifts towards more sophisticated user profiling methods, emphasizing implicit data collection, multi-behavior modeling, and the integration of graph data structures. We also address the critical need for privacy-preserving techniques and the push towards explainability and fairness in user modeling approaches. By examining the definitions of core terminology, we aim to clarify ambiguities and foster a clearer understanding of the field by proposing two novel encyclopedic definitions of the main terms. Furthermore, we explore the application of user modeling in various domains, such as fake news detection, cybersecurity, and personalized education. This survey serves as a comprehensive resource for researchers and practitioners, offering insights into the evolution of user modeling and profiling and guiding the development of more personalized, ethical, and effective AI systems.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"15 17","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Is Continual Learning Ready for Real-world Challenges? 持续学习能否应对现实世界的挑战?
Pub Date : 2024-02-15 DOI: 10.48550/arXiv.2402.10130
Theodora Kontogianni, Yuanwen Yue, Siyu Tang, Konrad Schindler
Despite continual learning's long and well-established academic history, its application in real-world scenarios remains rather limited. This paper contends that this gap is attributable to a misalignment between the actual challenges of continual learning and the evaluation protocols in use, rendering proposed solutions ineffective for addressing the complexities of real-world setups. We validate our hypothesis and assess progress to date, using a new 3D semantic segmentation benchmark, OCL-3DSS. We investigate various continual learning schemes from the literature by utilizing more realistic protocols that necessitate online and continual learning for dynamic, real-world scenarios (eg., in robotics and 3D vision applications). The outcomes are sobering: all considered methods perform poorly, significantly deviating from the upper bound of joint offline training. This raises questions about the applicability of existing methods in realistic settings. Our paper aims to initiate a paradigm shift, advocating for the adoption of continual learning methods through new experimental protocols that better emulate real-world conditions to facilitate breakthroughs in the field.
尽管持续学习在学术界有着悠久的历史和良好的口碑,但其在现实世界中的应用仍然相当有限。本文认为,造成这一差距的原因是持续学习的实际挑战与使用中的评估协议之间存在偏差,导致提出的解决方案无法有效解决现实世界中的复杂问题。我们利用新的三维语义分割基准 OCL-3DSS 验证了我们的假设,并评估了迄今为止取得的进展。我们利用更现实的协议来研究文献中的各种持续学习方案,这些协议要求在动态的真实世界场景(如机器人和三维视觉应用)中进行在线持续学习。结果令人警醒:所有考虑的方法都表现不佳,明显偏离了联合离线训练的上限。这就对现有方法在现实环境中的适用性提出了质疑。我们的论文旨在启动范式转变,倡导通过新的实验协议采用持续学习方法,更好地模拟现实世界的条件,以促进该领域的突破。
{"title":"Is Continual Learning Ready for Real-world Challenges?","authors":"Theodora Kontogianni, Yuanwen Yue, Siyu Tang, Konrad Schindler","doi":"10.48550/arXiv.2402.10130","DOIUrl":"https://doi.org/10.48550/arXiv.2402.10130","url":null,"abstract":"Despite continual learning's long and well-established academic history, its application in real-world scenarios remains rather limited. This paper contends that this gap is attributable to a misalignment between the actual challenges of continual learning and the evaluation protocols in use, rendering proposed solutions ineffective for addressing the complexities of real-world setups. We validate our hypothesis and assess progress to date, using a new 3D semantic segmentation benchmark, OCL-3DSS. We investigate various continual learning schemes from the literature by utilizing more realistic protocols that necessitate online and continual learning for dynamic, real-world scenarios (eg., in robotics and 3D vision applications). The outcomes are sobering: all considered methods perform poorly, significantly deviating from the upper bound of joint offline training. This raises questions about the applicability of existing methods in realistic settings. Our paper aims to initiate a paradigm shift, advocating for the adoption of continual learning methods through new experimental protocols that better emulate real-world conditions to facilitate breakthroughs in the field.","PeriodicalId":8425,"journal":{"name":"ArXiv","volume":"14 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139962691","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ArXiv
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1