Optimal word order for non-causal text generation with Large Language Models: The Spanish case

IF 3.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pattern Recognition Letters Pub Date : 2025-04-01 Epub Date: 2025-02-17 DOI:10.1016/j.patrec.2025.02.010
Andrea Busto-Castiñeira, Silvia García-Méndez, Francisco de Arriba-Pérez, Francisco J. González-Castaño
{"title":"Optimal word order for non-causal text generation with Large Language Models: The Spanish case","authors":"Andrea Busto-Castiñeira,&nbsp;Silvia García-Méndez,&nbsp;Francisco de Arriba-Pérez,&nbsp;Francisco J. González-Castaño","doi":"10.1016/j.patrec.2025.02.010","DOIUrl":null,"url":null,"abstract":"<div><div>Natural Language Generation (<span>nlg</span>) popularity has increased owing to the progress in Large Language Models (<span>llm</span>s), with zero-shot inference capabilities. However, most neural systems utilize decoder-only causal (unidirectional) transformer models, which are effective for English but may reduce the richness of languages with less strict word order, subject omission, or different relative clause attachment preferences. This is the first work that analytically addresses optimal text generation order for non-causal language models. We present a novel Viterbi algorithm-based methodology for maximum likelihood word order estimation. We analyze the non-causal most-likelihood order probability for <span>nlg</span> in Spanish and, then, the probability of generating the same phrases with Spanish causal <span>nlg</span>. This comparative analysis reveals that causal <span>nlg</span> prefers English-like <span>svo</span> structures. We also analyze the relationship between optimal generation order and causal left-to-right generation order using Spearman’s rank correlation. Our results demonstrate that the ideal order predicted by the maximum likelihood estimator is not closely related to the causal order and may be influenced by the syntactic structure of the target sentence.</div></div>","PeriodicalId":54638,"journal":{"name":"Pattern Recognition Letters","volume":"190 ","pages":"Pages 89-96"},"PeriodicalIF":3.3000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition Letters","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167865525000546","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/17 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Natural Language Generation (nlg) popularity has increased owing to the progress in Large Language Models (llms), with zero-shot inference capabilities. However, most neural systems utilize decoder-only causal (unidirectional) transformer models, which are effective for English but may reduce the richness of languages with less strict word order, subject omission, or different relative clause attachment preferences. This is the first work that analytically addresses optimal text generation order for non-causal language models. We present a novel Viterbi algorithm-based methodology for maximum likelihood word order estimation. We analyze the non-causal most-likelihood order probability for nlg in Spanish and, then, the probability of generating the same phrases with Spanish causal nlg. This comparative analysis reveals that causal nlg prefers English-like svo structures. We also analyze the relationship between optimal generation order and causal left-to-right generation order using Spearman’s rank correlation. Our results demonstrate that the ideal order predicted by the maximum likelihood estimator is not closely related to the causal order and may be influenced by the syntactic structure of the target sentence.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
大型语言模型非因果文本生成的最佳词序:西班牙语案例
由于具有零概率推理能力的大型语言模型(llms)的进步,自然语言生成(nlg)越来越受欢迎。然而,大多数神经系统使用仅解码的因果(单向)转换模型,这对英语是有效的,但可能会减少语言的丰富程度,因为语序不太严格,主语省略或不同的关系从句依恋偏好。这是第一个分析解决非因果语言模型的最佳文本生成顺序的工作。我们提出了一种新的基于Viterbi算法的最大似然词序估计方法。我们分析了西班牙语中nlg的非因果最似然顺序概率,然后,用西班牙语因果nlg生成相同短语的概率。这一对比分析表明,因果关系只倾向于英语类svo结构。我们还利用Spearman秩相关分析了最优生成顺序与因果从左到右生成顺序之间的关系。结果表明,最大似然估计量预测的理想顺序与因果顺序关系不大,可能受到目标句子句法结构的影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Pattern Recognition Letters
Pattern Recognition Letters 工程技术-计算机:人工智能
CiteScore
12.40
自引率
5.90%
发文量
287
审稿时长
9.1 months
期刊介绍: Pattern Recognition Letters aims at rapid publication of concise articles of a broad interest in pattern recognition. Subject areas include all the current fields of interest represented by the Technical Committees of the International Association of Pattern Recognition, and other developing themes involving learning and recognition.
期刊最新文献
DM-SR: Diffusion-based multimodal semantic restoration within semantic communication systems VoMarkSplat: Robust watermarking for 3D Gaussian splatting with patch and multi-convolutional voting Object – PSF: A unified representation framework for end-to-end panoptic segmentation forecasting Task-Driven learned image compression with explainability preservation for image classification VWENet: Volumetric wavelet and mixture-of-experts network for 3D medical image segmentation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1