大型语言模型中相互关联的类认知能力的证据:人工通用智能的迹象还是成就?

IF 3.3 2区 心理学 Q1 PSYCHOLOGY, MULTIDISCIPLINARY Intelligence Pub Date : 2024-09-01 DOI:10.1016/j.intell.2024.101858
{"title":"大型语言模型中相互关联的类认知能力的证据:人工通用智能的迹象还是成就?","authors":"","doi":"10.1016/j.intell.2024.101858","DOIUrl":null,"url":null,"abstract":"<div><p>Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (<em>Gf</em>), domain-specific knowledge (<em>Gkn</em>), reading/writing (<em>Grw</em>), and quantitative knowledge (<em>Gq</em>), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined <em>Gkn</em>/<em>Grw</em> group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and <em>Gkn</em>/<em>Grw</em> factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.</p></div>","PeriodicalId":13862,"journal":{"name":"Intelligence","volume":null,"pages":null},"PeriodicalIF":3.3000,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0160289624000527/pdfft?md5=fca3c71c01b2f51c86dae15548627371&pid=1-s2.0-S0160289624000527-main.pdf","citationCount":"0","resultStr":"{\"title\":\"Evidence of interrelated cognitive-like capabilities in large language models: Indications of artificial general intelligence or achievement?\",\"authors\":\"\",\"doi\":\"10.1016/j.intell.2024.101858\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (<em>Gf</em>), domain-specific knowledge (<em>Gkn</em>), reading/writing (<em>Grw</em>), and quantitative knowledge (<em>Gq</em>), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined <em>Gkn</em>/<em>Grw</em> group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and <em>Gkn</em>/<em>Grw</em> factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.</p></div>\",\"PeriodicalId\":13862,\"journal\":{\"name\":\"Intelligence\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":3.3000,\"publicationDate\":\"2024-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S0160289624000527/pdfft?md5=fca3c71c01b2f51c86dae15548627371&pid=1-s2.0-S0160289624000527-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Intelligence\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0160289624000527\",\"RegionNum\":2,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHOLOGY, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Intelligence","FirstCategoryId":"102","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0160289624000527","RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

大型语言模型(LLM)是一种先进的人工智能(AI)系统,可以完成人类智力测试中常见的各种任务,如定义词语、进行计算和语言推理。LLM 的能力也存在很大的个体差异。鉴于在人类样本中持续观察到正的流形和一般智能因子,以及群体级因子(如结晶智能),我们假设 LLM 测试分数也可能表现出正的相互关系,这有可能产生人工一般能力(AGA)因子和一个或多个群体级因子。基于 591 名法学硕士的样本和 12 项测试的分数,我们发现了能力的正流形和一般因子的强有力的经验证据,这些测试包括流体推理(Gf)、特定领域知识(Gkn)、阅读/写作(Grw)和定量知识(Gq)。此外,我们还发现了一个 Gkn/Grw 组级综合因子。最后,LLM参数的数量与一般能力因子和Gkn/Grw因子得分呈正相关,尽管其效果呈递减趋势。我们对结果的解释是,LLM 与人类的认知能力一样,可能在处理信息和解决问题方面具有共同的潜在效率,但 LLM 是否主要表现为成就/专长而非智力仍有待确定。最后,虽然参数数量较多的模型表现出较强的类似认知能力,类似于神经元密度较高与人类一般智力之间的联系,但其他特征也必须参与其中。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Evidence of interrelated cognitive-like capabilities in large language models: Indications of artificial general intelligence or achievement?

Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (Gf), domain-specific knowledge (Gkn), reading/writing (Grw), and quantitative knowledge (Gq), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined Gkn/Grw group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and Gkn/Grw factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Intelligence
Intelligence PSYCHOLOGY, MULTIDISCIPLINARY-
CiteScore
5.80
自引率
13.30%
发文量
64
审稿时长
69 days
期刊介绍: This unique journal in psychology is devoted to publishing original research and theoretical studies and review papers that substantially contribute to the understanding of intelligence. It provides a new source of significant papers in psychometrics, tests and measurement, and all other empirical and theoretical studies in intelligence and mental retardation.
期刊最新文献
Ecological cognitive assessment has incremental validity for predicting academic performance over and above single occasion cognitive assessments The threshold for teratogenic effects on child intelligence of prenatal exposure to phenylalanine g: Formative, reflective, or both? Inconsistent Flynn effect patterns may be due to a decreasing positive manifold: Cohort-based measurement-invariant IQ test score changes from 2005 to 2024 Sex differences of school grades in childhood and adolescence: A longitudinal analysis
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1