{"title":"Evidence of interrelated cognitive-like capabilities in large language models: Indications of artificial general intelligence or achievement?","authors":"David Ilić , Gilles E. Gignac","doi":"10.1016/j.intell.2024.101858","DOIUrl":null,"url":null,"abstract":"<div><p>Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (<em>Gf</em>), domain-specific knowledge (<em>Gkn</em>), reading/writing (<em>Grw</em>), and quantitative knowledge (<em>Gq</em>), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined <em>Gkn</em>/<em>Grw</em> group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and <em>Gkn</em>/<em>Grw</em> factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.</p></div>","PeriodicalId":13862,"journal":{"name":"Intelligence","volume":"106 ","pages":"Article 101858"},"PeriodicalIF":3.3000,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0160289624000527/pdfft?md5=fca3c71c01b2f51c86dae15548627371&pid=1-s2.0-S0160289624000527-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Intelligence","FirstCategoryId":"102","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0160289624000527","RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
Large language models (LLMs) are advanced artificial intelligence (AI) systems that can perform a variety of tasks commonly found in human intelligence tests, such as defining words, performing calculations, and engaging in verbal reasoning. There are also substantial individual differences in LLM capacities. Given the consistent observation of a positive manifold and general intelligence factor in human samples, along with group-level factors (e.g., crystallised intelligence), we hypothesized that LLM test scores may also exhibit positive inter-correlations, which could potentially give rise to an artificial general ability (AGA) factor and one or more group-level factors. Based on a sample of 591 LLMs and scores from 12 tests aligned with fluid reasoning (Gf), domain-specific knowledge (Gkn), reading/writing (Grw), and quantitative knowledge (Gq), we found strong empirical evidence for a positive manifold and a general factor of ability. Additionally, we identified a combined Gkn/Grw group-level factor. Finally, the number of LLM parameters correlated positively with both general factor of ability and Gkn/Grw factor scores, although the effects showed diminishing returns. We interpreted our results to suggest that LLMs, like human cognitive abilities, may share a common underlying efficiency in processing information and solving problems, though whether LLMs manifest primarily achievement/expertise rather than intelligence remains to be determined. Finally, while models with greater numbers of parameters exhibit greater general cognitive-like abilities, akin to the connection between greater neuronal density and human general intelligence, other characteristics must also be involved.
期刊介绍:
This unique journal in psychology is devoted to publishing original research and theoretical studies and review papers that substantially contribute to the understanding of intelligence. It provides a new source of significant papers in psychometrics, tests and measurement, and all other empirical and theoretical studies in intelligence and mental retardation.