{"title":"Understanding Foundation Models: Are We Back in 1924?","authors":"Alan F. Smeaton","doi":"arxiv-2409.07618","DOIUrl":null,"url":null,"abstract":"This position paper explores the rapid development of Foundation Models (FMs)\nin AI and their implications for intelligence and reasoning. It examines the\ncharacteristics of FMs, including their training on vast datasets and use of\nembedding spaces to capture semantic relationships. The paper discusses recent\nadvancements in FMs' reasoning abilities which we argue cannot be attributed to\nincreased model size but to novel training techniques which yield learning\nphenomena like grokking. It also addresses the challenges in benchmarking FMs\nand compares their structure to the human brain. We argue that while FMs show\npromising developments in reasoning and knowledge representation, understanding\ntheir inner workings remains a significant challenge, similar to ongoing\nefforts in neuroscience to comprehend human brain function. Despite having some\nsimilarities, fundamental differences between FMs and the structure of human\nbrain warn us against making direct comparisons or expecting neuroscience to\nprovide immediate insights into FM function.","PeriodicalId":501479,"journal":{"name":"arXiv - CS - Artificial Intelligence","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07618","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This position paper explores the rapid development of Foundation Models (FMs)
in AI and their implications for intelligence and reasoning. It examines the
characteristics of FMs, including their training on vast datasets and use of
embedding spaces to capture semantic relationships. The paper discusses recent
advancements in FMs' reasoning abilities which we argue cannot be attributed to
increased model size but to novel training techniques which yield learning
phenomena like grokking. It also addresses the challenges in benchmarking FMs
and compares their structure to the human brain. We argue that while FMs show
promising developments in reasoning and knowledge representation, understanding
their inner workings remains a significant challenge, similar to ongoing
efforts in neuroscience to comprehend human brain function. Despite having some
similarities, fundamental differences between FMs and the structure of human
brain warn us against making direct comparisons or expecting neuroscience to
provide immediate insights into FM function.