{"title":"基因瓶颈与通过扩展和高通量实现高智能","authors":"Arifa Khan, Saravanan P, Venkatesan S. K.","doi":"arxiv-2407.08743","DOIUrl":null,"url":null,"abstract":"We study the biological evolution of low-latency natural neural networks for\nshort-term survival, and its parallels in the development of low latency\nhigh-performance Central Processing Unit in computer design and architecture.\nThe necessity of accurate high-quality display of motion picture led to the\nspecial processing units known as the GPU, just as how special visual cortex\nregions of animals produced such low-latency computational capacity. The human\nbrain, especially considered as nothing but a scaled-up version of a primate\nbrain evolved in response to genomic bottleneck, producing a brain that is\ntrainable and prunable by society, and as a further extension, invents\nlanguage, writing and storage of narratives displaced in time and space. We\nconclude that this modern digital invention of social media and the archived\ncollective common corpus has further evolved from just simple CPU-based\nlow-latency fast retrieval to high-throughput parallel processing of data using\nGPUs to train Attention based Deep Learning Neural Networks producing\nGenerative AI with aspects like toxicity, bias, memorization, hallucination,\nwith intriguing close parallels in humans and their society. We show how this\npaves the way for constructive approaches to eliminating such drawbacks from\nhuman society and its proxy and collective large-scale mirror, the Generative\nAI of the LLMs.","PeriodicalId":501310,"journal":{"name":"arXiv - CS - Other Computer Science","volume":"47 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Genetic Bottleneck and the Emergence of High Intelligence by Scaling-out and High Throughput\",\"authors\":\"Arifa Khan, Saravanan P, Venkatesan S. K.\",\"doi\":\"arxiv-2407.08743\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We study the biological evolution of low-latency natural neural networks for\\nshort-term survival, and its parallels in the development of low latency\\nhigh-performance Central Processing Unit in computer design and architecture.\\nThe necessity of accurate high-quality display of motion picture led to the\\nspecial processing units known as the GPU, just as how special visual cortex\\nregions of animals produced such low-latency computational capacity. The human\\nbrain, especially considered as nothing but a scaled-up version of a primate\\nbrain evolved in response to genomic bottleneck, producing a brain that is\\ntrainable and prunable by society, and as a further extension, invents\\nlanguage, writing and storage of narratives displaced in time and space. We\\nconclude that this modern digital invention of social media and the archived\\ncollective common corpus has further evolved from just simple CPU-based\\nlow-latency fast retrieval to high-throughput parallel processing of data using\\nGPUs to train Attention based Deep Learning Neural Networks producing\\nGenerative AI with aspects like toxicity, bias, memorization, hallucination,\\nwith intriguing close parallels in humans and their society. We show how this\\npaves the way for constructive approaches to eliminating such drawbacks from\\nhuman society and its proxy and collective large-scale mirror, the Generative\\nAI of the LLMs.\",\"PeriodicalId\":501310,\"journal\":{\"name\":\"arXiv - CS - Other Computer Science\",\"volume\":\"47 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Other Computer Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.08743\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Other Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.08743","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
我们研究了低延迟自然神经网络的生物进化以实现短期生存,以及计算机设计和体系结构中低延迟高性能中央处理器的发展。人脑,尤其被认为是灵长类动物大脑的放大版,是在基因组瓶颈下进化而来的,它产生了一个可被社会训练和修剪的大脑,并作为进一步的延伸,发明了语言、书写和存储在时间和空间中移动的叙述。我们的结论是,社交媒体和归档的共同语料库这一现代数字发明,已经从简单的基于 CPU 的低延迟快速检索,进一步发展到使用 GPU 对数据进行高吞吐量并行处理,以训练基于注意力的深度学习神经网络,产生了具有毒性、偏差、记忆、幻觉等方面的生成式人工智能,与人类及其社会有着惊人的相似之处。我们展示了这如何为消除人类社会及其代理和集体大规模镜像--LLMs 的生成式人工智能--中的这些弊端铺平了建设性的道路。
Genetic Bottleneck and the Emergence of High Intelligence by Scaling-out and High Throughput
We study the biological evolution of low-latency natural neural networks for
short-term survival, and its parallels in the development of low latency
high-performance Central Processing Unit in computer design and architecture.
The necessity of accurate high-quality display of motion picture led to the
special processing units known as the GPU, just as how special visual cortex
regions of animals produced such low-latency computational capacity. The human
brain, especially considered as nothing but a scaled-up version of a primate
brain evolved in response to genomic bottleneck, producing a brain that is
trainable and prunable by society, and as a further extension, invents
language, writing and storage of narratives displaced in time and space. We
conclude that this modern digital invention of social media and the archived
collective common corpus has further evolved from just simple CPU-based
low-latency fast retrieval to high-throughput parallel processing of data using
GPUs to train Attention based Deep Learning Neural Networks producing
Generative AI with aspects like toxicity, bias, memorization, hallucination,
with intriguing close parallels in humans and their society. We show how this
paves the way for constructive approaches to eliminating such drawbacks from
human society and its proxy and collective large-scale mirror, the Generative
AI of the LLMs.