{"id":13256,"date":"2023-05-25T12:35:10","date_gmt":"2023-05-25T02:35:10","guid":{"rendered":"https:\/\/rationalemagazine.com\/?p=13256"},"modified":"2023-05-25T12:35:10","modified_gmt":"2023-05-25T02:35:10","slug":"the-impact-and-implications-of-large-language-models","status":"publish","type":"post","link":"https:\/\/rationalemagazine.com\/index.php\/2023\/05\/25\/the-impact-and-implications-of-large-language-models\/","title":{"rendered":"The impact and implications of large language models"},"content":{"rendered":"<p><span style=\"font-weight: 400;\">Large language models (LLMs) such as ChatGPT and Google Bard have attracted a great deal of publicity in recent months due to their impressive abilities to engage in conversation, answer open-ended questions, find information, write essays, generate computer code, produce poetry, and solve certain problems.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The improvement of these systems has generated <\/span><a href=\"https:\/\/futureoflife.org\/open-letter\/pause-giant-ai-experiments\/\"><span style=\"font-weight: 400;\">significant concern<\/span><\/a><span style=\"font-weight: 400;\"> regarding the potential harms of such technology. Various potential risks have been raised, including use of LLMs to rapidly <\/span><a href=\"https:\/\/www.nytimes.com\/2023\/02\/08\/technology\/ai-chatbots-disinformation.html\"><span style=\"font-weight: 400;\">spread misinformation<\/span><\/a><span style=\"font-weight: 400;\">, rapid <\/span><a href=\"https:\/\/www.cnbc.com\/2023\/03\/28\/ai-automation-could-impact-300-million-jobs-heres-which-ones.html\"><span style=\"font-weight: 400;\">automation<\/span><\/a><span style=\"font-weight: 400;\"> of large numbers of jobs, and artificial agents behaving in unexpected ways or even <\/span><a href=\"https:\/\/nypost.com\/2023\/01\/26\/rogue-ai-could-kill-everyone-scientists-warn\/\"><span style=\"font-weight: 400;\">escaping human control<\/span><\/a><span style=\"font-weight: 400;\"> entirely.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">A major driving force underpinning these concerns is the rapid rate of improvement of LLMs, and the belief that this rate of progress will soon lead to systems that exceed human capabilities in many tasks, leading to drastically disruptive effects on the economy and society.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">In this essay, I will argue that such fears are exaggerated. While there are legitimate concerns about the safety and reliability of LLMs, I do not think it is likely that such systems will soon reach human levels of intelligence or capability in a broad range of tasks.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Instead, I argue that such systems have intrinsic limitations which cannot be overcome within the existing development paradigm. Also, continual growth in capabilities based on increasing the number of parameters and size of the training data will only continue for a few more years before running its course.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">I also argue that the adoption of such systems will be slow, occurring over years to decades rather than months to years (as <\/span><a href=\"https:\/\/forum.effectivealtruism.org\/posts\/D8GitXAMt7deG8tBc\/how-quickly-ai-could-transform-the-world-tom-davidson-on-the#Why_AI_takeoff_might_be_shockingly_fast\"><span style=\"font-weight: 400;\">some have argued<\/span><\/a><span style=\"font-weight: 400;\">). Therefore, their impacts on society and the economy will be more gradual and evolutionary rather than sudden and revolutionary.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Current <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Large_language_model\"><span style=\"font-weight: 400;\">LLMs<\/span><\/a><span style=\"font-weight: 400;\"> are based on the <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Transformer_(machine_learning_model)\"><span style=\"font-weight: 400;\">transformer architecture<\/span><\/a><span style=\"font-weight: 400;\">. These are very large neural networks which are trained on huge corpuses of data, most of which is from the internet. The models are usually trained to predict the next word in a sentence, and during training they learn complex statistical associations between words in natural language.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Recently, OpenAI has extended this framework by adding a technique called <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Reinforcement_learning_from_human_feedback\"><span style=\"font-weight: 400;\">Reinforcement Learning from Human Feedback<\/span><\/a><span style=\"font-weight: 400;\"> (RLHF). This involves presenting queries and their corresponding LLM outputs to humans, who then provide ratings as to the quality of the responses.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">These ratings are then used to fine-tune the language model, altering its output to improve its ratings from human feedback. This technique has enabled language models to produce output that is more useful to humans, and has improved the performance of language models as chatbots.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The OpenAI team has also made other additions and modifications to its newest model (GPT-4) to improve its capabilities as a chatbot, though very few public details are available about this.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Judging by the number of contributors to the <\/span><a href=\"https:\/\/arxiv.org\/abs\/2303.08774\"><span style=\"font-weight: 400;\">GPT-4 paper<\/span><\/a><span style=\"font-weight: 400;\"> (which lists 93 \u2018core contributors\u2019 and hundreds of other contributors) relative to previous <\/span><a href=\"https:\/\/arxiv.org\/abs\/2005.14165\"><span style=\"font-weight: 400;\">GPT-3 paper<\/span><\/a><span style=\"font-weight: 400;\"> (which lists only 31 authors), it appears that OpenAI has devoted a lot of time to adjusting, augmenting, and modifying the model in various ways.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">We know that systems have been put in place to filter out queries likely to lead to harmful or offensive results. There is also <\/span><a href=\"https:\/\/arxiv.org\/abs\/2303.12712\"><span style=\"font-weight: 400;\">evidence<\/span><\/a><span style=\"font-weight: 400;\"> that GPT-4 has a limited ability to check for faulty assumptions in the queries or instructions it is given, though it is unclear how this has been done. Nonetheless, it appears that extensive development work has been done beyond the initial stage of training the transformer on a large text corpus.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">In my view, the fact that such extensive augmentations and modifications are necessary is an indication of the underlying weaknesses and limitations of the transformer architecture. These models learn complex associations between words, but do not form the same structured, flexible, multimodal representations of word meaning as humans do.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">As such, they <\/span><a href=\"https:\/\/arxiv.org\/abs\/2109.01247\"><span style=\"font-weight: 400;\">do not truly \u2018understand\u2019 language<\/span><\/a><span style=\"font-weight: 400;\"> in the same sense as humans can. For many applications, this does not matter. But in other cases it can manifest in extremely bizarre behaviour, including models accepting absurd premises, making <\/span><a href=\"https:\/\/w3nhao.github.io\/2023\/03\/30\/Limitation-ChatGPT-Causal-Inference\/\"><span style=\"font-weight: 400;\">faulty inferences<\/span><\/a><span style=\"font-weight: 400;\">, making <\/span><a href=\"https:\/\/www.scientificamerican.com\/article\/chatgpt-explains-why-ais-like-chatgpt-should-be-regulated1\/\"><span style=\"font-weight: 400;\">contradictory statements<\/span><\/a><span style=\"font-weight: 400;\">, and failing to incorporate information that is provided.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">A related issue is the known tendency of LLMs to \u2018<\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Hallucination_(artificial_intelligence)\"><span style=\"font-weight: 400;\">hallucinate<\/span><\/a><span style=\"font-weight: 400;\">\u2019, making up facts, information, or non-existent libraries of computer code when giving responses. I dislike the term \u2018hallucination\u2019 because it implies there is some fundamental distinction between veridical knowledge that the LLM has correctly learned and hallucinations, which it simply makes up.<\/span><\/p>\n<blockquote><p><strong>In my view, the fact that such extensive augmentations and modifications are necessary is an indication of the underlying weaknesses and limitations of the transformer architecture.<\/strong><\/p><\/blockquote>\n<p><span style=\"font-weight: 400;\">In fact, there is no such distinction, because <\/span><a href=\"https:\/\/spectrum.ieee.org\/ai-hallucination\"><span style=\"font-weight: 400;\">LLMs do not form memories<\/span><\/a><span style=\"font-weight: 400;\"> of events or facts in the way humans do. All they are capable of is storing complex statistical associations in their billions of learned parameters.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">When the model produces some string of words as an output, this is equally the product of its internal learned parameters regardless of whether humans would evaluate the string as true or false.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Furthermore, an LLM has no notion of truth or falsity; it simply learns word associations. (Here I am ignoring the possibility that GPT-4 may be augmented with capabilities beyond its basic transformer architecture, since there is no public information about this. And, at any rate, the underlying architecture is still a transformer model). As such, the problem of \u2018hallucinations\u2019 is not some teething issue or minor annoyance, but is intrinsic to the architecture and method of training of LLMs.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Of course, various proposals exist for how to mitigate this limitation, such as augmenting LLMs with curated datasets of encyclopaedic facts or <\/span><a href=\"https:\/\/arxiv.org\/abs\/1906.05317\"><span style=\"font-weight: 400;\">common-sense knowledge<\/span><\/a><span style=\"font-weight: 400;\">. While promising, such proposals are not new and <\/span><a href=\"https:\/\/arxiv.org\/abs\/2202.03629\"><span style=\"font-weight: 400;\">face many problems<\/span><\/a><span style=\"font-weight: 400;\"> of their own right. While they may be successful in the long run, I do not believe there is any simple or easily implemented solution to the problem of \u2018hallucinations\u2019 in LLMs.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Another core limitation of LLMs which has been the focus of extensive research is their difficulty in exhibiting compositionality. This refers to the ability to combine known elements in novel ways by following certain abstract rules.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Many cognitive scientists have argued that compositionality is a <\/span><a href=\"https:\/\/iep.utm.edu\/compositionality-in-language\/\"><span style=\"font-weight: 400;\">critical component<\/span><\/a><span style=\"font-weight: 400;\"> of the human ability to understand novel sentences with combinations of words and ideas never previously encountered.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Prior to the release of GPT-4, the best transformer models still <\/span><a href=\"https:\/\/www.annualreviews.org\/doi\/abs\/10.1146\/annurev-linguistics-031120-122924\"><span style=\"font-weight: 400;\">struggled to perform many compositional<\/span><\/a><span style=\"font-weight: 400;\"> tasks, often only succeeding when augmented with symbolic components (which is difficult to scale to real-world tasks), or when given special task-specific training.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">At the time of writing, I am not aware of GPT-4 having been subjected to these types of tests. Although I anticipate it would outperform most existing models \u2013 given that it shares the same transformer architecture \u2013 I doubt it will be able to completely solve the problem of compositionality.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The underlying limitation is that transformer-based language models do not learn explicit symbolic representations, and hence struggle to generalise appropriately in accordance with systematic rules.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">There have also been efforts to circumvent some of these limitations and use LLMs for a wider range of tasks by developing them into a partially autonomous agent. The approach is to chain together a series of instructions, allowing the model to step through subcomponents of a task and reason its way to the desired conclusion.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">One such project called <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Auto-GPT\"><span style=\"font-weight: 400;\">Auto-GPT<\/span><\/a><span style=\"font-weight: 400;\"> involves augmenting GPT with the ability to read and write from external memory, and allowing it access to various external software packages through their application programming interfaces (APIs).<\/span><\/p>\n<p><span style=\"font-weight: 400;\">It is too early to say what will become of such projects, though early investigations indicate some promising results but also plenty of <\/span><a href=\"https:\/\/jina.ai\/news\/auto-gpt-unmasked-hype-hard-truths-production-pitfalls\/\"><span style=\"font-weight: 400;\">difficulties<\/span><\/a><span style=\"font-weight: 400;\">. In particular, the model often gets stuck in loops, fails to correctly incorporate contextual knowledge to constrain solutions to the problem, and has no ability to generalise results to similar future problems.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Such difficulties illustrate that LLMs are not designed to be general purpose agents, and hence lack many cognitive faculties such as planning, learning, decision making, or symbolic reasoning.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Furthermore, it is exceedingly unlikely that simply \u2018plugging in\u2019 various components to an LLM in an ad hoc manner will result in an agent capable of performing competently in a diverse range of environments.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The way the components are connected and interact is absolutely crucial to the overall capabilities of the system. The structure of the different cognitive components of an agent is called a <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Cognitive_architecture\"><span style=\"font-weight: 400;\">cognitive architecture<\/span><\/a><span style=\"font-weight: 400;\">, and there has been decades of research into this topic in both cognitive psychology and computer science.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">As such, I think it is na\u00efve to believe that such research will be rendered irrelevant or obsolete by the simple expedient of augmenting LLMs with a few additional components. Instead, I expect that LLMs will form one component of many that will need to be incorporated into a truly general-purpose intelligent system, one which will likely take decades of further research to develop.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Recent improvements in LLMs have primarily occurred as a result of dramatic increases in both the number of model parameters and the size of the training datasets. This has led to a rapid increase in training costs, largely due to the electricity usage and rental or opportunity cost of the required hardware. For example, the cost of training GPT-3 was probably <\/span><a href=\"https:\/\/www.reddit.com\/r\/MachineLearning\/comments\/hwfjej\/d_the_cost_of_training_gpt3\/\"><span style=\"font-weight: 400;\">several million dollars<\/span><\/a><span style=\"font-weight: 400;\">, compared to over <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/GPT-4#Training\"><span style=\"font-weight: 400;\">one hundred million<\/span><\/a><span style=\"font-weight: 400;\"> for GPT-4.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Assuming current growth rates continue, within about five years further increasing model size will become infeasible even for the biggest governments and tech firms, as training costs will reach tens of billions of dollars.\u00a0<\/span><\/p>\n<blockquote><p><strong>The next few years will be a critical period for LLMs, in which there will be much experimentation and failed attempts as companies compete to find the <a href=\"https:\/\/www.ibm.com\/watson\/resources\/ai-adoption\">best way to deploy<\/a> the technology.<\/strong><\/p><\/blockquote>\n<p><span style=\"font-weight: 400;\">Separately from the issue of training cost, there is also the question of the availability of training data. Existing models require enormous training datasets, with the size increasing exponentially from one iteration to the next. For example, GPT-3 was trained on a primary corpus of <\/span><a href=\"https:\/\/gptblogs.com\/chatgpt-how-much-data-is-used-in-the-training-process#data-quality-5.1\"><span style=\"font-weight: 400;\">300 billion words<\/span><\/a><span style=\"font-weight: 400;\"> derived from the internet. Based on historical trends, <\/span><a href=\"https:\/\/epochai.org\/blog\/will-we-run-out-of-ml-data-evidence-from-projecting-dataset\"><span style=\"font-weight: 400;\">Epoch estimates<\/span><\/a><span style=\"font-weight: 400;\"> that high quality language data will be exhausted by 2024 or 2025, and low quality data by 2032.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">I am not arguing here that the development of LLMs will cease within five years or that further improvements are impossible. Rather my point is that the primary method by which improvements have been achieved over the past five years will cease to be feasible. As such, we cannot expect current rates of progress to continue indefinitely. Similar views have been expressed by other researchers, including <\/span><a href=\"https:\/\/www.reddit.com\/r\/singularity\/comments\/zmc7gn\/ben_goertzel_architecture_behind_chatgptgpt3gpt4\/\"><span style=\"font-weight: 400;\">Ben Goertzel<\/span><\/a><span style=\"font-weight: 400;\">, <\/span><a href=\"https:\/\/garymarcus.substack.com\/p\/gpt-4s-successes-and-gpt-4s-failures\"><span style=\"font-weight: 400;\">Gary Marcus<\/span><\/a><span style=\"font-weight: 400;\">, and <\/span><a href=\"https:\/\/www.wired.com\/story\/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over\/\"><span style=\"font-weight: 400;\">Sam Altman<\/span><\/a><span style=\"font-weight: 400;\">.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">In light of these considerations, along with the intrinsic limitations discussed above, I do not think it is plausible that LLMs will reach or exceed human performance in a wide range of tasks in the near future.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The next few years will be a critical period for LLMs, in which there will be much experimentation and failed attempts as companies compete to find the <\/span><a href=\"https:\/\/www.ibm.com\/watson\/resources\/ai-adoption\"><span style=\"font-weight: 400;\">best way to deploy<\/span><\/a><span style=\"font-weight: 400;\"> the technology. It will take considerable time and effort to turn LLMs into a viable product, and even longer to adapt its use to various speciality applications and for the technology to become <\/span><a href=\"https:\/\/www.weforum.org\/reports\/the-future-of-jobs-report-2020\/in-full\/2-1-technological-adoption\"><span style=\"font-weight: 400;\">widely adopted<\/span><\/a><span style=\"font-weight: 400;\">.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Many companies and organisations will seek ways to use LLMs to augment their existing internal processes and procedures, which also will take a great deal of time and trial and error.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Contrary to what some have implied, no new technology can ever simply be \u2018plugged in\u2019 to existing processes without substantial change or adaptation. Just as automobiles, computers, and the internet took decades to have major economic and social impacts, so too I expect LLMs will take decades to have such impacts. Yet other technologies, such as <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Nuclear_fusion\"><span style=\"font-weight: 400;\">nuclear fusion<\/span><\/a><span style=\"font-weight: 400;\">, <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Reusable_launch_vehicle\"><span style=\"font-weight: 400;\">reusable launch vehicles<\/span><\/a><span style=\"font-weight: 400;\">, commercial <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/Concorde\"><span style=\"font-weight: 400;\">supersonic flight<\/span><\/a><span style=\"font-weight: 400;\">, are still yet to achieve their promised substantial impact.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">One of the major limitations of using existing LLMs is their unreliability. No important processes can currently be trusted to LLMs, because we have very little understanding of how they work, limited knowledge of the limits of their capabilities, and a poor understanding of how and when they fail. They are able to perform impressive feats, but then fail in particularly unexpected and surprising ways.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Unpredictability and unreliability both make it very difficult to use LLMs for many business or government tasks. Of course, humans regularly make mistakes, but human capabilities and fallibilities are better understood than those of LLMs, and existing political, economic, and governance systems have been developed over many decades to manage human mistakes and imperfections.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">I expect it will similarly take many years to build systems to effectively work around the limitations of LLMs and achieve sufficient reliability for widespread deployment.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">It is also valuable to take a historical perspective, as the field of artificial intelligence has seen numerous examples of excessive hype and inflated expectations. In the late 1950s and early 1960s, there was a wave of enthusiasm about the promise of logic-based systems and automated reasoning, which were thought to be capable of overtaking humans in many tasks within a matter of years.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The failure of many of these predictions led to the <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/AI_winter\"><span style=\"font-weight: 400;\">first \u2018AI winter<\/span><\/a><span style=\"font-weight: 400;\">\u2019 of the 1970s. The 1980s saw a resurgence of interest in AI, this time based on new approaches such as expert systems, the backpropagation algorithm, and initiatives such as Japan&#8217;s Fifth Generation computer initiative. Underperformance of these systems and techniques led to <\/span><a href=\"https:\/\/towardsdatascience.com\/history-of-the-second-ai-winter-406f18789d45\"><span style=\"font-weight: 400;\">another \u2018AI winter<\/span><\/a><span style=\"font-weight: 400;\">\u2019 in the 1990s and early 2000s.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">The most recent resurgence of interest in AI has largely been driven by breakthroughs in machine learning and the availability of much larger sources of data for training. Progress in the past 15 years has been rapid and impressive, but even so there have been numerous instances of inflated expectations and failed promises.\u00a0<\/span><\/p>\n<p><span style=\"font-weight: 400;\">IBM\u2019s Watson system which won <\/span><i><span style=\"font-weight: 400;\">Jeopardy!<\/span><\/i><span style=\"font-weight: 400;\"> in 2011 was heralded by IBM as a critical breakthrough in AI research, but subsequently they spent years attempting to adapt the system for use in medical diagnosis with <\/span><a href=\"https:\/\/spectrum.ieee.org\/how-ibm-watson-overpromised-and-underdelivered-on-ai-health-care\"><span style=\"font-weight: 400;\">little success<\/span><\/a><span style=\"font-weight: 400;\">.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Self-driving cars developed by Google attracted <\/span><a href=\"https:\/\/en.wikipedia.org\/wiki\/History_of_self-driving_cars#The_2010s\"><span style=\"font-weight: 400;\">substantial publicity in 2012<\/span><\/a><span style=\"font-weight: 400;\"> with their ability to drive autonomously on public roads with minimal human intervention. But, a decade later, there remain <\/span><a href=\"https:\/\/www.bloomberg.com\/news\/features\/2022-10-06\/even-after-100-billion-self-driving-cars-are-going-nowhere#xj4y7vzkg\"><span style=\"font-weight: 400;\">considerable challenges<\/span><\/a><span style=\"font-weight: 400;\"> in closing the last few small portions of the journey where humans still need to take over.\u00a0<\/span><\/p>\n<p><a href=\"https:\/\/rationalist.com.au\/make-a-donation\/\"><img fetchpriority=\"high\" decoding=\"async\" class=\"alignnone size-large wp-image-11873\" src=\"https:\/\/rationalemagazine.com\/wp-content\/uploads\/2022\/07\/Rationale-donation-1024x256.png\" alt=\"\" width=\"1024\" height=\"256\" srcset=\"https:\/\/rationalemagazine.com\/wp-content\/uploads\/2022\/07\/Rationale-donation-1024x256.png 1024w, https:\/\/rationalemagazine.com\/wp-content\/uploads\/2022\/07\/Rationale-donation-300x75.png 300w, https:\/\/rationalemagazine.com\/wp-content\/uploads\/2022\/07\/Rationale-donation-768x192.png 768w, https:\/\/rationalemagazine.com\/wp-content\/uploads\/2022\/07\/Rationale-donation-1536x384.png 1536w, https:\/\/rationalemagazine.com\/wp-content\/uploads\/2022\/07\/Rationale-donation.png 1600w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/a><\/p>\n<p><span style=\"font-weight: 400;\">While such comparisons can never be definitive, I believe these historical precedents should temper our expectations about the rate of progress of the latest set of techniques in artificial intelligence research.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">In conclusion, LLMs have intrinsic limitations which are unlikely to be resolved without fundamental new paradigms. The increasing costs of training and limited stock of quality training data will mean that growth of LLMs at present rates will not be able to continue for more than a few years. Furthermore, historical parallels indicate that it will take years for LLMs to become widely adopted and integrated into existing economic and social processes.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">Overall, there is little reason to believe that LLMs are likely to exceed human capabilities in a wide range of tasks within a few years, or displace large fractions of the workforce. These outcomes may occur in 30 or 50 years time, but almost certainly not within the next five or 10 years \u2013 and not solely due to the continued development of LLMs.<\/span><\/p>\n<p><span style=\"font-weight: 400;\">While there are legitimate concerns and problems associated with the rapid improvement of LLMs, we should not be distracted by inflated concerns about catastrophic impacts in the near term.<\/span><\/p>\n<p><b><i>If you wish to republish this original article, please attribute to\u00a0<\/i><\/b><a href=\"https:\/\/rationalemagazine.com\/\"><b><i>Rationale<\/i><\/b><\/a><b><i>.\u00a0<\/i><\/b><a href=\"https:\/\/rationalemagazine.com\/index.php\/publishing-guidelines\/\"><b><i>Click here<\/i><\/b><\/a><b><i>\u00a0to find out more about republishing under Creative Commons.<\/i><\/b><\/p>\n<p><b><i>Photo by <a href=\"https:\/\/unsplash.com\/photos\/1Na806ZwUPg\">Mojahid Mottakin<\/a><\/i><\/b><b><i>\u00a0on Unsplash.<\/i><\/b><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Large language models (LLMs) such as ChatGPT and Google Bard have attracted a great deal of publicity in recent months<\/p>\n","protected":false},"author":139,"featured_media":13257,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[65],"tags":[562],"coauthors":[156],"class_list":["post-13256","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-science-health","tag-artificial-intelligence"],"acf":[],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/posts\/13256","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/users\/139"}],"replies":[{"embeddable":true,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/comments?post=13256"}],"version-history":[{"count":2,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/posts\/13256\/revisions"}],"predecessor-version":[{"id":13259,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/posts\/13256\/revisions\/13259"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/media\/13257"}],"wp:attachment":[{"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/media?parent=13256"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/categories?post=13256"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/tags?post=13256"},{"taxonomy":"author","embeddable":true,"href":"https:\/\/rationalemagazine.com\/index.php\/wp-json\/wp\/v2\/coauthors?post=13256"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}