2 Sources
[1]
AI LLMs Learn Like Us, But Without Abstract Thought - Neuroscience News
Summary: A new study finds that large language models (LLMs), like GPT-J, generate words not by applying fixed grammatical rules, but by drawing analogies, mirroring how humans process unfamiliar language. When faced with made-up adjectives, the LLM chose noun forms based on similarity to words it had seen in training, just as humans do. However, unlike humans, LLMs don't build mental dictionaries; they treat each instance of a word as unique, relying heavily on memorized examples. This analogy-based pattern-matching explains both their impressive fluency and their need for massive data inputs. A new study led by researchers at the University of Oxford and the Allen Institute for AI (Ai2) has found that large language models (LLMs) - the AI systems behind chatbots like ChatGPT - generalize language patterns in a surprisingly human-like way: through analogy, rather than strict grammatical rules. The findings were published on 9 May in the journal PNAS. The research challenges a widespread assumption about LLMs: that these learn how to generate language primarily by inferring rules from their training data. Instead, the models rely heavily on stored examples and draw analogies when dealing with unfamiliar words, much as people do. To explore how LLMs generate language, the study compared judgments made by humans with those made by GPT-J (an open-source large language model developed by EleutherAI in 2021) on a very common word formation pattern in English, which turns adjectives into nouns by adding the suffix "-ness" or "-ity". The LLM's responses were compared to the choices made by people, and to predictions made by two well-established cognitive models. One model generalises using rules, and another uses analogical reasoning based on similarity to stored examples. The results revealed that the LLM's behaviour resembled human analogical reasoning. Rather than using rules, it based its answers on similarities to real words it had "seen" during training - much as people do when thinking about new words. For instance, friquish is turned into friquishness on the basis of its similarity to words like selfish, whereas the outcome for cormasive is influenced by word pairs such as sensitive, sensitivity. The study also found pervasive and subtle influences of how often word forms had appeared in the training data. The LLM's responses on nearly 50,000 real English adjectives were probed, and its predictions matched the statistical patterns in its training data with striking precision. The LLM behaved as if it had formed a memory trace from every individual example of every word it has encountered during training. Drawing on these stored 'memories' to make linguistic decisions, it appeared to handle anything new by asking itself: "What does this remind me of?" The study also revealed a key difference between how human beings and LLMs form analogies over examples. Humans acquire a mental dictionary - a mental store of all the word forms that they consider to be meaningful words in their language, regardless of how often they occur. They easily recognize that forms like friquish and cormasive are not words of English at this time. To deal with these potential neologisms, they make analogical generalizations based on the variety of known words in their mental dictionaries. The LLMs, in contrast, generalize directly over all the specific instances of words in the training set, without unifying instances of the same word into a single dictionary entry. Senior author Janet Pierrehumbert, Professor of Language Modelling at Oxford University, said: "Although LLMs can generate language in a very impressive manner, it turns out that they do not think as abstractly as humans do. "This probably contributes to the fact that their training requires so much more language data than humans need to learn a language." Co-lead author Dr. Valentin Hofman (Ai2 and University of Washington) said: "This study is a great example of synergy between Linguistics and AI as research areas. "The findings give us a clearer picture of what's going on inside LLMs when they generate language, and will support future advances in robust, efficient, and explainable AI." The study also involved researchers from LMU Munich and Carnegie Mellon University. Derivational Morphology Reveals Analogical Generalization in Large Language Models What mechanisms underlie linguistic generalization in large language models (LLMs)? This question has attracted considerable attention, with most studies analyzing the extent to which the language skills of LLMs resemble rules. As of yet, it is not known whether linguistic generalization in LLMs could equally well be explained as the result of analogy. A key shortcoming of prior research is its focus on regular linguistic phenomena, for which rule-based and analogical approaches make the same predictions. Here, we instead examine derivational morphology, specifically English adjective nominalization, which displays notable variability. We introduce a method for investigating linguistic generalization in LLMs: Focusing on GPT-J, we fit cognitive models that instantiate rule-based and analogical learning to the LLM training data and compare their predictions on a set of nonce adjectives with those of the LLM, allowing us to draw direct conclusions regarding underlying mechanisms. As expected, rule-based and analogical models explain the predictions of GPT-J equally well for adjectives with regular nominalization patterns. However, for adjectives with variable nominalization patterns, the analogical model provides a much better match. Furthermore, GPT-J's behavior is sensitive to the individual word frequencies, even for regular forms, a behavior that is consistent with an analogical account but not a rule-based one. These findings refute the hypothesis that GPT-J's linguistic generalization on adjective nominalization involves rules, suggesting analogy as the underlying mechanism. Overall, our study suggests that analogical processes play a bigger role in the linguistic generalization of LLMs than previously thought.
[2]
Like humans, ChatGPT favors examples and 'memories,' not rules, to generate language
A new study led by researchers at the University of Oxford and the Allen Institute for AI (Ai2) has found that large language models (LLMs) -- the AI systems behind chatbots like ChatGPT -- generalize language patterns in a surprisingly human-like way: through analogy, rather than strict grammatical rules. The work is published in the journal Proceedings of the National Academy of Sciences. The research challenges a widespread assumption about LLMs: that they learn how to generate language primarily by inferring rules from their training data. Instead, the models rely heavily on stored examples and draw analogies when dealing with unfamiliar words, much as people do. To explore how LLMs generate language, the study compared judgments made by humans with those made by GPT-J (an open-source large language model developed by EleutherAI in 2021) on a very common word formation pattern in English, which turns adjectives into nouns by adding the suffix "-ness" or "-ity." For instance, "happy" becomes "happiness," and "available" becomes "availability." The research team generated 200 made-up English adjectives that the LLM had never encountered before -- words such as "cormasive" and "friquish." GPT-J was asked to turn each one into a noun by choosing between -ness and -ity (for example, deciding between "cormasivity" and "cormasiveness"). The LLM's responses were compared to the choices made by people, and to predictions made by two well-established cognitive models. One model generalizes using rules, and another uses analogical reasoning based on similarity to stored examples. The results revealed that the LLM's behavior resembled human analogical reasoning. Rather than using rules, it based its answers on similarities to real words it had encountered during training -- much as people do when thinking about new words. For instance, "friquish" is turned into "friquishness" on the basis of its similarity to words like "selfish," whereas the outcome for "cormasive" is influenced by word pairs such as sensitive, sensitivity. The study also found pervasive and subtle influences of how often word forms had appeared in the training data. The LLM's responses on nearly 50,000 real English adjectives were probed, and its predictions matched the statistical patterns in its training data with striking precision. The LLM behaved as if it had formed a memory trace from every individual example of every word it has encountered during training. Drawing on these stored memories to make linguistic decisions, it appeared to handle anything new by asking itself: "What does this remind me of?" The study also revealed a key difference between how human beings and LLMs form analogies over examples. Humans acquire a mental dictionary -- a mental store of all the word forms that they consider to be meaningful words in their language, regardless of how often they occur. They easily recognize that forms like friquish and cormasive are not words of English at this time. To deal with these potential neologisms, they make analogical generalizations based on the variety of known words in their mental dictionaries. The LLMs, in contrast, generalize directly over all the specific instances of words in the training set, without unifying instances of the same word into a single dictionary entry. Senior author Janet Pierrehumbert, Professor of Language Modelling at Oxford University, said, "Although LLMs can generate language in a very impressive manner, it turns out that they do not think as abstractly as humans do. This probably contributes to the fact that their training requires so much more language data than humans need to learn a language." Co-lead author Dr. Valentin Hofman (Ai2 and University of Washington) said, "This study is a great example of synergy between linguistics and AI as research areas. The findings give us a clearer picture of what's going on inside LLMs when they generate language, and will support future advances in robust, efficient, and explainable AI." The study also involved researchers from LMU Munich and Carnegie Mellon University.
Share
Copy Link
A new study finds that large language models (LLMs) like GPT-J generate language through analogy rather than fixed grammatical rules, similar to humans. However, unlike humans, LLMs don't form mental dictionaries and rely heavily on memorized examples.
A groundbreaking study led by researchers from the University of Oxford and the Allen Institute for AI (AI2) has revealed that large language models (LLMs), the AI systems powering chatbots like ChatGPT, learn and generalize language patterns in a surprisingly human-like manner. The research, published in the Proceedings of the National Academy of Sciences, challenges prevailing assumptions about how these AI models process language 1.
The study focused on GPT-J, an open-source LLM developed by EleutherAI in 2021. Researchers compared its performance to human judgments on a common English word formation pattern: turning adjectives into nouns by adding "-ness" or "-ity" suffixes 2.
Key findings include:
The research uncovered subtle influences of word frequency in the AI's training data:
While the study revealed similarities between human and AI language processing, it also highlighted crucial differences:
Janet Pierrehumbert, Professor of Language Modelling at Oxford University and senior author of the study, noted, "Although LLMs can generate language in a very impressive manner, it turns out that they do not think as abstractly as humans do. This probably contributes to the fact that their training requires so much more language data than humans need to learn a language" 1.
Dr. Valentin Hofman, co-lead author from AI2 and the University of Washington, emphasized the study's significance in bridging linguistics and AI research. He stated, "The findings give us a clearer picture of what's going on inside LLMs when they generate language, and will support future advances in robust, efficient, and explainable AI" 2.
This research provides valuable insights into the inner workings of AI language models and highlights areas for potential improvement in making AI systems more efficient and human-like in their language processing capabilities.
Summarized by
Navi
[1]
Google launches its new Pixel 10 smartphone series, showcasing advanced AI capabilities powered by Gemini, aiming to challenge competitors in the premium handset market.
20 Sources
Technology
3 hrs ago
20 Sources
Technology
3 hrs ago
Google's Pixel 10 series introduces groundbreaking AI features, including Magic Cue, Camera Coach, and Voice Translate, powered by the new Tensor G5 chip and Gemini Nano model.
12 Sources
Technology
3 hrs ago
12 Sources
Technology
3 hrs ago
NASA and IBM have developed Surya, an open-source AI model that can predict solar flares and space weather with improved accuracy, potentially helping to protect Earth's infrastructure from solar storm damage.
6 Sources
Technology
11 hrs ago
6 Sources
Technology
11 hrs ago
Google's latest smartwatch, the Pixel Watch 4, introduces significant upgrades including a curved display, enhanced AI features, and improved health tracking capabilities.
17 Sources
Technology
3 hrs ago
17 Sources
Technology
3 hrs ago
FieldAI, a robotics startup, has raised $405 million to develop "foundational embodied AI models" for various robot types. The company's innovative approach integrates physics principles into AI, enabling safer and more adaptable robot operations across diverse environments.
7 Sources
Technology
3 hrs ago
7 Sources
Technology
3 hrs ago