2 Sources
[1]
Researchers take AI to 'kindergarten' in order to learn more complex tasks
In their work, published in the journal Nature Machine Intelligence, researchers found that when recurrent neural networks (RNNs) are first trained on simple cognitive tasks, they are better equipped to handle more difficult and complex ones later on. The paper's authors labeled this form of training kindergarten curriculum learning as it centers on first instilling an understanding of basic tasks and then combining knowledge of these tasks in carrying out more challenging ones. "From very early on in life, we develop a set of basic skills like maintaining balance or playing with a ball," explains Cristina Savin, an associate professor in NYU's Center for Neural Science and Center for Data Science. "With experience, these basic skills can be combined to support complex behavior -- for instance, juggling several balls while riding a bicycle. "Our work adopts these same principles in enhancing the capabilities of RNNs, which first learn a series of easy tasks, store this knowledge, and then apply a combination of these learned tasks to successfully complete more sophisticated ones." RNNs -- neural networks that are designed to process sequential information based on stored knowledge -- are particularly useful in speech recognition and language translation. However, when it comes to complex cognitive tasks, training RNNs with existing methods can prove difficult and fall short of capturing crucial aspects of animal and human behavior that AI systems aim to replicate. To address this, the study's authors -- who also included David Hocker, a postdoctoral researcher in NYU's Center for Data Science, and Christine Constantinople, a professor in NYU's Center for Data Science -- first conducted a series of experiments with laboratory rats. The animals were trained to seek out a water source in a box with several compartmentalized ports. However, in order to know when and where the water would be available, the rats needed to learn that delivery of the water was associated with certain sounds and the illumination of the port's lights -- and that the water was not delivered immediately after these cues. In order to reach the water, then, the animals needed to develop basic knowledge of multiple phenomena (e.g., sounds precede water delivery, waiting after the visual and audio cues before trying to access the water) and then learn to combine these simple tasks in order to complete a goal (water retrieval). These results pointed to principles of how the animals applied knowledge of simple tasks in undertaking more complex ones. The scientists took these findings to train RNNs in a similar fashion -- but, instead of water retrieval, the RNNs managed a wagering task that required these networks to build upon basic decision making in order to maximize the payoff over time. They then compared this kindergarten curriculum learning approach to existing RNN-training methods. Overall, the team's results showed that the RNNs trained on the kindergarten model learned faster than those trained on current methods. "AI agents first need to go through kindergarten to later be able to better learn complex tasks," observes Savin. "Overall, these results point to ways to improve learning in AI systems and call for developing a more holistic understanding of how past experiences influence learning of new skills." This research was funded by grants from the National Institute of Mental Health (1R01MH125571-01, 1K01MH132043-01A1) and conducted using research computing resources of the Empire AI consortium, with support from the State of New York, the Simons Foundation, and the Secunda Family Foundation.
[2]
AI goes to 'kindergarten' in order to learn more complex tasks
We need to learn our letters before we can learn to read and our numbers before we can learn how to add and subtract. The same principles are true with AI, a team of New York University scientists has shown through laboratory experiments and computational modeling. In their work, published in the journal Nature Machine Intelligence, researchers found that when recurrent neural networks (RNNs) are first trained on simple cognitive tasks, they are better equipped to handle more difficult and complex ones later on. The paper's authors labeled this form of training kindergarten curriculum learning as it centers on first instilling an understanding of basic tasks and then combining knowledge of these tasks in carrying out more challenging ones. "From very early on in life, we develop a set of basic skills like maintaining balance or playing with a ball," explains Cristina Savin, an associate professor in NYU's Center for Neural Science and Center for Data Science. "With experience, these basic skills can be combined to support complex behavior -- for instance, juggling several balls while riding a bicycle. "Our work adopts these same principles in enhancing the capabilities of RNNs, which first learn a series of easy tasks, store this knowledge, and then apply a combination of these learned tasks to successfully complete more sophisticated ones." RNNs -- neural networks that are designed to process sequential information based on stored knowledge -- are particularly useful in speech recognition and language translation. However, when it comes to complex cognitive tasks, training RNNs with existing methods can prove difficult and fall short of capturing crucial aspects of animal and human behavior that AI systems aim to replicate. To address this, the study's authors -- who also included David Hocker, a postdoctoral researcher in NYU's Center for Data Science, and Christine Constantinople, a professor in NYU's Center for Data Science -- first conducted a series of experiments with laboratory rats. The animals were trained to seek out a water source in a box with several compartmentalized ports. However, in order to know when and where the water would be available, the rats needed to learn that delivery of the water was associated with certain sounds and the illumination of the port's lights -- and that the water was not delivered immediately after these cues. In order to reach the water, then, the animals needed to develop basic knowledge of multiple phenomena (e.g., sounds precede water delivery, waiting after the visual and audio cues before trying to access the water) and then learn to combine these simple tasks in order to complete a goal (water retrieval). These results pointed to principles of how the animals applied knowledge of simple tasks in undertaking more complex ones. The scientists took these findings to train RNNs in a similar fashion -- but, instead of water retrieval, the RNNs managed a wagering task that required these networks to build upon basic decision-making in order to maximize the payoff over time. They then compared this kindergarten curriculum learning approach to existing RNN-training methods. Overall, the team's results showed that the RNNs trained on the kindergarten model learned faster than those trained on current methods. "AI agents first need to go through kindergarten to later be able to better learn complex tasks," observes Savin. "Overall, these results point to ways to improve learning in AI systems and call for developing a more holistic understanding of how past experiences influence learning of new skills."
Share
Copy Link
Researchers at NYU have developed a new training method for recurrent neural networks (RNNs) inspired by early childhood learning, showing improved performance on complex tasks.
Researchers at New York University have developed a novel training method for artificial intelligence systems, drawing inspiration from early childhood education. The study, published in Nature Machine Intelligence, demonstrates that recurrent neural networks (RNNs) can learn complex tasks more effectively when first trained on simpler, foundational skills 12.
The research team, led by Associate Professor Cristina Savin from NYU's Center for Neural Science and Center for Data Science, introduced what they call "kindergarten curriculum learning." This approach involves training AI systems on basic tasks before progressing to more complex ones, mirroring the way humans acquire skills from an early age 1.
"From very early on in life, we develop a set of basic skills like maintaining balance or playing with a ball," explains Savin. "With experience, these basic skills can be combined to support complex behavior -- for instance, juggling several balls while riding a bicycle" 2.
The study combined both animal experiments and computational modeling to validate their approach:
Rat Experiments: Laboratory rats were trained to locate water in a compartmentalized box. The animals had to learn to associate water delivery with specific sounds and light cues, and understand that water wasn't immediately available after these cues 12.
RNN Training: Using insights from the rat experiments, the researchers applied a similar training method to RNNs. The networks were tasked with a wagering exercise that required building upon basic decision-making skills to maximize payoffs over time 1.
Comparative Analysis: The team compared their kindergarten curriculum learning approach with existing RNN training methods. Results showed that RNNs trained using the new approach learned faster and more effectively than those trained with current techniques 2.
This research has significant implications for enhancing AI learning capabilities:
Improved Learning Efficiency: The kindergarten approach allows AI systems to acquire complex skills more rapidly and effectively 12.
Bridging the Gap: This method could help AI systems better replicate crucial aspects of animal and human behavior, addressing current limitations in RNN training for complex cognitive tasks 1.
Holistic Learning: The study emphasizes the importance of understanding how past experiences influence the acquisition of new skills in AI systems 2.
The research team suggests that this approach could lead to more sophisticated AI systems capable of handling increasingly complex tasks. "AI agents first need to go through kindergarten to later be able to better learn complex tasks," notes Savin 2.
This study, funded by grants from the National Institute of Mental Health and supported by the Empire AI consortium, opens new avenues for AI research and development. It underscores the potential of interdisciplinary approaches, combining insights from neuroscience, psychology, and computer science to advance artificial intelligence 1.
Summarized by
Navi
Google launches its new Pixel 10 series, featuring improved AI capabilities, enhanced camera systems, and the new Tensor G5 chip. The lineup includes the base Pixel 10, Pixel 10 Pro, Pixel 10 Pro XL, and Pixel 10 Pro Fold, all showcasing Google's commitment to AI-driven smartphone technology.
70 Sources
Technology
18 hrs ago
70 Sources
Technology
18 hrs ago
Google launches its new Pixel 10 smartphone series, featuring advanced AI capabilities powered by Gemini, aiming to challenge competitors in the premium handset market.
24 Sources
Technology
18 hrs ago
24 Sources
Technology
18 hrs ago
Google's latest Pixel Watch 4 introduces a curved display, AI-powered health coaching, and satellite communication, setting new standards in the smartwatch market.
19 Sources
Technology
18 hrs ago
19 Sources
Technology
18 hrs ago
FieldAI, an Irvine-based startup, has raised $405 million to develop "foundational embodied AI models" for various robots, aiming to create adaptable and safe AI systems for real-world applications.
8 Sources
Technology
18 hrs ago
8 Sources
Technology
18 hrs ago
Mustafa Suleyman, CEO of Microsoft AI, cautions about the risks of AI systems that appear conscious, urging the industry to avoid creating illusions of sentience in AI products.
5 Sources
Technology
18 hrs ago
5 Sources
Technology
18 hrs ago