Curated by THEOUTPOST
On Tue, 11 Mar, 8:02 AM UTC
2 Sources
[1]
New technique overcomes spurious correlations problem in AI
AI models often rely on "spurious correlations," making decisions based on unimportant and potentially misleading information. Researchers have now discovered these learned spurious correlations can be traced to a very small subset of the training data and have demonstrated a technique that overcomes the problem. "This technique is novel in that it can be used even when you have no idea what spurious correlations the AI is relying on," says Jung-Eun Kim, corresponding author of a paper on the work and an assistant professor of computer science at North Carolina State University. "If you already have a good idea of what the spurious features are, our technique is an efficient and effective way to address the problem. However, even if you are simply having performance issues, but don't understand why, you could still use our technique to determine whether a spurious correlation exists and resolve that issue." Spurious correlations are generally caused by simplicity bias during AI training. Practitioners use data sets to train AI models to perform specific tasks. For example, an AI model could be trained to identify photographs of dogs. The training data set would include pictures of dogs where the AI is told a dog is in the photo. During the training process, the AI will begin identifying specific features that it can use to identify dogs. However, if many of the dogs in the photos are wearing collars, and because collars are generally less complex features of a dog than ears or fur, the AI may use collars as a simple way to identify dogs. This is how simplicity bias can cause spurious correlations. "And if the AI uses collars as the factor it uses to identify dogs, the AI may identify cats wearing collars as dogs," Kim says. Conventional techniques for addressing problems caused by spurious correlations rely on practitioners being able to identify the spurious features that are causing the problem. They can then address this by modifying the data sets used to train the AI model. For example, practitioners might increase the weight given to photos in the data set that include dogs that are not wearing collars. However, in their new work, the researchers demonstrate that it is not always possible to identify the spurious features that are causing problems -- making conventional techniques for addressing spurious correlations ineffective. "Our goal with this work was to develop a technique that allows us to sever spurious correlations even when we know nothing about those spurious features," Kim says. The new technique relies on removing a small portion of the data used to train the AI model. "There can be significant variation in the data samples included in training data sets," Kim says. "Some of the samples can be very simple, while others may be very complex. And we can measure how 'difficult' each sample is based on how the model behaved during training. "Our hypothesis was that the most difficult samples in the data set can be noisy and ambiguous, and are most likely to force a network to rely on irrelevant information that hurt a model's performance," Kim explains. "By eliminating a small sliver of the training data that is difficult to understand, you are also eliminating the hard data samples that contain spurious features. This elimination overcomes the spurious correlations problem, without causing significant adverse effects." The researchers demonstrated that the new technique achieves state-of-the-art results -- improving performance even when compared to previous work on models where the spurious features were identifiable. The peer-reviewed paper, "Severing Spurious Correlations with Data Pruning," will be presented at the International Conference on Learning Representations (ICLR), being held in Singapore from April 24-28. First author of the paper is Varun Mulchandani, a Ph.D. student at NC State.
[2]
New technique overcomes spurious correlations problem in AI
AI models often rely on "spurious correlations," making decisions based on unimportant and potentially misleading information. Researchers have now discovered these learned spurious correlations can be traced to a very small subset of the training data and have demonstrated a technique that overcomes the problem. "This technique is novel in that it can be used even when you have no idea what spurious correlations the AI is relying on," says Jung-Eun Kim, corresponding author of a paper on the work and an assistant professor of computer science at North Carolina State University. "If you already have a good idea of what the spurious features are, our technique is an efficient and effective way to address the problem. However, even if you are simply having performance issues, but don't understand why, you could still use our technique to determine whether a spurious correlation exists and resolve that issue." Spurious correlations are generally caused by simplicity bias during AI training. Practitioners use data sets to train AI models to perform specific tasks. For example, an AI model could be trained to identify photographs of dogs. The training data set would include pictures of dogs where the AI is told a dog is in the photo. During the training process, the AI will begin identifying specific features that it can use to identify dogs. However, if many of the dogs in the photos are wearing collars, and because collars are generally less complex features of a dog than ears or fur, the AI may use collars as a simple way to identify dogs. This is how simplicity bias can cause spurious correlations. "And if the AI uses collars as the factor it uses to identify dogs, the AI may identify cats wearing collars as dogs," Kim says. Conventional techniques for addressing problems caused by spurious correlations rely on practitioners being able to identify the spurious features that are causing the problem. They can then address this by modifying the data sets used to train the AI model. For example, practitioners might increase the weight given to photos in the data set that include dogs that are not wearing collars. However, in their new work, the researchers demonstrate that it is not always possible to identify the spurious features that are causing problems -- making conventional techniques for addressing spurious correlations ineffective. "Our goal with this work was to develop a technique that allows us to sever spurious correlations even when we know nothing about those spurious features," Kim says. The new technique relies on removing a small portion of the data used to train the AI model. "There can be significant variation in the data samples included in training data sets," Kim says. "Some of the samples can be very simple, while others may be very complex. And we can measure how 'difficult' each sample is based on how the model behaved during training. "Our hypothesis was that the most difficult samples in the data set can be noisy and ambiguous, and are most likely to force a network to rely on irrelevant information that hurt a model's performance," Kim explains. "By eliminating a small sliver of the training data that is difficult to understand, you are also eliminating the hard data samples that contain spurious features. This elimination overcomes the spurious correlations problem, without causing significant adverse effects." The researchers demonstrated that the new technique achieves state-of-the-art results -- improving performance even when compared to previous work on models where the spurious features were identifiable. The peer-reviewed paper, "Severing Spurious Correlations with Data Pruning," will be presented at the International Conference on Learning Representations (ICLR 2025), being held in Singapore April 24-28.
Share
Share
Copy Link
A new method developed by researchers at North Carolina State University addresses the problem of spurious correlations in AI models, even when the specific correlations are unknown, by removing a small portion of difficult training data samples.
Researchers at North Carolina State University have developed a groundbreaking technique to overcome the problem of spurious correlations in artificial intelligence (AI) models. This new method, which involves removing a small portion of difficult training data samples, has shown promising results in improving AI performance, even when the specific spurious features are unknown 12.
Spurious correlations occur when AI models make decisions based on unimportant or misleading information. This issue often arises due to simplicity bias during the training process. For example, an AI model trained to identify dogs in photographs might rely on the presence of collars rather than more complex features like ears or fur 1.
Jung-Eun Kim, assistant professor of computer science at North Carolina State University and corresponding author of the study, explains:
"If the AI uses collars as the factor it uses to identify dogs, the AI may identify cats wearing collars as dogs." 1
Traditional methods for addressing spurious correlations typically require practitioners to identify the problematic features and modify the training data accordingly. However, the researchers demonstrated that it is not always possible to identify these spurious features, rendering conventional techniques ineffective 2.
The new technique focuses on removing a small portion of the training data that is considered "difficult" for the AI model to process. Kim elaborates:
"Our hypothesis was that the most difficult samples in the data set can be noisy and ambiguous, and are most likely to force a network to rely on irrelevant information that hurt a model's performance. By eliminating a small sliver of the training data that is difficult to understand, you are also eliminating the hard data samples that contain spurious features." 1
This approach overcomes the spurious correlations problem without causing significant adverse effects on the model's overall performance 2.
The researchers demonstrated that their new technique achieves state-of-the-art results, improving performance even when compared to previous work on models where the spurious features were identifiable 12.
This innovative method has significant implications for AI development and deployment across various sectors. It can be particularly useful in scenarios where:
The technique's versatility and effectiveness make it a valuable tool for AI practitioners and researchers working to improve the reliability and accuracy of AI models in diverse applications.
The peer-reviewed paper titled "Severing Spurious Correlations with Data Pruning" will be presented at the International Conference on Learning Representations (ICLR 2025) in Singapore from April 24-28 12. This presentation is expected to generate significant interest in the AI research community and potentially lead to further advancements in addressing AI model biases and improving overall performance.
Reference
[1]
[2]
MIT researchers have created a novel method to identify and remove specific data points in AI training datasets that contribute to bias, improving model performance for underrepresented groups while preserving overall accuracy.
3 Sources
3 Sources
Researchers warn that the proliferation of AI-generated web content could lead to a decline in the accuracy and reliability of large language models (LLMs). This phenomenon, dubbed "model collapse," poses significant challenges for the future of AI development and its applications.
8 Sources
8 Sources
Scientists at Los Alamos National Laboratory have created a novel AI defense method called Low-Rank Iterative Diffusion (LoRID) that effectively shields neural networks from adversarial attacks, setting a new benchmark in AI security.
2 Sources
2 Sources
A study reveals AI models can make accurate but nonsensical predictions from knee X-rays, highlighting the risks of 'shortcut learning' in medical AI applications.
3 Sources
3 Sources
MIT researchers have created a new algorithm called Model-Based Transfer Learning (MBTL) that significantly improves the efficiency and reliability of training AI agents for complex decision-making tasks.
3 Sources
3 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved