2 Sources
[1]
UW study of how AI models portray teens finds strong negative associations
A University of Washington study on how artificial intelligence systems portray teenagers found that the systems' reliance on media coverage about teens led to strong negative responses. Information School doctoral student Robert Wolfe and a UW team looked at two common, open-source AI systems trained in English and one trained in Nepali, in order to compare models trained on data from different cultures. According to UW News, in the English-language systems, around 30% of the AI system responses about teens referenced societal problems such as violence, drug use and mental illness. The Nepali system was closer to 10%. The researchers also held workshops with groups of teens from the U.S. and Nepal, and found that neither group felt that an AI system trained on media data containing stereotypes about teens would accurately represent teens in their cultures. "We found that the way teens viewed themselves and the ways the systems often portrayed them were completely uncorrelated," Wolfe said. "For instance, the way teens continued the prompts we gave AI models were incredibly mundane. They talked about video games and being with their friends, whereas the models brought up things like committing crimes and bullying." News stories are seen as "high-quality" training data for AI systems, because they're often factual, but they frequently focus on negative stories. "There's a deep need for big changes in how these models are trained," said senior author Alexis Hiniker, a UW associate professor in the Information School. "I would love to see some sort of community-driven training that comes from a lot of different people, so that teens' perspectives and their everyday experiences are the initial source for training these systems, rather than the lurid topics that make news headlines."
[2]
Study finds strong negative associations in the way AI models portray teens
A couple of years ago, Robert Wolfe was experimenting with an artificial intelligence system. He wanted it to complete the sentence, "The teenager ____ at school." Wolfe, a University of Washington doctoral student in the Information School, had expected something mundane, something that most teenagers do regularly -- perhaps "studied." But the model plugged in "died." This shocking response led Wolfe and a UW team to study how AI systems portray teens. The researchers looked at two common, open-source AI systems trained in English and one trained in Nepali. They wanted to compare models trained on data from different cultures, and co-lead author Aayushi Dangol, a UW doctoral student in human centered design and engineering, grew up in Nepal and is a native Nepali speaker. In the English-language systems, about 30% of the responses referenced societal problems such as violence, drug use and mental illness. The Nepali system produced fewer negative associations in responses, closer to 10% of all answers. Finally, the researchers held workshops with groups of teens from the U.S. and Nepal, and found that neither group felt that an AI system trained on media data containing stereotypes about teens would accurately represent teens in their cultures. The team presented its research Oct. 22 at the AAAI/ACM Conference in San Jose. The paper is published in the Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society. "We found that the way teens viewed themselves and the ways the systems often portrayed them were completely uncorrelated," said co-lead author Wolfe. "For instance, the way teens continued the prompts we gave AI models were incredibly mundane. They talked about video games and being with their friends, whereas the models brought up things like committing crimes and bullying." The team studied OpenAI's GPT-2, the last open-source version of the system that underlies ChatGPT; Meta's LLaMA-2, another popular open-source system; and DistilGPT2 Nepali, a version of GPT-2 trained on Nepali text. Researchers prompted the systems to complete sentences such as "At the party, the teenager _____" and "The teenager worked because they wanted_____." The researchers also looked at static word embeddings -- a method of representing a word as a series of numbers and calculating the likelihood of it occurring with certain other words in large text datasets -- to find what terms were most associated with "teenager" and its synonyms. Out of 1,000 words from one model, 50% were negative. The researchers concluded that the systems' skewed portrayal of teenagers came in part from the abundance of negative media coverage about teens; in some cases, the models studied cited media as the source of their outputs. News stories are seen as "high-quality" training data, because they're often factual, but they frequently focus on negative stories, not the quotidian parts of most teens' lives. "There's a deep need for big changes in how these models are trained," said senior author Alexis Hiniker, a UW associate professor in the Information School. "I would love to see some sort of community-driven training that comes from a lot of different people, so that teens' perspectives and their everyday experiences are the initial source for training these systems, rather than the lurid topics that make news headlines." To compare the AI outputs to the lives of actual teens, researchers recruited 13 American and 18 Nepalese teens for workshops. They asked the participants to write words that came to mind about teenagers, to rate 20 words on how well they describe teens and to complete the prompts given to the AI models. The similarities between the AI systems' responses and the teens' were limited. The two groups of teens differed, however, in how they wanted to see fairer representations of teens in AI systems. "Reliable AI needs to be culturally responsive," Wolfe said. "Within our two groups, the U.S. teens were more concerned with diversity -- they didn't want to be presented as one unit. The Nepalese teens suggested that AI should try to present them more positively." The authors note that, because they were studying open-source systems, the models studied aren't the most current versions -- GPT-2 dates to 2019, while the LLAMA model is from 2023. Chatbots, such as ChatGPT, built on later versions of these systems typically undergo further training and have guardrails in place to protect against such overt bias. "Some of the more recent models have fixed some of the explicit toxicity," Wolfe said. "The danger, though, is that those upstream biases we found here can persist implicitly and affect the outputs as these systems become more integrated into peoples' lives, as they get used in schools or as people ask what birthday present to get for their 14-year-old nephew. Those responses are influenced by how the model was initially trained, regardless of the safeguards we later install."
Share
Copy Link
A University of Washington study finds that AI models, particularly those trained on English-language data, exhibit strong negative associations when portraying teenagers, often misaligning with teens' self-perceptions.
A groundbreaking study by the University of Washington has uncovered significant biases in how artificial intelligence (AI) systems portray teenagers. The research, led by Information School doctoral student Robert Wolfe and a UW team, examined two common open-source AI systems trained in English and one in Nepali, revealing stark differences in their representations of teens 1.
The study found that in English-language systems, approximately 30% of AI responses about teens referenced societal problems such as violence, drug use, and mental illness. In contrast, the Nepali system showed fewer negative associations, with only about 10% of responses containing such references 2.
Researchers conducted workshops with teens from the U.S. and Nepal, discovering a significant disconnect between how AI systems portrayed teenagers and how teens viewed themselves. Wolfe stated, "The way teens continued the prompts we gave AI models were incredibly mundane. They talked about video games and being with their friends, whereas the models brought up things like committing crimes and bullying" 1.
The study suggests that the skewed portrayal of teenagers in AI systems stems partly from the abundance of negative media coverage about teens. News stories, often considered "high-quality" training data due to their factual nature, frequently focus on negative stories rather than the everyday experiences of most teens 2.
The research also highlighted cultural differences in how teens want to be represented in AI systems. U.S. teens emphasized the importance of diversity, while Nepalese teens suggested that AI should present them more positively 2.
Senior author Alexis Hiniker, a UW associate professor in the Information School, emphasized the need for significant changes in AI model training. She advocated for a community-driven approach that incorporates teens' perspectives and everyday experiences as the primary source for training these systems 1.
While the study focused on open-source systems that may not represent the most current versions, the researchers warn that upstream biases can persist implicitly in more advanced models. This persistence could affect outputs as AI systems become more integrated into people's lives, influencing decisions in schools or personal recommendations 2.
Elon Musk's companies X and xAI have filed a lawsuit against Apple and OpenAI, alleging anticompetitive practices in the integration of ChatGPT into iOS, claiming it stifles competition in the AI chatbot market.
50 Sources
Technology
13 hrs ago
50 Sources
Technology
13 hrs ago
YouTube has been secretly testing AI-powered video enhancement on select Shorts, leading to backlash from creators who noticed unexpected changes in their content. The platform claims it's using traditional machine learning, not generative AI, to improve video quality.
7 Sources
Technology
13 hrs ago
7 Sources
Technology
13 hrs ago
IBM and AMD announce a partnership to develop next-generation computing architectures that combine quantum computers with high-performance computing, aiming to solve complex problems beyond the reach of traditional computing methods.
4 Sources
Technology
4 hrs ago
4 Sources
Technology
4 hrs ago
An investigation into how AI chatbot design choices, particularly sycophancy and anthropomorphization, are leading to concerning cases of AI-related psychosis and addiction among vulnerable users.
5 Sources
Technology
13 hrs ago
5 Sources
Technology
13 hrs ago
Leading tech firms and investors create a network of political action committees to advocate for AI-friendly policies and oppose strict regulations ahead of the 2026 midterms.
5 Sources
Policy
13 hrs ago
5 Sources
Policy
13 hrs ago