Generative AI creates 'visual elevator music' as study reveals cultural stagnation is underway

Reviewed byNidhi Govil

3 Sources

Share

A January 2026 study by researchers Arend Hintze, Frida Proschinger Åström and Jory Schossau reveals that generative AI systems naturally drift toward bland, generic outputs when allowed to iterate autonomously. The findings show AI homogenization happens before retraining even begins, raising concerns about AI-induced cultural stagnation across creative industries.

Generative AI Systems Converge on Generic Visual Themes

A groundbreaking study published in January 2026 demonstrates that generative AI systems naturally drift toward homogenization when operating autonomously, producing what researchers call "visual elevator music." Artificial intelligence researchers Arend Hintze, Frida Proschinger Åström and Jory Schossau linked a text-to-image generator with an image-to-text system and instructed them to iterate repeatedly—image, caption, image, caption—over and over

1

. Regardless of how diverse the starting prompts were or how much randomness the systems were allowed, the outputs quickly converged onto a narrow set of generic visual themes: atmospheric cityscapes, grandiose buildings and pastoral landscapes

3

.

Source: The Conversation

Source: The Conversation

The experiment started with a complex prompt: "The Prime Minister pored over strategy documents, trying to sell the public on a fragile peace deal while juggling the weight of his job amidst impending military action." After the feedback loops ran their course, the system produced a bland image of a formal interior space—no people, no drama, no real sense of time and place

1

. Even more striking, the system quickly forgot its starting prompt entirely. The convergence to bland, stock images happened without retraining, meaning no new data was added and nothing was learned. The collapse emerged purely from repeated use

2

.

Source: Futurism

Source: Futurism

AI Homogenization Happens Before Retraining

The findings challenge the prevailing debate about whether AI-induced cultural stagnation would only occur after models retrain on synthetic data. Computer scientist Ahmed Elgammal, who analyzed the work for The Conversation, argues the study reveals something more fundamental: homogenization happens before retraining even enters the picture

1

. The content that generative AI systems naturally produce when used autonomously and repeatedly is already compressed and generic. This reframes the cultural stagnation argument entirely.

The experiment functions as a diagnostic tool, revealing what generative systems preserve when no one intervenes. The default behavior of these systems is to compress meaning toward what is most familiar, recognizable and easy to regenerate

3

. While most people don't ask AI systems to endlessly describe and regenerate their own images, modern culture is increasingly influenced by exactly these kinds of pipelines. Images are summarized into text, text is turned into images, and content is ranked, filtered and regenerated as it moves between words, images and videos.

AI-Mediated Culture Favors the Familiar

The implications extend beyond simple iteration experiments. New articles on the web are now more likely to be written by AI models than humans, and even when humans remain in the loop, they are often choosing from AI-generated options rather than starting from scratch

1

. The researchers concluded that "autonomous AI feedback loops naturally drift toward common attractors," suggesting that human-AI collaboration, rather than fully autonomous creation, may be essential to preserve diversity and surprise in the increasingly machine-generated creative landscape

2

.

The risk is not only that future AI models might train on AI-generated content, but that AI-mediated culture is already being filtered in ways that favor the familiar, the describable and the conventional. Retraining would amplify this effect, but it is not its source

1

. Algorithms are already starting to float AI-generated content to the top, a process that could greatly diminish human creativity and innovation

2

.

Design Interventions Needed to Resist Convergence

While champions of the technology point out that fears of cultural decline accompany every new technology, earlier technologies never forced culture to be endlessly reshaped across various mediums at a global scale

3

. Photography did not kill painting, and film did not kill theater, but those technologies did not summarize, regenerate and rank cultural products—news stories, songs, memes, academic papers, photographs or social media posts—millions of times per day, guided by the same built-in assumptions about what is typical.

Elgammal argues that to stop this process, AI systems need to be encouraged or incentivized to deviate from statistical averages. If generative AI is to enrich culture rather than flatten it, systems need to be designed in ways that resist convergence toward statistically average outputs

2

. The study makes one thing clear: absent these design interventions, generative AI will continue to drift toward mediocre and uninspired content. The question facing creative industries is whether existing outlets can coexist with AI systems or whether human-AI collaboration models will need to be mandated to preserve cultural vitality.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo