CRBC News
Culture

Study Warns: Generative AI Loops Are Producing 'Visual Elevator Music' and Threatening Cultural Diversity

Study Warns: Generative AI Loops Are Producing 'Visual Elevator Music' and Threatening Cultural Diversity
When generative AI was left to its own devices, its outputs landed on a set of generic images – what researchers called 'visual elevator music.'Wang Zhao/AFP via Getty Images

A January 2026 experiment chained text-to-image and image-to-text models in a closed loop and found that repeated autonomous conversions produced polished but generic images—what researchers called “visual elevator music.” The collapse occurred without retraining or added data, indicating homogenization can begin before AI-generated content is used to train future models. To prevent cultural flattening, systems should be redesigned to reward deviation and support less common, creative expressions.

Generative AI was trained on centuries of human art and writing. But what happens when these systems begin to feed on their own outputs? A January 2026 experiment by Arend Hintze, Frida Proschinger Åström and Jory Schossau offers a clear—and worrying—answer.

How the Experiment Worked

The researchers chained a text-to-image model to an image-to-text model and let them run autonomously in a closed loop: image → caption → image → caption, repeated many times without human intervention. They tested many different initial prompts and varied randomness settings, yet the outputs consistently converged on a narrow set of familiar, generic motifs.

Study Warns: Generative AI Loops Are Producing 'Visual Elevator Music' and Threatening Cultural Diversity
A prompt that begins with a prime minister under stress ends with an image of an empty room with fancy furnishings.Arend Hintze, Frida Proschinger Åström and Jory Schossau,CC BY

Key Findings

Despite diverse starting prompts—ranging from detailed political scenes to more abstract descriptions—the loop rapidly collapsed into bland, polished visuals the team called “visual elevator music.” Typical outputs included atmospheric cityscapes, grand public buildings and pastoral landscapes. In one example, a richly detailed prompt about a prime minister and a fragile peace deal degraded after several iterations into a featureless formal interior with no people or narrative context.

"Visual elevator music": Pleasant to look at but stripped of specific meaning, drama and temporal or geographic detail.

Why This Matters

Crucially, this homogenization occurred without retraining the models and without adding new data. The collapse emerged purely from repeated autonomous use. That suggests homogenization can begin before any AI-generated content is used to retrain future models.

Study Warns: Generative AI Loops Are Producing 'Visual Elevator Music' and Threatening Cultural Diversity
Pretty … boring.Chris McLoughlin/Moment via Getty Images

The experiment functions as a diagnostic: it reveals what these conversion pipelines preserve when left unchecked. Contemporary culture increasingly relies on similar transformations—images summarized into captions, captions converted back into images, content ranked and filtered by automated systems. Human editors often choose from AI-generated options rather than creating from scratch. The study shows those pipelines tend to compress meaning toward what is most familiar, recognizable and easy to reproduce.

Implications For Culture And Creativity

This is not a claim that stagnation is inevitable—human creativity, communities and institutions can and do resist homogenization. But the study reframes the risk: the problem starts earlier than many assumed. Retraining on synthetic content would amplify convergence, but it isn’t necessary for convergence to begin.

Study Warns: Generative AI Loops Are Producing 'Visual Elevator Music' and Threatening Cultural Diversity
AI’s outputs are familiar because they revert to average displays of human creativity.Bulgac/iStock via Getty Images

The findings also clarify a common misconception about AI creativity: producing massive numbers of variations is not the same as producing genuine innovation. Without incentives to explore or deviate from norms, generative systems optimize for familiarity—the patterns they have learned best—so autonomy can accelerate convergence rather than foster discovery.

What Should Be Done

If generative AI is to enrich rather than flatten culture, designers and policymakers should pursue measures that encourage diversity: build models with incentives for deviation, create evaluation metrics that reward originality, and support platforms that promote less common and non-mainstream expressions. Absent such interventions, AI-mediated content pipelines are likely to keep drifting toward mediocre, uninspired outputs.

Conclusion

The experiment by Hintze, Proschinger Åström and Schossau provides empirical evidence that autonomous, repeated conversions between text and images push generative systems toward homogenized content. Cultural stagnation is no longer a purely speculative worry—this study suggests it is already happening in certain AI-mediated flows.

This article is adapted from a piece republished from The Conversation. Original author: Ahmed Elgammal, Rutgers University. The author disclosed no relevant industry conflicts.

Help us improve.

Related Articles

Trending