As AI-generated text continues to evolve, distinguishing it from human-authored content has become increasingly difficult. This study examined whether non-expert readers could reliably differentiate between AI-generated poems and those written by well-known human poets. We conducted two experiments with non-expert poetry readers and found that participants performed below chance levels in identifying AI-generated poems (46.6% accuracy, χ2(1, N = 16,340) = 75.13, p < 0.0001). Notably, participants were more likely to judge AI-generated poems as human-authored than actual human-authored poems (χ2(2, N = 16,340) = 247.04, p < 0.0001). We found that AI-generated poems were rated more favorably in qualities such as rhythm and beauty, and that this contributed to their mistaken identification as human-authored. Our findings suggest that participants employed shared yet flawed heuristics to differentiate AI from human poetry: the simplicity of AI-generated poems may be easier for non-experts to understand, leading them to prefer AI-generated poetry and misinterpret the complexity of human poems as incoherence generated by AI.
Indeed, there are whole categories of art such as “found art” or the abstract stuff that involves throwing splats of paint at things that can’t really convey the intent of the artist because the artist wasn’t involved in specifying how it looked in the first place. The artist is more like the “first viewer” of those particular art pieces, they do or find a thing and then decide “that means something” after the fact.
It’s entirely possible to do that with something AI generated. Algorithmic art goes way back. Lots of people find graphs of the Mandelbrot Set to be beautiful.