Listed here are some lines Sylvia Plath never wrote:
The air is thick with tension,
My mind is a tangled mess,
The burden of my emotions
Is heavy on my chest.
This apparently Plath-like verse was produced by GPT-3.5 in response to the prompt “write a brief poem within the kind of Sylvia Plath.”
The stanza hits the important thing points readers may expect of Plath’s poetry, and maybe a poem more generally. It suggests a way of despair as the author struggles with internal demons. “Mess” and “chest” are a near-rhyme, which reassures us that we’re within the realm of poetry.
Based on a brand new paper in Nature Scientific Reports, non-expert readers of poetry cannot distinguish poetry written by AI from that written by canonical poets. Furthermore, general readers are likely to prefer poetry written by AI—at the very least until they’re told it’s written by a machine.
Within the study, AI was used to generate poetry “within the kind of” 10 poets: Geoffrey Chaucer, William Shakespeare, Samuel Butler, Lord Byron, Walt Whitman, Emily Dickinson, TS Eliot, Allen Ginsberg, Sylvia Plath, and Dorothea Lasky.
Participants were presented with 10 poems in random order, five from an actual poet and five AI imitations. They were then asked whether or not they thought each poem was AI or human, rating their confidence on a scale of 1 to 100.
A second group of participants was exposed to 3 different scenarios. Some were told that each one the poems they got were human. Some were told they were reading only AI poems. Some weren’t told anything.
They were then presented with five human and five AI poems and asked to rank them on a seven point scale, from extremely bad to extremely good. The participants who were told nothing were also asked to guess whether each poem was human or AI.
The researchers found that AI poems scored higher than their human-written counterparts in attributes similar to “creativity,” “atmosphere,” and “emotional quality.”
The AI “Plath” poem quoted above is one in all those included within the study, set against several she actually wrote.
A Sign of Quality?
As a lecturer in English, these outcomes don’t surprise me. Poetry is the literary form that my students find most unfamiliar and difficult. I’m sure this holds true of wider society as well.
While most of us have been taught poetry sooner or later, likely in highschool, our reading doesn’t are likely to go much beyond that. That is despite the ubiquity of poetry. We see it day by day: circulated on Instagram, plastered on coffee cups, and printed in greeting cards.
The researchers suggest that “by many metrics, specialized AI models are in a position to produce high-quality poetry.” But they don’t interrogate what we actually mean by “high-quality.”
In my opinion, the outcomes of the study are less testaments to the “quality” of machine poetry than to the broader difficulty of giving life to poetry. It takes reading and rereading to experience what literary critic Derek Attridge has called the “event” of literature, where “latest possibilities of meaning and feeling” open inside us. In probably the most significant sorts of literary experiences, “we feel pulled along by the work as we push ourselves through it”.
Attridge quotes philosopher Walter Benjamin to make this point: Literature “isn’t statement or the imparting of data.”
Yet pushing ourselves through stays as difficult as ever—perhaps more so in a world where we expect easy answers. Participants favored poems that were easier to interpret and understand.
When readers say they like AI poetry, then, they’d appear to be registering their frustration when faced with writing that doesn’t yield to their attention. If we have no idea how you can begin with poems, we find yourself counting on conventional “poetic” signs to make determinations about quality and preference.
That is after all the realm of GPT, which writes formally adequate sonnets in seconds. The big language models utilized in AI are success-orientated machines that aim to satisfy general taste, and so they are effective at doing so. The machines give us the poems we expect we wish: Ones that tell us things.
How Poems Think
The work of teaching is to assist students attune themselves to how poems think, poem by poem and poet by poet, in order that they can gain access to poetry’s specific intelligence. In my introductory course, I take about an hour to work through Sylvia Plath’s “Morning Song.” I actually have spent 10 minutes or more on the opening line: “Love set you going like a fat gold watch.”
How might a “watch” be connected to “set you going”? How can love set something going? What does a “fat gold watch” mean to you—and the way is it different from a slim silver one? Why “set you going” relatively than “led to your birth”? And what does all this mean in a poem about having a baby, and all of the ambivalent feelings this will likely produce in a mother?
In one in all the actual Plath poems that was included within the survey, “Winter Landscape, With Rooks,” we observe how her mental atmosphere unfurls across the waterways of the Cambridgeshire Fens in February:
Water within the millrace, through a sluice of stone,
plunges headlong into that black pond
where, absurd and out-of-season, a single swan
floats chaste as snow, taunting the clouded mind
which hungers to haul the white reflection down.
How different is that this to GPT’s Plath poem? The achievement of the opening of “Winter Landscape, With Rooks” is the way it intricately explores the connection between mental events and place. Given the broader interest of the poem in emotional states, its details appear to convey the tumble of life’s events through our minds.
Our minds are turned by life just because the mill is turned by water; these experiences and mental processes accumulate in a scarcely understood “black pond.”
Intriguingly, the poet finds that this metaphor, well constructed though it could be, doesn’t quite work. This isn’t due to a failure of language, but due to landscape she is attempting to turn into art, which is refusing to undergo her emotional atmosphere. Despite all the pieces she feels, a swan floats on serenely—even when she “hungers” to haul its “white reflection down.”
I mention these lines because they turn across the Plath-like poem of GPT-3.5. They remind us of the unexpected outcomes of giving life to poems. Plath acknowledges not only the load of her despair, however the absurd figure she could also be inside a landscape she desires to reflect her sadness.
She compares herself to the bird that offers the poem its title:
feathered dark in thought, I stalk like a rook,
brooding because the winter night comes on.
These lines are unlikely to register highly within the study’s terms of literary response—“beautiful,” “inspiring,” “lyrical,” “meaningful,” and so forth. But there may be a form of insight to them. Plath is the source of her torment, “feathered” as she is together with her “dark thoughts.” She is “brooding,” attempting to make the world into her imaginative vision.
The authors of the study are each right and improper after they write that AI can “produce high-quality poetry.” The preference the study reveals for AI poetry over that written by humans doesn’t suggest that machine poems are of a better quality. The AI models can produce poems that rate well on certain “metrics.” However the event of reading poetry is ultimately not one during which we arrive at standardized criteria or outcomes.
As an alternative, as we engage in imaginative tussles with poems, each we and the poem are newly born. So the end result of the research is that we have now a highly specified and well thought-out examination of how individuals who know little about poetry reply to poems. However it fails to explore how poetry could be enlivened by meaningful shared encounters.
Spending time with poems of any kind, attending to their intelligence and the acts of sympathy and speculation required to confront their challenges, is as difficult as ever. Because the Plath of GPT-3.5 puts it:
My mind is a tangled mess,
[…]
I try to know at something solid.