Like many teachers at every level of education, I have spent the past two years trying to wrap my head around the question of generative AI in my English classroom. To my thinking, this is a question that ought to concern all people who like to read and write, not just teachers and their students. Today’s English students are tomorrow’s writers and readers of literature. If you enjoy thoughtful, consequential, human-generated writing—or hope for your own human writing to be read by a wide human audience—you should want young people to learn to read and write. College is not the only place where this can happen, of course, but large public universities like UVA, where I teach, are institutions that reliably turn tax dollars into new readers and writers, among other public services. I see it happen all the time.
There are valid reasons why college students in particular might prefer that AI do their writing for them: most students are overcommitted; college is expensive, so they need good grades for a good return on their investment; and AI is everywhere, including the post-college workforce. There are also reasons I consider less valid (detailed in a despairing essay that went viral recently), which amount to opportunistic laziness: if you can get away with using AI, why not?
It was this line of thinking that led me to conduct an experiment in my English classroom. I attempted the experiment in four sections of my class during the 2024-2025 academic year, with a total of 72 student writers. Rather than taking an “abstinence-only” approach to AI, I decided to put the central, existential question to them directly: was it still necessary or valuable to learn to write? The choice would be theirs. We would look at the evidence, and at the end of the semester, they would decide by vote whether A.I. could replace me.
What could go wrong?
In the weeks that followed, I had my students complete a series of writing assignments with and without AI, so that we could compare the results.
My students liked to hate on AI, and tended toward food-based metaphors in their critiques: AI prose was generally “flavorless” or “bland” compared to human writing. They began to notice its tendency to hallucinate quotes and sources, as well as its telltale signs, such as the weird prevalence of em-dashes, which my students never use, and sentences that always include exactly three examples. These tics quickly became running jokes, which made class fun: flexing their powers of discernment proved to be a form of entertainment. Without realizing it, my students had become close readers.
During these conversations, my students expressed views that reaffirmed their initial survey choices, finding that AI wasn’t great for first drafts, but potentially useful in the pre- or post-writing stages of brainstorming and editing. I don’t want to overplay the significance of an experiment with only 72 subjects, but my sense of the current AI discourse is that my students’ views reflect broader assumptions about when AI is and isn’t ethical or effective.
It’s increasingly uncontroversial to use AI to brainstorm, and to affirm that you are doing so: just last week, the hosts of the New York Times’s tech podcast spoke enthusiastically about using AI to brainstorm for the podcast itself, including coming up with interview questions and summarizing and analyzing long documents, though of course you have to double-check AI’s work. One host compares AI chatbots to “a very smart assistant who has a dozen Ph.D.s but is also high on ketamine like 30 percent of the time.”
myfavouritename@beehaw.org 1 day ago
This was an excellent read. The professor clearly thought a lot about what they wanted their students to get out of investigating the limits of AI use. It seems like it was a well crafted experiment as well as being a well crafted learning experience. Thanks for sharing the link.