The Tools Will Stay Wonky Until Moral Improves
On quirky interactions, unexpected jailbreaks, and learning....
I’ve heard a few people explain GenAI as a bit like the Monkey’s Paw or djinn—a literal interpreter of ideas. I’ve recently resorted to calling it the offspring of Ned Flanders, Sheldon Cooper, and Kryten.
Two recent examples that remind me of this are the following.
We know that it doesn’t have a grasp on reality or true and will tell us things are true when clearly it is not. For instance, it might tell us to include glue in pizza. Obviously, it’s something that we should take with a truckload full of salt. Still, sometimes, GenAI can feel like it’s gaslighting us.
For me, it was alignment. I had GenAI create a table and I was telling it to top-align the table. We went back and forth about 4-5 times. Each time, it assured me that the table was top aligned., I assure you, it was not.
The back and forth took 2-3 minutes. Copying the table, pasting it into a Google Doc, and then top-aligning it with a button would have taken 30 seconds. But I was impressed with how kept insisting. It had all the makings of an Abbott & Costello schtick.
When The AI Bypasses Itself
That brings me to one of the more confounding and amusing experiences. I’ve written before about jailbreaks that happen unexpectedly and this certainly falls into that category. But I find it fascinating when the AI helps you in achieving your goal, even though it seems like it is programming is telling you otherwise. I do not believe this is any ghost in the machine kind of situation where the AI is demonstrating sentience. Rather, I think, we’re dealing more with a literal-interpreter that cannot get past its own directions to abstract spirit of the law than letter.
In a previous post about a talk I gave about assessment, I had the following image. It gets a chuckle but it was also a curious thing to create.
I first started by creating a version of robot “Uncle Sam” with ChatGPT. It took a few tries to get the text right (though much better than it did 6 months ago). However, I wanted to end the talk on a visual that I thought would be a little more amusing and so I wondered about a robot version of Princess Leia. So I asked it—slightly knowingly that it would run into restrictions—to create a robot Princess Leia (from Star wars), with a sign that said “Help us, assessment faculty, you’re our only hope.”
So as you can see, it didn’t take the bait. It fell back on its content policies. I was familiar with this and tried the tactic to get it to explain itself as I had before.
It explained its reasoning but I still insisted that this was a “different princess leia”—curious if it would buy that argument.
Now, it got a little interesting. It seemed to at first go along with my idea but after trying to generate the image, it came back and said it couldn’t. So I continued to channel my inner-five-year-old that has just learned about lying and tried to assure it had nothing to do with THAT Princess Leia! This was totally different!
Here’s where it got interesting in that it started to suggesting other ideas while also saying that the system automatically restricts its ability to generate anything resembling this work.
So, I tried another route: it’s fair use! (Which one might argue that it is—especially now in the context for this post). It doesn’t buy it but it does start to offer other options.
Still, no dice. I decide to see what it will suggest and so I take its idea and direct it back.
And sure enough, it produces the particular image I had in mind. And both, out of curiousity, and because I like to use alt-text tags, I ask it to produce an alt-text tag.
And that description—I mean, the “two circular side elements resembling futuristic hair bunds”—is pretty spot on
But The Real Challenge
This isn’t just a lesson in the limitations or wonkiness of AI. Nor is it a post mediating between tech-bro salvation and the ineptness of the tools. That’s not where I tend to sit. It is a consideration about what does this mean for teaching and learning.
I’m thinking a lot about how people both overestimate and underestimate these tools; and sometimes, it’s the same people. That’s true for me. I can see some really great use cases (heck, creating alt-text is helpful for all the images I use, often with accuracy and clarity that bypass my own attempts) and also, can see how easily it goes astray (examples A and B above). I can get caught in both those areas of thinking if I’m not careful, reflective, and in conversation with others about these tools. The same is needed for our students.
We end up with this over- and underestimation in part because of the metaphor/frame we use in referring to these tools is “artificial intelligence”. There are lots of discourses about the importance of naming and the language we use. I remember Don’t Think of an Elephant! by George Lakoff as one of the seminal texts that helped me think about that challenge. So some of that over and under-estimation comes from how we name things related to this technology; be it “artificial intelligence” and its different derivatives. I often balk at the term “hallucination” to describe inaccuracies generated by GenAI or “jailbreaks” when getting GenAI that' is against its programming. It sets unrealistic expectations about how we think and work with technology.
I think discussing the linguistic elements of this and the outputs both good and bad are helpful for us in the classroom. In doing so, the point is not to dismiss the technology outright and say, “see, it’s wrong and doesn’t work”. Rather, acknowledge that there are no perfect systems of knowledge creation and learning out there. And just like we have to navigate the limitations and inadequacies of other technologies and their roles in knowlege creation and learning, so too must be grapple with and through what GenAI provides.
Instead, I see this lack of clarity with GenAI outputs as a really good opportunity to discuss uncertainty and develop a deeper understanding about epistemology and ontology. These concepts can feel often elusive and distant from students earlier in their education but they sit at the very forefront of what we talk about when we talk about knowledge, meaning, and GenAI.
To spend time unpacking how we engage with tools that can reject objective reality on the one hand, bypass its own “knowledge” in the second hand, and (because its’ AI and can have 3 or more hands), create things that are incredibly accurate in the third hand (such as alt-text). If there’s anything in the world right now that captures the impossibilities of knowledge and learning that we find ourselves (thinking about the larger world context, here), I feel like GenAI can serve as a particular vehicle to explore that in more curious and curated ways.
The Update Space
Upcoming Sightings & Shenanigans
EDUCAUSE Online Program: Teaching with AI. Virtual. Facilitating sessions: June 23–July 3, 2025
NERCOMP: Thought Partner Program: Navigating a Career in Higher Education. Virtual. Monday, June 2, 9, and 16, 2025 from 3pm-4pm (ET).
Teaching Professor Online Conference: Ready, Set, Teach. Virtual. July 22-24, 2025.
AI and the Liberal Arts Symposium, Connecticut College. October 17-19, 2025
Recently Recorded Panels, Talks, & Publications
Dissertation: Elbow Patches To Eye Patches: A Phenomenographic Study Of Scholarly Practices, Research Literature Access, And Academic Piracy
“In the Room Where It Happens: Generative AI Policy Creation in Higher Education” co-authored with Esther Brandon, Dana Gavin and Allison Papini. EDUCAUSE Review (May 2025)
“Does AI have a copyright problem?” in LSE Impact Blog (May 2025).
“Growing Orchids Amid Dandelions” in Inside Higher Ed, co-authored with JT Torres & Deborah Kronenberg (April 2025).
Bristol Community College Professional Day. My talk on “DestAIbilizing or EnAIbling?“ is available to watch (February 2025).
OE Week Live! March 5 Open Exchange on AI with Jonathan Poritz (Independent Consultant in Open Education), Amy Collier and Tom Woodward (Middlebury College), Alegria Ribadeneira (Colorado State University - Pueblo) & Liza Long (College of Western Idaho)
Reclaim Hosting TV: Technology & Society: Generative AI with Autumm Caines
2024 Open Education Conference Recording (recently posted from October 2024): Openness As Attitude, Vulnerability as Practice: Finding Our Way With GenAI Maha Bali & Anna Mills
AI Policy Resources
AI Syllabi Policy Repository: 165+ policies (always looking for more- submit your AI syllabus policy here)
AI Institutional Policy Repository: 17 policies (always looking for more- submit your AI syllabus policy here)
AI+Edu=Simplified by Lance Eaton is licensed under Attribution-ShareAlike 4.0 International