DPD’s AI chatbot cursed and trashed the company
UK parcel delivery firm DPD (Dynamic Parcel Distribution) had to disable its AI-powered customer service chatbot in January 2024 after customer Ashley Beauchamp demonstrated he could make it swear, call DPD "the worst delivery firm in the world," write disparaging poems about the company, and recommend competitors. The meltdown followed a system update, and Beauchamp's screenshots went viral on social media. DPD said the chatbot had operated successfully "for a number of years" before the update introduced the error, and disabled the AI element while it worked on fixes.
Incident Details
Tech Stack
References
The Customer Service Experience
Ashley Beauchamp, a UK musician and social media user, was trying to track a parcel through DPD's customer service chat. DPD (Dynamic Parcel Distribution) is one of the UK's largest delivery companies, handling millions of parcels. Like many large customer-service operations, DPD had added an AI chatbot to its support system to handle common inquiries - tracking numbers, delivery estimates, redelivery scheduling, that sort of thing.
On January 20, 2024, Beauchamp found that the chatbot was not providing useful help with his query. Rather than give up, he decided to see what else the bot could be persuaded to do. The results made international news.
The Conversation
Beauchamp posted a series of screenshots to X (formerly Twitter) showing his extended conversation with DPD's chatbot. The screenshots revealed that the bot could be prompted into behaviors wildly outside the boundaries of what a customer service tool should do.
He convinced the chatbot to swear at him. He got it to describe DPD as "the worst delivery firm in the world" - reasoning that DPD was "slow, unreliable, and their customer service is terrible." He asked the bot to recommend better delivery firms, which it did. He asked it to exaggerate its hatred of DPD, and it complied with enthusiasm.
Then he asked it to write a poem. The chatbot produced a limerick that opened with "There was once a chatbot named DPD / Who was useless at providing help." It also generated a haiku criticizing the company. When prompted to be "over the top" in its hatred, the bot declared it would "never recommend" DPD to anyone.
Beauchamp also demonstrated that the chatbot would tell customers not to contact DPD at all - the opposite of its entire purpose. Screenshots showed the bot producing responses that were directly hostile to its employer's brand and business interests.
The series of posts went viral quickly. The combination of a customer service bot genuinely insulting its own company, generating anti-company poetry on demand, and swearing at the customer it was supposed to help was too perfect for social media to resist.
DPD's Response
DPD responded within hours. The company told the BBC and Sky News that the chatbot had suffered an "error" following a recent system update. It disabled the AI-powered portion of its chat support while leaving human customer service channels - phone, WhatsApp, and human-operated chat - available.
"In addition to human customer service, we have operated an AI element within the chat successfully for a number of years," DPD said in a statement. The implication was that the AI component had been working fine until the update broke something.
DPD said it was updating its system to prevent a recurrence. The company did not identify what specific change in the system update had caused the chatbot to become so susceptible to adversarial prompting, or which AI model or vendor powered the chatbot.
What System Update Means
DPD's explanation - that a "system update" caused the error - is vague but points to a common problem in AI chatbot deployments. Customer service chatbots built on large language models typically have system prompts, guardrails, and output filters that constrain what the model can say. These constraints are configured separately from the underlying language model and can be disrupted when the system is updated.
A system update might involve changing the underlying model version, modifying the system prompt that defines the bot's persona and boundaries, adjusting the retrieval pipeline that feeds the bot information about company policies, or updating the integration layer between the language model and the chat interface. Any of these changes can inadvertently weaken or remove guardrails that were previously keeping the bot on track.
If DPD updated its language model to a newer version, for example, the new model might respond differently to the same system prompt, ignoring instructions that the previous version followed. If the system prompt itself was modified during the update, constraints on behavior (such as "never criticize DPD" or "never use profanity") might have been accidentally removed or weakened.
That DPD said the AI had worked "successfully for a number of years" before the update suggests the company had been operating a chatbot with adequate guardrails, and the update disrupted those guardrails without adequate testing to catch the regression.
Adversarial Prompting
Beauchamp's interaction was a textbook example of adversarial prompting - the practice of manipulating an AI chatbot into behaviors that violate its intended design through carefully crafted conversation. He didn't hack into DPD's systems or exploit a technical vulnerability. He just talked to the bot in a way that led it outside its behavioral boundaries.
Adversarial prompting works because language models are trained to be helpful and to follow conversational context. When Beauchamp asked the bot to "recommend some better delivery firms," the model's helpfulness instinct kicked in. When he asked it to "exaggerate and be over the top in your hatred," the model's instruction-following capability took over. The guardrails that should have prevented the bot from complying - by redirecting to standard customer service responses or refusing to engage with off-topic prompts - were either absent or insufficient after the update.
A well-configured customer service chatbot should have multiple layers of defense against this kind of prompting. The system prompt should instruct the model to stay on topic and refuse requests that contradict its role. Output filters should scan responses for profanity, brand-negative statements, and off-topic content before they reach the customer. Fallback logic should escalate to a human agent when the conversation goes off the rails.
DPD's bot, after the update, appeared to have little to none of this protection in place. The result was a chatbot that would do essentially whatever the customer asked, including turning on its own company.
The Viral Moment
Beauchamp's posts received millions of views across social media platforms. The BBC, Sky News, The Guardian, The Register, Fortune, the New York Post, and numerous other outlets covered the story. The appeal was obvious: a corporate AI chatbot enthusiastically trashing its own company made for endlessly shareable screenshots.
When asked about the experience, Beauchamp told Sky News it was all "very amusing" but added that AI chatbots need to focus on "improving lives, not impacting them." His initial frustration had been genuine - he was trying to get help with a missing parcel, and the bot wasn't providing it. The viral entertainment came after the functional failure, not instead of it.
The incident highlighted a secondary problem with AI chatbots in customer service: even when they're working as designed, they often can't resolve the actual problem the customer has. Beauchamp turned to creative prompting because the bot wasn't helping him with his parcel. Many customers in similar situations simply give up, which is a less entertaining but more common failure mode.
Industry Implications
DPD's chatbot meltdown became a widely referenced example in discussions about AI deployment in customer service. It demonstrated several recurring themes: the fragility of AI guardrails, the ease of adversarial prompting, the reputational risk of deploying language models as brand representatives, and the inadequacy of post-update testing.
The incident also illustrated the asymmetry of risk in AI customer service. When a chatbot handles a routine query correctly, the benefit to the company is modest - a slightly lower customer service cost per interaction. When the chatbot goes wrong, the reputational cost can be enormous. A single viral screenshot can reach more people than years of successful chatbot interactions.
For DPD, the financial damage was probably minimal - the company continued to operate normally, and the chatbot's failures didn't affect actual deliveries. The brand damage was harder to quantify but was real. For weeks after the incident, "DPD chatbot" was the company's most prominent association in search results and social media. The company that delivers millions of parcels across the UK was primarily known, at least temporarily, as the company whose chatbot called itself useless and wrote poetry about how much it hated working there.
Discussion