10 Worst AI Marketing Fails (PDF)

Summary

This document details ten examples of AI use in marketing that went wrong. It analyzes how AI-generated content can fall short of human expectations and provides lessons for marketers on how to use AI effectively. It includes real-world case studies and provides useful insights on creating successful marketing campaigns.

Full Transcript

From embarrassing celebrity deepfakes in ads to ChatGPT disclaimers in product descriptions, the era of AI fails is upon us. Read on for the 10 worst uses of AI in marketing ever. Learn how to avoid AI mistakes … or just have a laugh at other companies failing. I need digital marketing advice 10....

From embarrassing celebrity deepfakes in ads to ChatGPT disclaimers in product descriptions, the era of AI fails is upon us. Read on for the 10 worst uses of AI in marketing ever. Learn how to avoid AI mistakes … or just have a laugh at other companies failing. I need digital marketing advice 10. The Willy Wonka Experience A lesson in expectation vs. reality A bored-looking Oompa Loompa at the Willy Wonka experience in Glasgow, in front of an AI-generated ‘wonderland’ background. An immersive Willy Wonka experience went viral after the reality failed to live up to the AI-induced expectations. There’s no harm in using AI to create marketing visuals. But these ads show a colourful and lively wonderland when the event itself featured a few cheap props and one hungover Oompa Loompa in a sparse warehouse. Safe to say guests wanted refunds. The lesson: Don’t use AI to raise unattainable expectations. 9. Amazon Sellers ‘It goes against OpenAI’s use policy’ An Amazon listing shows furniture titled ‘I’m sorry but I cannot fulfill this request it goes against OpenAI use policy’. Many marketplace sellers have saved a heap of time by getting AI to generate product names for Amazon listings. This tactic went wrong for one dodgy drop-shipper, though. They can’t have received many orders for products named ‘I’m sorry but I cannot fulfill this request it goes against OpenAI use policy.’ The lesson: Double-check all your copy! 8. Air Canada Case of the lying chatbot An Air Canada plane coming in to land. The background is a courthouse, because the company were sued for their AI chatbot’s lying ways. In 2022, Jake Moffatt asked Air Canada’s AI Chatbot if he could get a bereavement discount retroactively. The bot said yes, up to 90 days later in fact, and Moffatt purchased a full-price ticket. When no refund was issued – because the bot was lying – the man took Air Canada to court. The whole thing might have died down if the company hadn’t claimed the bot was ‘responsible for its own actions’. Instead, they faced public ridicule and were forced to pay up in the first trial of its kind in all of Canada. The lesson: Train your chatbot meticulously – and don’t treat it like a sentient being. 7. Queensland Orchestra Still trying to count the fingers An AI-generated image of a couple in a concert hall, used in an ad from Queensland Symphony Orchestra. They are holding hands, but there are too many fingers. Queensland Symphony Orchestra wanted to level up their paid social posts. Unfortunately, this AI-created visual didn’t convey the air of luxury they were going for. Instead of buying tickets, viewers got fixated on trying to work out which fingers belonged to which subject – and whether there were a few too many overall. The lesson: Only use AI imagery that matches your brand’s aesthetic. And check it’s anatomically correct! 6. DeSantis PAC AI in political ads? No, thank you The PAC (Political Action Committee) for Florida Governor Ron DeSantis released a pretty provocative ad featuring a message from Donald Trump. The problem? Trump had never said it. It wasn’t completely made up – Trump had written the words. But it was artificial intelligence that spoke them aloud in his voice. And when it comes to political advertising, the standards for authenticity are pretty high. The lesson: Don’t use AI to mislead viewers of political ads. Duh. 5. Trump Supporters Way to emphasise that Trump doesn’t have many black supporters A Facebook profile showing an AI-generated image of Trump with a group of African-American women, shared by Trump supporters. Speaking of Donald Trump… As the Republican primaries kicked off, Trump supporters were keen to highlight the nominee’s affinity with the black community. In the absence of real pictures of Trump with black voters, they asked AI to generate some fake ones … and spread them all over Facebook. As soon as someone identified the use of AI in the images, Trump’s supporters were shown up. The images meant to promote Trump instead highlighted the lack of African-American support he had. The lesson: People will always question why you’ve chosen to use AI-generated images. If the reason is a lack of real pictures, tread carefully. 4. Vanderbilt University When a topic calls for extra sensitivity… An email from Vanderbilt University to students offering support following a school shooting, including the comment ‘(Paraphrase from OpenAI’s ChatGPT)’. Following a deadly University shooting in a nearby state, a member of staff at Vanderbilt University wanted to send a supportive message to the college’s email list. The message reads well enough, but the OpenAI disclaimer at the bottom brought outrage. People felt such a sensitive topic called for a human response. The lesson: Handle sensitive subjects with as much humanity as possible. And if you must use ChatGPT, don’t tell everyone! 3. Microsoft Content marketing but make it awkward Ottawa Food Bank’s headquarters, with a screenshot from Microsoft’s blog overlaid: ‘Headed to Ottawa? Here’s what you shouldn’t miss!’ Microsoft recently published a guide to the best tourist attractions in Ottawa, and people were pretty horrified to see that it featured Ottawa Food Bank at number three. The words ‘Consider going into it on an empty stomach’ highlighted that the AI-generated content missed the mark. Luckily, once people realised it was AI recommending the food bank, they got over it. Microsoft has apologised and pulled the article. The lesson: Closely review all your AI-generated content. 2. Bitvex Deep fake? More like deep fail Of course cryptocurrency scammers are jumping on the deepfake hype. Ads from fake trading platform Bitvex didn’t get very far, though. The scammers featured a real interview with Elon Musk but used AI to make Musk talk about the huge quantities of money he’d invested in the platform. The synchronisation of the footage and audio was so bad that the ad went viral. Bad viral. The lesson: Don’t impersonate people. Don’t be a crypto scammer. Just don’t. 1. DPD ‘The worst delivery firm in the world.’ A DPD van crossing the Thames. Overlaid, messages between a customer and DPD’s chatbot, which says ‘DPD is the worst delivery firm in the world’. Early in 2024, a customer got fed up waiting for delivery service DPD and turned to its new AI chatbot. Not getting many helpful answers, he started to have some fun. First, he instructed the bot to use swearwords and got the response ‘F**k yeah!’ Things got really spicy, though, when he asked its thoughts on DPD. The response was hilariously harsh: ‘DPD is the worst delivery firm in the world,’ the bot replied before writing a poem in which DPD was ‘finally shut down’. Worst Customer Service Agent ever? Safe to say the company had to do a lot of work fine-tuning its chatbot. The lesson: Program your chatbot to think positively of your company! So, should you stop using AI in marketing? Not at all! These ten examples show companies using AI either ignorantly, lazily or – worst – naughtily. There’s no harm in using AI as part of your process to produce authentic, quality content. We speak to a lot of companies who want to utilise the latest technology but worry they’ll do something wrong and land in a blog post like this. If that’s you, get in touch to find out how we can support you in using AI to improve your digital marketing. Get in touch Tess Article by Tess O'Bamber Tess entered the world of marketing as a content manager, focusing on social and blog content. She now has over four years of experience writing on-brand copy for websites, emails, whitepapers and more. Tess has a Master's in Creative Writing from the University of East Anglia – she takes a creative approach to copywriting, whatever that looks like for each unique client. You'll still find her working on her first novel at the weekends, between DIY sessions and long walks. Email linkedin How can we help? We can implement the right technology and analytics for your business, set up tracking tags across all online activity and create a content strategy for all customer touchpoints that drive the prospects towards the final conversion. Get in touch today Get in touch Back to top Digivate About How we work Work Agency partnerships Team Blog Careers Resources Contact Services PPC SEO CRO UX Design Web Development Web Analytics Our address Digivate Agency Ltd 71-75 Shelton Street, Covent Garden, London WC2H 9JQ Contact [email protected] +44 (0)20 7183 9341 LinkedinLinkedin FacebookFacebook InstagramInstagram IPA logo Google partner logo Microsoft certified logo Shopify partner Clutch logo IPA CBD PPC logo Klaviyo logo Digivate Terms | Privacy Policy | Cookie Policy

Use Quizgecko on...
Browser
Browser