Do You Think That Little Chat Bubble on a Website Deserves Your Attention?
I just wanted to place an order. Simple right?
Landed on a website. Saw the chat bubble. Clicked it.
"Hi! How can I help you?"
I explained what I needed. The response came back and came across helpful, friendly and seemingly human.
I had an issue with the order form. My mobile phone number wasn't being accepted in an order field, so I asked for help via the chat bubble. Nothing complicated.
I got a reply directing me to other means to have the issue resolved, but the messaging seemed off as was instant. That's when I realised I was never talking with a human.
The chatbot suggested I contact their customer service team and one option was on WhatsApp.
"Great", I thought. I'll get a real person there.
I opened WhatsApp and started the conversation.
At first, the conversation was going well and seemed fine. But then I was getting the same response. Over and Over.
"I'm sorry you are experiencing this issues. Let me know if you need further assistance"
No matter how I rephrased my question, the answer was the same. Then I realised I was never having a conversation with a customer service person. The whole entire time I was talking to another AI bot. And at no point, not on the website, not on WhatsApp, did anyone tell me I was chatting with an AI bot and not a real human being from customer service.
I felt so frustrated when I realised this and by the whole process as all I had wanted to do was place an order, simple.
In my frustration, I reflected back and noticed no disclaimer. Not one statement saying "You're chatting with an AI assistant". No transparency. No notices. There was nothing. Just 20 minutes of my time wasted, looping through automated responses that couldn't actually help me.
So I left and brought from their competitor.
For those unfamiliar, an AI chat widget is that small chat bubble you see on websites usually in the bottom right corner. You click it, type your question, and an AI (not a human) responds. It's like texting a smart assistant that helps with customer support, finding information, or navigating the site.
The problem? Most customers don't realise they're talking to AI, not a person.
What Bothered Me Most
It wasn't that the company used AI chat bots. It was that they designed the experience to make me think this wasn't the case. It felt like that they wanted me to feel like I was talking to one of their customer service team members....so from my experience they pretended.
The chat bubble felt like I was engaging with a human. The responses read like a human wrote it as I also saw the typing indicator active before each response was received. Even the WhatsApp messages read like they were from a human. But it wasn't a human on the other end, and the business I was engaging with never told me. As a customer, the customer service experience felt deceptive.
As someone who's spent 23+ years fixing operational problems, I immediately thought "Do they even realise the gaps and risk they have created here".
Most Businesses Don't See
Because here's what happens in that interaction I had experienced:
I shared information thinking I was talking to a person
The AI couldn't resolve my problem but kept pretending it could
I wasted 20 minutes getting increasingly frustrated
I left and have told my story to others just like you who are reading this now
Their competitors got my business instead
But it could have been worse.
What if I shared other details like my credit card? My address? A compliant about a previous order?
Here's what you probably don't realise. The entire conversation would have been recorded, processed by a third-party AI systems, and stored (note: some third-party agreements do have no AI training and storing of conversations in place, but as a customer you don't know).
Without me knowing. Without my consent. Without transparency about who processes the conversation and associated data, where it's stored, how long it's kept, and who has access to it. From a business perspective:
Lost sale (I went elsewhere)
Brand damage (I'm telling this story and to others)
Data liability (What happens to conversation data?)
Potential exposure (no transparency about AI usage at any stage of the customer service process. No transparency whether my information is being processed by a third-party provider. No transparency that my information would be recorded and stored for who know how long.)
All because the company added a chat widget focused on delivering "better customer service" without thinking through the operational foundations needed to use it responsibly.
This Isn't Rare
You have most like experienced something similar yourself. Maybe you have:
Asked a question and got an irrelevant answer
Been stuck in a loop with same automatic response
Realised too late in the conversation you were never talking to a human and gave too much information about your situation like experiencing financial hardship, health information that is really very personal, or uploaded and shared a copy of your ID or other sensitive information while all the time thinking you are engaging with a human from customer service.
Felt frustrated by a "helpful" AI that couldn't actually help at all.
As a customer, it's annoying.
As a founder or business leader, it's a problem you might not even know you have.
When Chat Widgets Go Really Wrong
My story ended with frustration and the business lost a sale. But it can go much worse.
I discovered through research Air Canada learned this the hard way.
A customer asked their website chatbot about bereavement fares. The chatbot responded by stating "Buy a full-price ticket now and then request a refund later with proof of death". The customer did exactly that.
Air Canada refused the refund. "Our chatbot was wrong. That's not our policy." The customer took them to tribunal. Air Canada lost. The ruling: You're responsible for what your chatbot says, even if it's wrong, even if you didn't write it, even if it's a third-party service. That chatbot conversation cost them money plus made international headlines. (The Guardian | Forbes)
In another case, a delivery company's chatbot was manipulated by a frustrated customer into swearing (yes swearing) and calling the company "the worst delivery service in the world". The conversation went viral on social media. The company had to disable the entire chatbot.
Image sourced from Instagram
The Pattern
Three stories. Three different problems. One common thread: Businesses added AI chat widgets without operational foundations to use them responsibly:
No clear guidelines on what the chatbot can say
No transparency about when customers are talking to AI
No escalation path when AI can't help
Appears to be no monitoring of what's actually happening in conversations
No consent process for recording and storing conversations
Most likely there is no accountability for what goes wrong
Most founders and business leaders think they have thought it through. They tested the chatbot. It seemed fined. They launched it.
But they never asked (May not even have documented):
What if the solution gives wrong information?
What if customers don't realise it's AI?
What is the impact on our business overall?
What if the AI chat bot can't solve the customer's problem but won't admit it?
What if a customer intentionally overshares information?
Are we getting proper consent to record and store conversations?
Do customers know their conversations and the data they are sharing is being processed by third parties?
Who's monitoring this? Who is responsible - the company, the third-party provider, who?
Those questions above? They are operational foundations that should be thought about before launching for members of the public to use.
Why Businesses Add AI Chat Widgets Anyway
Despite the risks such as the two cases mentioned, chatbots are appearing on more websites every day and including in emails and WhatsApp, the benefits for businesses are:
Respond instantly to customer questions
Provide help 24 hours a day
Reduce repetitive support requests
But also to capture leads before visitors leave the website
Scale customer service without scaling headcount
For customers, like myself and you, the benefits is that instead of waiting hours or sometimes even days for a reply, customers can get answers immediately.
The small chat bubbles removes friction from the customer service experience. The technology itself isn't the problem. The problem is implementing the technology without operational frameworks to use it responsibly.
What Founders and Business Leaders Need to Know Before Adding a Chat Widget
Image sourced from All-in Consulting Newsletter - Substack, 2024
If you are considering an AI chat widget or you've already activated one already, recommend reading further so your business doesn't end up in headlines like Air Canada, or goes rogue like the delivery company, or hands potential business to competitors
Know what the AI Chat Bot is saying
Your chatbot represents your brand every time someone engages with that bubble, think of it as another employee.
Monitor conversations regularly, especially in the first month. If it starts giving wrong responses, you need to know before your customers start sharing screenshots on social media like what happened with the delivery company.
Ask yourself (and your team): When was the last time you reviewed actual chat transcripts?
Be transparent about AI
Customers deserve to know when they're talking to AI, not a human, whether on your website, in WhatsApp, voice chat, or anywhere else.
It doesn't have to be complicated. A simple notice: such as for example - "You're chatting with our AI assistant. Need a human? Click here."
The lack of transparency is what frustrated me most in my experience, and I'm sure other customers experience the same. What it does is it damages trust.
Get proper consent
Before the conversation starts, inform customers that the conversation is being recorded, how and where it's stored, why it's being recorded, and who processes the data (especially if third-party providers are involved). And get their consent. This is good practice and it's law in many jurisdictions.
Don't let the AI make promises you can't keep
The Air Canada case happened because the chatbot made commitments the company didn't honour.
Make sure your chatbot (just in the same way as an employee) can only provide clear and established information, not make binding promises about refunds, pricing, or policies.
Ask yourself (and your team): If your chatbot told a customer they could get a full refund, would our company honour it?
Think about other situations where a promise made by AI could be broken.
Have a human backup
AI should help your team, not replace them entirely.
Customers need a way to reach a real person when the chatbot can't help, not get stuck and be in endless response loops. This is a safety net when things go wrong.
In my experience, I couldn't find a way to reach a human, so I left.
Test it like a skeptical customer would
Before you launch the AI chatbot widget, try to break it.
Ask confusing questions. Try to get it to say something wrong. See what happens when someone shares sensitive information. Better you find the problems than your customers.
Remember: You may be responsible for what the AI Chatbot says
Air Canada learned this the hard way.
Even if the chatbot is powered by a third-party service, even if you didn't write this response, even if it's "just AI" - you may own the consequences.
If it gives bad advice, makes a promise, or says something damaging, that may be on you.
The Real Question
The question isn't "Should we be using an AI chatbot widget?", the question is "Does our company have the operational foundations to use AI responsibly?"
The best technology in the world won't help if:
Your team doesn't know who's responsible for monitoring the conversations
There is no clear documented guidelines about what it can and can't say
There's no process for handling problems when they arise
No one's checking if it's actually helping or hurting the business
Customers don't know when they talking to AI
You're not getting proper consent to record and process conversations
If you can't answer any of the above, this is an operational foundation problem, not IT.
AI solutions such as AI chatbot widgets promise to help your business scale. But without clear operational systems (like guidelines for employees), it's just scaling chaos and most likely without businesses realising it.
From Reactive Hope to Proactive Confidence
Here is the transformation most founders and business leader need and most likely haven't thought about:
Before establishing operational foundations:
"I hope our chatbot doesn't say anything wrong"
"I/We think customers know it's AI...right?"
"If something goes wrong, we will figure it out?"
"I think someone on the team is monitoring it or at least our third-party providers will monitor it for us."
After thinking about and designing operational foundations:
"I/We know what our chatbot can and can't say as there are clear guidelines in place just in the same way we have for our employees."
"Customer's know when they are talking to AI, we're transparent about it in our privacy policy, on our website, in a WhatsApp message, in our emails, in our voice chat, and everywhere else."
'We've informed our customers that conversations are stored. We've explained how, where, and why. And we've obtained their consent."
"If something goes wrong, we know exactly was to do, we have practiced with role plays and covered scenarios like what happened to the delivery company and Air Canada."
"We review chat transcripts weekly, so nothing surprises us."
"We make confident operational decisions as we scale the use of AI chatbot widgets beyond just the business website, but into email, WhatsApp, and voice chat."
The Opportunity
When implemented thoughtfully, AI chat widgets can transform how business interact with customers. They can reduce friction, improves response times, and make support more accessible.
But that little chat bubble also carries responsibility.
Every time a customer types a message into that window, they're doing something simple but significant.
Customers are trusting your business to respond correctly and protecting that trust is what good technology implementation is really about.
Many founders and business leaders adopt AI tools before busines operations are ready.
Hyplon helps founders and business leaders build the operational foundations needed to adopt AI safely through the 5 Foundations Frameworks.
7 Checks Before Adding an AI Chatbot Widget to Your Website, WhatsApp and other tools
Before launching an AI chatbot on your website, founders and business leaders should review the following operational checks.
Transparency
Customers should know whether they are speaking with an AI assistant or a human.
Privacy and Data Handling
Ensure the chatbot clearly explains how customer information will be stored, processed, and used.
Escalation to a Human
Customers should be able to transfer to a human when the chatbot cannot solve their issue.
Guardrails and Behaviour Controls
AI chatbots need boundaries to prevent inappropriate or misleading responses, just in the same way as an organisation has a guidebook for employees.
Customer Frustration Monitoring
Track when conversations fail or when customers abandon the interaction.
Legal and Compliance Risks
Businesses should review risks around consumer law, misleading advice, and privacy obligations (including security for data handling).
Operational Ownership
Someone in the business must be responsible for maintaining and reviewing chatbot behaviour over time and regularly, not just adhoc or annually.
Want's Next
Want to see where your operational gaps are before they become a problem?
Benchmark your operational foundations and discover the gaps that could slow your growth before they become expensive problems and create burn out.
Or learn more about building stronger operational systems with the AI-Ready Business Program.
You can also read my free book 'Don't Let Your Business Get Cooked' or grab a hard copy of Marnie's book , it's available on Amazon.
Written by Marnie McLeod
Marnie McLeod has spent 23+ years turning around deficit businesses, sitting on boards, and improving business operations. She teaches The 5 Foundations Framework to help founders scale with AI confidently.

