Ai&E: the chatbot ‘GP’ has arrived — and it operates outside the law
Vendan Ananda Kumararajah
- Published
- Opinion & Analysis

Wearables, chatbots and digital assistants now provide guidance that once came from nurses and GPs. Millions of people act on this advice when deciding whether to wait, book a GP appointment or go to A&E, and this carries real implications for risk, cost and triage. The absence of clear duties, liability and oversight has created a new advisory layer that sits outside the medical rules the system relies on, warns Vendan Kumararajah
People now ask AI systems about health in the same way they once asked doctors, nurses or pharmacists. A mother with a child who keeps coughing may type symptoms into a chatbot in the kitchen while making breakfast. A young man who worries about hair loss might ask an AI assistant during his commute. Someone who feels chest discomfort late at night might talk to a voice agent on a smart speaker because it feels easier than bothering the NHS. Wearable devices already nudge people about their sleep, their stress and their heart rate, and many of these tools speak in natural language about what to do next.
Health advice has crept into homes, offices and pockets without the need for appointments and without long waiting times. This has changed the structure of healthcare because the advisory role has shifted into tools that sit outside the rules and norms that have traditionally governed medicine.
The advice that AI provides directly steers people how to behave. A woman who has pelvic pain and is told it may settle on its own will usually wait, plan a GP visit for next week or try home remedies. Another woman with the same pain who is told it could be serious may call 111 or go to A&E that same evening. These forks in decision-making shape demand, cost and risk within the health system. Doctors and nurses train for years to understand this power. They hold licences, professional duties and liability because their advice carries real consequences. But AI agents influence the same decisions but do so without the moral, legal and institutional guardrails that surround clinical guidance.
Current regulation focuses on data privacy, transparency and recall procedures for unsafe products. These efforts matter, but they do not answer a basic question about responsibility. When a chatbot or wearable gives advice that alters behaviour around pregnancy, medication, psychiatric symptoms or care for vulnerable relatives, for example, who owns that advisory outcome? Developers place disclaimers on their products that push judgement back onto the user. The average person does not read probabilistic outputs as mere probabilities but as expert guidance that should be followed. Clinicians then meet patients whose choices have been shaped before the appointment, yet have no insight into the digital pathways that produced those choices. Regulators and courts tend to appear only after failures become visible, long after the moment where early stewardship could have prevented harm.
Good governance needs to move to the point where advice is given. At that point, clear rules should define what an AI agent may handle alone and what must trigger escalation to a human clinician. Chest pain, suicidal ideation, pregnancy complications, signs of stroke and medication misuse are not edge cases; they appear daily in primary care and emergency departments, which means any system that advises on them should face defined obligations. Certification could recognise these agents as active participants within the healthcare system rather than as neutral information tools.
Other sectors already understand advisory power. Banks cannot recommend complex financial products without suitability checks, for instance. Airlines require escalation from pilots to ground control when instruments behave unpredictably. Pharmaceutical companies must run post-marketing surveillance with real-time reporting duties. AI healthcare deserves rules of similar sophistication.
AI models change all the time. Updates alter their tone, their confidence levels and their triage suggestions. These changes look subtle at first and usually become visible only when patterns emerge across large populations. Approval regimes that sign off a product once and then step away cannot manage technology that rewrites itself every few weeks.
Continuous monitoring, longitudinal audits and performance tracking would provide a realistic foundation for safety. Without such structures, society ends up running a vast uncontrolled experiment on millions of people who have no way to understand the risks, the incentives or the absence of governance around these advisory conversations.
Europe already has institutions that understand advice, risk and public interest. Its medical device framework, its pharmaceutical rules and its financial supervisory systems already deal with high-stakes guidance and cross-border enforcement. The task ahead is not to invent an entirely new philosophy of responsible AI but to adapt existing machinery to a new actor that sits in a grey zone between clinical care and consumer technology. Influence needs liability, function needs certification and patients need protections at the point where behaviour is shaped.
Without it, healthcare across Europe and beyond will develop a hidden advisory layer with no clear owner, no monitoring and no accountability. That would erode public trust in medicine and technology alike, and it would place a burden on patients that they never agreed to shoulder.

Vendan Ananda Kumararajah is an internationally recognised transformation architect and systems thinker. The originator of the A3 Model—a new-order cybernetic framework uniting ethics, distortion awareness, and agency in AI and governance—he bridges ancient Tamil philosophy with contemporary systems science. A Member of the Chartered Management Institute and author of Navigating Complexity and System Challenges: Foundations for the A3 Model (2025), Vendan is redefining how intelligence, governance, and ethics interconnect in an age of autonomous technologies.
READ MORE: ‘Robots can’t care — and believing they can will break our health system‘. Artificial intelligence is being hailed as the next frontier in healthcare but as broadcaster and disability rights advocate Matthew Kayne writes, empathy cannot be automated. Real care exists in human presence, in the moments of understanding that only people can offer.
Do you have news to share or expertise to contribute? The European welcomes insights from business leaders and sector specialists. Get in touch with our editorial team to find out more.
Main image: Karola G/Pexels
RECENT ARTICLES
-
First Adolescence, now Inside the Manosphere. How do we protect boys from misogynistic alpha male influencers? -
NATO reluctance signals limits on U.S. leadership -
Iran, nuclear proliferation and the hard choices facing democracies -
When AI customer service fails, don’t blame technology — it’s leadership at fault -
SUCCESS London conference highlights challenge of life after cure for brain tumour survivors -
A new generation of disability rights leaders is reshaping Europe -
Trump hasn’t broken America — he’s exposed what it really is -
AI is rewriting Europe’s networks from the inside out — and the continent isn’t ready -
Europe’s new gender strategy may be solving yesterday’s problems -
Why Britain still needs reporters in the courtroom -
Rivers run deeper than we think -
Spain’s rocket builder just landed €180 million — and Europe’s case for space sovereignty just got harder to ignore -
Why jobs and housing must be solved together to deliver real disability inclusion -
The new gender divide is already reshaping Europe’s future leaders -
The Arctic’s unfinished cold war -
Highway robbery: how the UK’s post-Brexit electric car policy blew a fuse -
Nokia built the brains for the AI network revolution — so why is American capital leading the charge? -
What the UK SEND reform whitepaper means and what it might take to deliver it -
Europe cannot call itself ‘equal’ while disabled citizens are still fighting for access -
Is Europe regulating the future or forgetting to build it? The hidden flaw in digital sovereignty -
The era of easy markets is ending — here are the risks investors can no longer ignore -
Is testosterone the new performance hack for executives? -
Can we regulate reality? AI, sovereignty and the battle over what counts as real -
NATO gears up for conflict as transatlantic strains grow -
Facial recognition is leaving the US border — and we should be concerned


























