Tag Archives: healthcare

AI and the NHS

The next time you speak to an NHS healthcare professional it is quite likely that your conversation will be recorded before being transcribed and summarised for your medical record using AI provided by a private company. Other AI will be reading your hospital records and summarising your clinical letters. If you complain about your treatment, AI will probably be used to write a reply. It will be used to interpret your scans and blood results and plan your appointments. AI in the NHS is suddenly everywhere, all at once.

From £330m deals with Palantir to run the NHS Federated Data Platform to contracts between GPs and ambient scribe vendors, the scope is already too wide to clearly comprehend. Likewise, most clinicians like me who are using it have very little understanding about how it works, how much it costs, what risks it entails or who is profiting. Crucially, we have no idea whether it can improve quality, safety, equity, access or the experiences of staff or patients. We’re like kids with a new toy, too enthralled to ask questions.

‘The quintuple aim for improving healthcare includes: improved health outcomes, enhanced patient experience, reduced costs, professional wellbeing, and health equity. Health equity is defined as ‘the state in which everyone has the opportunity to attain their full health potential and no one is disadvantaged from achieving this potential because of social position or other socially determined circumstances’  (Nundy et al 2022)

Like the laundry and the dishes, I want AI to take action on the social determinants of health, but all it wants to do is listen to my consultations. Perhaps the quintuple aim lacks profitability.

NHS England are clearly enthralled by AI (they are mostly of a certain age and health status after all) and more concerned with how to allay fears so that they can roll it out as far and as fast as possible [i]. While certain risks are acknowledged including racial profiling, IT capacity, and mistakes (so called ‘hallucinations’), there are few safeguards and they have nothing to say about the human, environmental, or political risks.

There was a surge in enthusiasm in September 2025 when Great Ormond Street Hospital published (‘announced’ is more accurate, since the actual trial has not been published) the results of a trial of AI technology that showed that AST (Ambient Scribe Transcription or AVT, Ambient Voice Transcription) used in outpatient departments meant that clinicians spent almost 25% more time with patients at the same time as having 8% shorter appointment times. In A&E settings 13.4% more patients were seen per shift. Such extraordinary results surely demand transparency over the details of the study and a cautious reception, but instead, ‘AVT will allow us to reimagine healthcare,’ says NHS England’s CCIO (Alec Price Forbes). Forbes told delegates at the HETT leadership summit that AI scribing software is key to the ‘digital by default’ model of care and that the NHS must ‘fundamentally reimagine how care is designed, delivered, and experienced’. He stressed that the problem isn’t whether AVT is technically feasible but how it will be deployed at scale.

I have recently started trialling AVT in my practice – I’m a GP in east London. The AI ‘listens’ to the consultation and then generates a summary of the transcription for the medical record. It is supposed to allow the GP to spend more time listening to the patient and less time looking at the computer screen while producing a more comprehensive and accurate record.

When I started working as a GP I would have 18 patients booked face to face in a half-day session. Some were complex and took longer than 10 minutes, some were straightforward and took less time, and some didn’t turn up. Overall, I would usually run late because of the added complexity of working in a deprived area where almost every physical complaint is mixed up compounded by social problems, long-term mental illnesses, chronic pain, and many patients who didn’t speak English as a first language. In recent years we have moved to a system of total triage where every patient request is screened by a doctor before being given an appointment. Overall, it has been good for continuity for patients with long-term conditions and improves quick access for patients with urgent issues. But it has meant that only complex patients are booked with me and days when I am  triaging can mean making decisions about over 200 patient requests in a single half-day session. Triage has made the work more intense and more complex.

Typing up consultation notes after each appointment is a precious moment to reflect on what was said, what was seen and what was felt, emotionally as well as in a physical examination. It is a judgement about what it all meant. When AVT generates a summary of the transcript of the consultation it is as if it is making a judgment about what to include and what to leave out. But AI can only give the illusion of judgment, it has no morality or empathy. It simply predicts what seems most likely that an average doctor would want to include. What remains ought to have salience – which is to say that the words that are selected to represent the encounter stand for what was most important, at least to the clinician, and hopefully too, for the patient. When I look back at consultations written by AVT I cannot work out what mattered. The salience is missing.

AI predicts that we will ask certain things like, ‘was your sputum blood-stained?’, ‘was there blood in your stool?’, or ‘did your chest pain radiate to your neck or left arm?’ and occasionally adds these details even if you never mentioned them. It’s like having your lawyer edit your consultations.  

We must click a box to confirm that we have reviewed the AVT summary and we can edit it to add our own judgement and humanity, which I do. So far it is entirely indifferent to coding which means filing information under a problem title which links to a clinical code. To be fair so are many doctors, but I am not, so that is something else I must edit. If you want AVT to save you time, which is surely the main point of it, then you need to take the path of least resistance. This must have been what the GOS doctors did, especially given their version of AVT was several generations before mine.

AVT is particularly poor at predictions when patients are distressed, confused, skip from one topic to another and don’t speak English as a first language. Thanks to Total Triage, these are most of my consultations. I am not inclined to keep using it, but we have signed up to another trial from a different provider, so I will keep on investigating. I can see a time when it is mandated for medico-legal reasons where there is a dispute between and doctor and a patient about what is said, although for now at least, the original transcript which is held in UK data centres, is deleted after 30 days. There is an option where I can simply dictate my consultation notes, but using AI to do speech to text is like using a Tesla Cybertruck to drive around your neighbourhood. It’s ridiculously over-powered, over-spec’d, and over-priced for the task.

AI is a solution in search of a problem. We’ve got too much of it, so we’re trying to put it in everything. It has come along so fast, that we didn’t stop to diagnose the problems we really have and figure out ways to solve them. Before AI came along, I don’t remember ever thinking that typing consultation notes was the bugbear of my day. On the contrary, these tiny, reflective, creative acts are enjoyable. I am not alone in suspecting that AI wants the creative work that I enjoy, making my days even more intense and complex than before.

Energy consumption by AI is vast. A Chat GPT query requires on average ten times more electricity than a typical Google search. EOAI 275 Minutes of the UK AI Energy meeting June 2025 state that data centre electricity demand is projected to double by 2030 and nearly triple by 2035, becoming equivalent to the total electricity demand of Japan.

In the US at the current rate of development data centres are predicted to consume 8% of total electricity compared with 3% at present. The December minutes of the UK AI Energy meeting state ‘Senior officials underlined the urgency of clearing barriers to energy supply, the need for rapid and practical delivery, and the importance of acting within tight timescales given rising electricity demand.’

Globally, most of the electricity (outside of China) is going to be produced by fossil fuels. In addition to this existential threat of global warming, the consumption of fresh, potable water for cooling data centres is already having a significant effect on the availability of drinking water, especially in drought prone areas, but also along the M4 corridor in the UK where home building is being cancelled to make way for UK data centres. At the time of writing the escalating conflict in Iran and surrounding countries has led to a surge in energy prices that may continue for years.

Given its environmental impact it is extraordinary that AI is being introduced so widely, without any assessment of whether the benefits are worth the costs, or whether any other technology might be appropriate. To give a simple example, the problems that confront me every day as a GP such as slow/ unreliable IT, fragmented care, waiting times, poor communication between actual people in specialties and departments including hospitals, social care and third sector, difficulty maintaining continuity, staff sickness and disputes, long waits for hospital appointments, chronic pain, mental illness/ distress, poor housing and other effects of deprivation, long referral forms, messy problem lists, medical waste (excessive medical tests and interventions), environmental waste (medications, energy, paper, etc.) are not solved by AI.

Other significant concerns are the impact on staff since we are being encouraged (strongly) to use it to do admin work done by administrators and receptionists. The likely outcome is fewer jobs for healthcare administrators and receptionists and because GPs will have to do more reading, we will likely be more isolated in our rooms in front of our computers.

AI cannot be separated from the handful of US companies that own and control it, or the US government under Trump with whom they are closely aligned, which had a lot to do with the administration’s denial of climate change and gung-ho approach to the energy production and the environment.

Palantir owned by Peter Thiel and Alex Karp has been given nearly £330m worth of NHS contracts and according to a report in the Guardian on 5 Feb 2026, the UK Government has repeatedly blocked attempts by campaigners and MPs to find out details of contracts, including deals made with Boris Johnson and Keir Starmer. The Information Commissioner is investigating a refusal by the Department of Health and Social Care in June to release official reports about Palantir’s NHS federated data platform on the grounds that confidentiality is needed to allow the formation of government policy.

The US Foreign Intelligence and Surveillance Act (FISA) gives law enforcement and intelligence the right to access data anywhere in the world if held by a US company and to date there have been no assurances that NHS data will be protected. In the US Palantir are providing the software for 911 calls and data management for Medicare and patient data has been used to direct ICE agents to areas where there are suspected immigrants who have been aggressively targeted and even murdered.

The British Medical Association has had a sustained objection to the role of Palantir in the NHS and voted in its ARM in December 2025 to scrutinise and halt any further involvement of Palantir in handling patient data. NHS England’s medium term planning framework, published in October, said all trusts should be using FDP core products from April.

A company which is profiting from war, is closely associated with far-right ideologues, that cannot guarantee NHS sovereignty of patient data, and whose contacts are hidden from public scrutiny is being given access to the world’s most comprehensive and valuable patient records. MPs, journalists, campaigners, and the BMA are being ignored and the majority of ICBs and NHS trusts are now using it. Concerns about transparency, mistrust, vendor lock-in, value for money and data security aren’t being taken seriously enough according to Chi Onwurah, a Labour MP and chair of the science, innovation and technology select committee.

The majority of GPs are eagerly and uncritically adopting AI in spite of the BMA’s position. While there are a number of UK companies providing the services for GPs the US tech giants have been ruthlessly assimilating smaller companies and so that may not last long before there are just a few or even one provider. The human, environmental and political risks are enormous and I am very concerned. 

Selected references


https://www.theguardian.com/politics/2026/feb/05/calls-to-halt-uk-palantir-contracts-grow-amid-lack-of-transparency-over-deals

https://youtu.be/YftdpU4ElUQ?si=twlKsRjOFdX1FvYf
https://www.thenerve.news/p/palantir-technologies-uk-government-contracts-size-nuclear-deterrent-atomic-peter-thiel-louis-mosley
https://www.thenerve.news/p/technofascism-us-america-fascism-trump-palantir-peter-thiel-uk-nigel-farage-reform

https://youtu.be/IY1kS0Htk6I?si=9hG7wRTGaA0hlFY0

Strategic litigation against public participation, aka Lawfare, ie, The law protects those it does not bind and binds those it does not protect.

Time saving = work harder http://gosh.nhs.uk/news/implementing-ai-scribing-into-outpatient-clinics-to-enhance-care-and-reduce-burnout/


[i] https://www.gov.uk/government/news/ai-to-be-trialled-at-unprecedented-scale-across-nhs-screening

Confidence and trust in GPs.

Confidence and trust in GPs.

This weekend I went to a ‘nine-night’, a memorial for a Caribbean patient I had known for almost 20 years. I remember vividly the first time we met. She walked into my consulting room and stood right in front of me just as I sat down. Almost six-foot tall and considerably more than a hundred kilos, she leaned over me and banged her walking stick wrapped in silver lamé on the floor, “I don’t know you and I don’t like doctors, especially doctors I don’t know. What you need to know is that I need a doctor who can be comfortable in front of me,” I looked up to see her frowning down at me and I was about to say, “I think I can do that,” when she cracked into a grin and slapped me hard on my shoulder and said, “I’m only kidding! Dr Fuller said I should see you and I trust him! But he’s left. You had better not leave, because if you’re not staying, I’m going somewhere else.”  She wasn’t kidding, especially about needing a doctor that could remain comfortable in her considerable presence. Philosopher Julia Kristeva wrote ‘Powers of Horror: an essay on abjection’ in which she described abjection as having both visceral and symbolic powers. Visceral powers relate to its ability to disgust and repel, forcing people in its presence to recoil, repel or escape. The symbolic powers relate to the ways that abjection transgresses boundaries between life and death, clean and contaminated, what is permitted and what is not. Janet crossed a boundary by ignoring the patient seat in my room and leaning over me. Her body made huge by binge eating and scarred by surgical procedures and self-harm was a battleground where she was both victim and perpetrator, abject even to her. She had met too many doctors who recoiled in her presence, but having been assured by her previous GP that I would be up to the task she wasted no time before testing me. Looking back through her medical record I can see that it wasn’t long after our first meeting that she told me about some of the abject horror of her childhood that helped make sense of her present situation. In recent years when she was housebound she would insist on cooking for me when ever I came to visit. if she was unwell she might postpone the visit for a day or two until she had mustered the strength to prepare something. She bought a portable hob and electric wok so that she could sit on the side of her bed in the lounge and cook up Caribbean fusion dishes with meat, fish, chicken, rice, past and vegetables and always with hot pepper sauce on the side. She was impressed and amused by how much I enjoyed the spices. I would always be sure to skip breakfast and do a training session before work on days when I was due to visit because she would be disappointed if I didn’t eat A LOT. While I was eating, I couldn’t interrupt her and she had my full attention. My mouth was full and my ears were open. We would catch up on her hospital visits and her precarious health, and she would cry and try desperately to get me to understand how she felt, especially after her son died. Empathy came more from shared feeling than from words. I never saw her eat anything herself. After she died I looked through the dozens of consultations that I had recorded and wondered how many people knew as much as I did about her life.

“When you [a GP] are so poor that you cannot afford to refuse eighteenpence from a man who is too poor to pay you any more, it is useless to tell him that what he or his sick child needs is not medicine, but more leisure, better clothes, better food, and a better drained and ventilated house. It is kinder to give him a bottle of something almost as cheap as water, and tell him to come again with another eighteenpence if it does not cure him. When you have done that over and over again every day for a week, how much scientific conscience have you left? If you are weak-minded enough to cling desperately to your eighteenpence as denoting a certain social superiority to the sixpenny doctor, you will be miserably poor all your life; whilst the sixpenny doctor, with his low prices and quick turnover of patients, visibly makes much more than you do and kills no more people.

A doctor’s character can no more stand out against such conditions than the lungs of his patients can stand out against bad ventilation. The only way in which he can preserve his self-respect is by forgetting all he ever learnt of science, and clinging to such help as he can give without cost merely by being less ignorant and more accustomed to sick-beds than his patients. Finally, he acquires a certain skill at nursing cases under poverty-stricken domestic conditions.”

George Bernard Shaw, The Doctor’s Dilemma 1909

Dr Alfred Slater was a GP who worked in Bermondsey at that time and may have been an inspiration. He charged six pence for consultations if patients could afford it and nothing if they could not and in spite of this he was soon able to recruit four more doctors to the practice. He worked with his wife Ada all their lives to try to alleviate the effects of poverty in the area. He upset other doctors because of his low fees, and his popularity with patients. Unlike easily manipulated and faked Google reviews that are a proxy for GP popularity today, Dr Slater earned his status by the quality of his care, his political advocacy, and his physical presence. Unlike GPs today who rarely or never visit patients at home and consult remotely, he chose to live in the heart of the community where he practiced and was, and still is, held in high esteem by the community.

https://www.atlasobscura.com/places/dr-salters-daydream

I have worked for twenty-five years in a practice that actively supports continuity of care and doctors and nurses do home visits every week. Even as we have expanded from four to ten GPs and from seven thousand to seventeen thousand patients, we have kept the same philosophy of relationship centred care. Our social standing and the confidence and trust on which we and our patients depend, is sustained by this.

This is not just a philosophical or moral issue, but there is an abundance of research supporting my view that the more distant GPs are from their patients, the lower the confidence and trust patients will have in us. Our ability to contain anxiety, manage uncertainty, diagnose promptly, and take care of patients depends on the strength of our relationships with patients and the community.

The Royal College of GPs curriculum for GP training states that ‘Continuity of care, along with generalism is a fundamental feature of general practice,’ and yet continuity is vanishing and patients are increasingly viewing GPs as gatekeepers, shopkeepers, dealers and travel agents. And a lot of GPs are quietly slipping into these roles, some passively and some actively. The emphasis on work-life balance has come about because meaning and purpose is intrinsic to a working practice bound up in relationships with patients and community, but is entirely lacking from the transactional model which alienates GPs from their patients, their work and themselves.

Just have my patients have taught me the value of continuity of care, my trainees have shown me that if they are given the chance to experience some of the satisfaction and pride that comes from investing in relationships with patients they will want more of it. This way we can build a better future for general practice. Where there is a will there is a way. I want to be proved wrong, and the 2025 GP patient survey suggests that I may be, But I think that the truth is that the status of our profession like Wile E. Coyte who has run off the edge of a cliff but doesn’t realise it and hangs briefly in the air before plummeting down to the canyon before. We only have ourselves to blame.

References:

List of Critically Endangered Crafts in which General Practice will soon be included: https://www.heritagecrafts.org.uk/skills/crafts/

The Doctor’s Dilemma, Preface on Doctors by George Bernard Shaw https://www.gutenberg.org/files/5069/5069-h/5069-h.htm#link2H_4_0008

Dr Alfred Slater, the man who created an NHS before the NHS was created https://stephenliddell.co.uk/2017/11/30/dr-alfred-salter-the-man-who-created-an-nhs-before-the-nhs-was-created/

Man in the Arena speech by Theodore Roosevelt https://www.worldfuturefund.org/Documents/maninarena.htm

Alienation and the Crisis of NHS Morale https://abetternhs.net/2023/08/28/alienation-and-the-crisis-of-nhs-staff-morale/

Factors influencing confidence and trust in health professionals: A cross-sectional study of English general practices https://bjgp.org/content/early/2025/10/03/BJGP.2025.0154

The GP Patient Survey 2025 https://www.bma.org.uk/news-and-opinion/the-results-are-in-gps-do-an-amazing-job