AI transcription tools inserted suicidal ideation into social work records
A February 2026 Ada Lovelace Institute report on AI transcription tools in UK social care found that social workers were catching fabricated and mangled details in draft records, including false references to suicidal ideation, invented wording in children's accounts, and blocks of outright gibberish. Councils had adopted tools such as Magic Notes and Microsoft Copilot in the name of efficiency, but the frontline workers still carried full responsibility for correcting the output. In social work, a made-up sentence is not just a typo. It can follow a family through the system.
Incident Details
Tech Stack
References
The Pitch Was Less Paperwork
Social work is full of documentation. Case notes, assessments, meeting records, safeguarding concerns, chronology updates, court-facing paperwork, and internal summaries all compete with the actual work of talking to children, families, carers, and other professionals. That makes the sales pitch for AI transcription tools fairly obvious. Record the meeting, let the software turn speech into text, generate a cleaner draft note, and give frontline staff some time back.
By early 2026, that pitch had gone well beyond pilots. The Ada Lovelace Institute reported that AI transcription tools were being used across multiple UK councils, with products such as Magic Notes and Microsoft Copilot appearing in social care workflows. Some practitioners told researchers the tools could reduce paperwork burden and free up more time for direct work. That part is the story vendors prefer.
The more important part is what happened inside the records.
The Drafts Were Not Safe to Trust
Ada's February 2026 report documented repeated concerns about accuracy in AI-generated transcripts and summaries. Social workers described cases where the software inserted material that nobody had said, including false references to suicidal ideation. Others described children's words being rendered inaccurately or transformed into nonsense. Some workers reported blocks of gibberish. Others found that the tool changed the tone of notes into something more clinical, formal, or detached than the actual interaction.
In social work, those errors are not cosmetic. A case note can shape future visits, internal escalation, manager decisions, and court proceedings. It can influence how the next professional approaching a family interprets risk. A fabricated reference to suicidal ideation is not just a bad transcription. It is a false clinical signal inserted into an official record.
One social worker quoted in the Ada material described catching an inaccuracy and immediately recognizing the stakes: if it had been left in the file, it could have affected care further down the line. That is the central problem with these tools in high-trust public services. The output arrives wearing the costume of administrative assistance, but the burden of catching every invented sentence stays with the human.
"Human in the Loop" Still Means the Human Owns the Mess
AI advocates often respond to cases like this by saying that the tools are only drafts and that staff are expected to review them. That is true, and it is also why the productivity argument starts to wobble. If a social worker has to read every transcript closely for hallucinated content, distorted wording, and changes in tone that could alter the meaning of a child's account, then the tool is not replacing documentation work. It is changing the shape of documentation work.
The Ada report makes that clear. Social workers remained fully responsible for the final record, but there were no widely accepted standards for how much review time was necessary, how accurate a tool needed to be before deployment, or how councils should evaluate the risks of these systems in practice. In other words, the tools were being introduced into a sensitive public service faster than the governance around them had matured.
Community Care had already reported on the same tension in February 2025. Practitioners described AI note-taking tools as helpful for reducing admin load, while also raising accuracy concerns and warning against treating them as a quick fix for deeper resource shortages. A year later, Ada's report showed why those early warnings mattered. The tools were not simply transcribing speech. They were summarizing, rephrasing, and occasionally inventing.
Documentation Errors Become Service Errors
That matters because social work records do not stay local to the moment they are written. They travel. A case note may be read by another social worker, a manager, a safeguarding team, a health professional, a school, or a court. Once a false sentence enters the record, it can be repeated, relied upon, or treated as contemporaneous evidence unless someone notices and fixes it.
The risk is even sharper when the subject is a child. Children's disclosures are often partial, emotional, fragmentary, and context-dependent. Adults already have enormous interpretive power over how those statements are recorded. Handing that first draft to a system that sometimes fills gaps with its own plausible-looking language is a bad fit for the work. The AI does not know when a pause matters, when ambiguity is meaningful, or when a child's exact phrasing must be preserved because it may matter later.
There is also a style problem that is not really about style. Ada found concern that generated summaries could sound more formal or more clinical than the underlying interaction. That can subtly change the record even when the facts are mostly correct. A blunt, messy conversation gets converted into polished institutional prose. The resulting note may read better to a manager, but not necessarily more truthfully.
Efficiency Is a Thin Justification for Risky Records
None of this means every transcription tool is useless. It means the deployment logic matters. Councils are under intense staffing and budget pressure. Social workers spend too much time on paperwork. Vendors arrive with a product that promises relief. That sequence is understandable. It is also how fragile technology ends up in the middle of public services before anyone has settled basic questions about evaluation, accuracy thresholds, red lines, and audit trails.
The Ada report is careful on this point. It does not claim that every AI transcription deployment is doomed. It does say that the evidence base is thin, frontline accountability remains absolute, and current governance is too light for the sensitivity of the work. That is a restrained way of saying public bodies are buying software first and working out the rules afterward.
For Vibe Graveyard purposes, the key issue is concrete harm potential tied directly to AI behavior. These systems did not just save time badly or produce awkward wording. They generated inaccurate content in records used to make decisions about vulnerable people. That is a product failure in a safety-relevant setting.
The administrative dream behind these tools is that machines can turn messy real-world conversations into neat official notes at scale. The social work reality is less flattering. The machine drafts the note. The worker checks whether the machine quietly invented a crisis, mangled a child's words, or wrote a paragraph no one can explain. The paperwork does not disappear. It simply acquires a second failure mode.
Discussion