Rendered at 11:07:03 GMT+0000 (Coordinated Universal Time) with Cloudflare Workers.
samglass09 2 days ago [-]
Meanwhile they are pushing AI transcription and note taking solutions hard.
Patients are guilted into allowing the doctors to use it. I have gotten pushback when asked to have it turned off.
The messaging is that it all stays local. In reality it’s not and when I last looked it was running on Azure OpenAI in Australia.
I spoke to a practice nurse a few days ago to discuss this.
She said she didn’t think patients would care if they knew the data would be shipped off site. She said people’s problems are not that confidential and their heath data is probably online anyway so who cares.
taikon 2 days ago [-]
It's honestly such a big problem. One of my colleagues uses an AI scribe. I can't rely on any of his chart notes because the AI sometimes hallucinates (I've already informed him). It also tends to write a ridiculous amount of detail that are totally unnecessary and leave out important details such that I still need to comb through patient charts for (med rec, consults, etc). In the end it ends up creating more work for me. And if my colleague ever gets a college complaint I have no clue how he's gonna navigate any AI generated errors. I'm all for AI and it's great for things like copywriting, brainstorming and code generation. But from what I'm seeing, it's creating a lot more headache in the clinical setting.
If you're why doesn't this guy just check the AI scribe notes? Well, probably because with the amount of detail it writes, he'd be better off writing a quick soap note.
batshit_beaver 1 days ago [-]
> I'm all for AI and it's great for things like copywriting, brainstorming and code generation
It's funny how the assumption is always that LLMs are very useful in an industry other than your own.
tempest_ 1 days ago [-]
I mean they are not wrong.
For all the whinging about bugs and errors around here the software industry in general (some niche sub-fields excepted) long ago decided 80% is good enough to ship and we will figure the rest out later. This entire site is based on startup culture which largely prided itself on MVP moonshots.
Plus plenty of places are perfectly fine with tech dept and the AI fire hose is effectively tech debt on steroids but while it creates it at scale it can also help in understanding it.
It is is own panacea in a way.
I think it is gonna be a while before the industry figures out how to handle this better so might as well just ride the wave and not worry too much about it in software.
Still software is not medicine even if software is required in basically every industry now. It should more conservative and wait till things settle down before jumping in.
_AzMoo 1 days ago [-]
My (extensive) experience with LLM code generation is that it has the same issues you describe in your field. Hallucinations, over-engineering, misses important requirements/patterns.
But engineers have these same problems. The key is that the content creator (engineers for codegen, doctors for medicine) is still responsible for the output of the AI, as if they wrote it themselves. If they make a mistake with an AI (eg, include false data - hallucinations), they should be held accountable in the same way they would if they made a mistake without it.
shimman 1 days ago [-]
Okay but since we know how humans actually behave, they will fully trust the indeterministic machine and give away their thinking. Sadly there is a large swath of humans that will act like this, maybe 20-30%.
Are you willing to put your life in the hands of these people fully using the machines to do everything?
Acting like that smart people aren't getting one shot'ed by these machines is very dangerous. Even worse is how quickly your skills actual degrade. If knew my doctor was using anything LLM related, I would switch doctors.
rconti 1 days ago [-]
It feels very much like AI is creating AI lock-in (if not AI _vendor_ lock-in) by creating so much detailed information that it's futile to consume it without AI tools.
I was updating some gitlab pipelines and some simple testing scripts and it created 3 separate 300+ line README type metadata files (I think even the QUCIKSTART.md was 300 lines).
acuozzo 1 days ago [-]
> I'm all for AI and it's great for things like copywriting, brainstorming and code generation
That's funny. I would have said the same thing about your field prior to reading your comment.
dmtroyer 1 days ago [-]
sounds like they need a better instructions.md
SpaceNoodled 2 days ago [-]
Is there nothing like HIPAA there or what?
samglass09 2 days ago [-]
Very little protections. The entire medical records of a significant percentage of the NZ population were stolen recently and put up for sale online. Zero consequences for the medical practices who adopted the hacked software.
mixmastamyk 1 days ago [-]
Interesting, a person was telling me recently that NZ privacy laws were quite strong. Perhaps a different category.
The laws are, the policing is not. At least not in medical data
lights0123 2 days ago [-]
Many AI companies, including Azure with their OpenAI hosting, are more than willing to sign privacy agreements that allow processing sensitive medical data with their models.
Ucalegon 2 days ago [-]
The devil is in the details. For example, OAI does not have regional processing for AU [0] and their ZDR does not cover files[1]. Anthropic's ZDR [2] also does not cover files, so you really need to be careful, as a patient/consumer, to ensure that your health, or other sensitive data, that is being processed by SaaS frontier models is not contained in files. Which is asking a a lot of the medical provider to know how their systems work, they won't, which is why I will never opt in.
The models are licensed to Microsoft, and you pay them for the inference.
Ucalegon 1 days ago [-]
There is no way to upload files as a part of context with Azure deployments, you have to use the OAI API [0], and without having an architecture diagram of the solution, I am not going to trust it based off of the known native limitations with Azure's OAI implementation.
The wilful ignorance and total apathy is appalling.
I've had similar experiences in Australia. I emailed one of my docs' practices asking if they use Heidi AI (or anything similar) and that I do not consent. They were using it without my consent.
In the consultation, he tried to give me the schpiel, including the 'it stays local' thing. The Heidi AI website has the scripts for clinicians; he ran through them all.
Oh, their documents for clinicians also mention every two sentences that patient/client consent is not required at all. I wonder why they keep saying that? Hmm.
This doctor knows I am a developer. When I asked him to explain what he meant by 'local data', he said the servers were in Australia. I almost flipped the desk. Aside from the fact that it is mandatory (it's the law! they do not have a choice!), it's ...kind of meaningless where the servers are, especially when he (on behalf of Heidi AI) was trying to sell it as a security or privacy feature. When I pointed that out, he just couldn't wrap his head around it. Of course he can't, he doesn't understand.
AHPRA's "Meeting your professional obligations when using Artificial Intelligence in healthcare" guideline[0] (not any kind of enforceable requirement, unfortunately) has great stuff in it. It encourages using it with the informed consent of patients. Even if my doctor read it and agreed with it, and cared about getting consent, how the hell can he inform patients sufficiently when he has absolutely no idea about, well, anything?
He keeps pushing it and asking me about whether I've changed my mind about allowing him to use it. No! He keeps asking me questions that only confirm he hasn't even done a perfunctory web search about why some people hate LLMs, especially in the context of PII and PHI.
I really do feel for clinicians, but these products are not the answer.
I spec'd up an implementation of this that uses a hardware button with colors that is in reach of either party. The customer went with a different vendor based on price/"complexity"/training.
jimjimjim 2 days ago [-]
The New Zealand Chief Digital Officer allowed Australian cloud providers to be used as there weren't suitable NZ data centers and this was many years ago.
samglass09 1 days ago [-]
Health NZ adopted Snowflake. It was about costs/fancy tech. We have always had data centres. Nobody *needs* snowflake. They could have used Apache Spark.
1 days ago [-]
bfivyvysj 1 days ago [-]
What are you talking about, NZ has had suitable DC's for decades now.
tjpnz 1 days ago [-]
Didn't Health NZ just suffer a major data breach and have patient records ransomed?
antod 1 days ago [-]
There were two serious breaches recently but they were at private companies not HNZ.
1 days ago [-]
tokioyoyo 1 days ago [-]
I genuinely think people “care” about it (in quotes). It’s one of the things where nobody cares unless something bad happens, and when bad thing happens, they shrug it off and forget about it a week later.
I’d go as far as saying she’s right. And we’re in a tiny minority for even thinking about it.
gurachek 1 days ago [-]
The union rep gets it - people improvise when you cut their tools and then threaten discipline for improvising.
That memo is how you make staff hide things instead of asking for help.
The scarier part though is that LLM-written clinical notes probably look fine. That's the whole problem. I built a system where one AI was scoring another AI's work, and it kept giving high marks because the output read well. I had to make the scorer blind to the original coaching text before it started catching real issues. Now imagine that "reads well, isn't right" failure mode in clinical documentation.
Nobody's re-reading the phrasing until a patient outcome goes wrong.
totetsu 1 days ago [-]
ASR models can output a confidence score along with the text, but it is rarely used in the UI to display the results.. or maybe lost entirely in a subsequent LLM layer.
salawat 1 days ago [-]
Physicians need to have it pounded into them that every hallucination is downstream harm. AI has no place in medicine. If they insist on it, then all transcripts must be stored with the raw audio. Which should be accessible side by side, with lines of transcript time coded. It's the only way to actually use these safely, while guarding against hallucinations.
gurachek 1 days ago [-]
Raw audio is a cool idea! I've seen a similar approach in other domains, "keep the source of truth accessible so you can verify the AI output against it".
I wouldn't go as far as "no place in medicine" though. The Heidi scribe tool mentioned in the article is a good example, because in the end it's the doctor who reviews and signs off.
IMO the problem is AI doing the work with no human verification step, but I can 100% agree I don't want to have vibe-doctor for my next surgery/consult :D
bonsai_spool 1 days ago [-]
> Physicians need to have it pounded into them that every hallucination is downstream harm.
I think any person using 'AI' knows it makes mistakes. In a medical note, there are often errors at present. A consumer of a medical note has to decide what makes sense and what to ignore, and AI isn't meaningfully changing that. If something matters, it's asked again in follow up.
theshackleford 1 days ago [-]
> I think any person using 'AI' knows it makes mistakes.
You think wrong. I’m now encountering people on a regular basis arguing “those days are behind us” and it’s “old news.”
jmward01 1 days ago [-]
I have seen the evolution of these tools and I think they are going to push a fundamental change to medical care. Notes have been getting more and more abused, at least in the US. Big health systems want them for a lot of reasons that have nothing to do with helping a practitioner improve the care of their patient. They want to capture every billable moment of that encounter and potentially prep things like labs, appointments, clinical trial screening, pre-auths, etc. Some of this is good for the patient but a lot isn't. Also, the reality is that many practitioners spend as much, or more time, on the note than on the patient. That clearly isn't to their benefit. There is a reason they sit there and type constantly while talking to you and that doesn't stop when you leave the room. The demands on them to document everything so that all the accounting can happen are actually harming healthcare.
I think there is a chance that these systems will lead to a change where the note isn't the fundamental record of the encounter. Instead different artifacts are created specifically for each entity that needs it. Billing gets their view, and scheduling gets theirs, and, etc etc... It will, hopefully, give the practitioners a chance to get back to focusing on the patient and not ensuring their note quality captured one more billable code. Of course the negative is also likely to happen here too. As practitioners spend less time on the note they will likely not get that back in time with individual patients, but instead on seeing more patients. It will also likely lead to higher bills as the health systems do start squeezing more out of every encounter. There is no perfect here when profit is the driving motivator but with this much change happening I can only hope that it causes the industry as a whole to shake up enough to maybe find a new better optimum to land in.
shigawire 1 days ago [-]
>I think there is a chance that these systems will lead to a change where the note isn't the fundamental record of the encounter. Instead different artifacts are created specifically for each entity that needs it. Billing gets their view, and scheduling gets theirs, and, etc etc..
This is what an EHR does somewhat. The discrete data elements in the DB and the way they are displayed in the system are a better record than free text notes.
The problem is creating standards so this data is easily exchanged. Anyone can read and parse a free text note - but if we had standards this would be less necessary.
jimbokun 1 days ago [-]
This will always happen as long as there is a combative relationship between private insurers and providers over reimbursements. Each side is using documentation or lack thereof to make their financial case.
burnte 2 days ago [-]
Yeah, no privacy or security there. There are some tools explicitly designed at helping healthcare providers produce better notes faster, and a couple of them are AMAZING. I'm an AI-half-empty guy, I'm keenly aware of its shortcomings and deploy it thoughtfully, and even with my skepticism there are a couple of tools that are just plain great. I think using LLMs to create overviews and summaries is a great use of the tech.
javascriptfan69 2 days ago [-]
The one my doctor was using got my obs numbers completely wrong.
We had to correct them at the end of the consultation.
simmerup 2 days ago [-]
Gotta break a few eggs to save 2 minutes of thinking and work
burnte 16 hours ago [-]
No different than if that doctor was using a human scribe and they typoed. We make all our doctors proof their notes, it's SOP even long before AI.
2 days ago [-]
beatthatflight 1 days ago [-]
I know at least one GP that has stopped using Heidi Health for transcription. He (and as I've noticed with transcriptions from my medical professionals) has noticed many errors, far too many to be comfortable. Things might improve, but not yet.
dwilkox 1 days ago [-]
This is where I'm at as a GP. Every few months I give Heidi another try, but I haven't noticed any real improvement over the last two years. It spends lots of words on trivial nonsense and misses clinically significant points and sometimes entire issues. It takes far more time to review and fix the notes than it saves in typing. Presumably it will be good enough one day, but it's not there yet.
MengerSponge 1 days ago [-]
It's Gell-Mann amnesia: you notice the errors in fields where you're an expert.
I'm a female doctor, for me the hardest thing so far was watching AI shut down a discussion on harassment in healthcare in real time. Women were contributing stories to a website documenting their experience until one claiming to be written by a male was published. All contributors were fluent but this hit different - lower information density (harder to get a read on the age, ethnicity, social class of the writer) but higher emotional impact. Women usually seemed to have interrogated themselves, identifying choices they had made (joining after work drinks, accepting a ride, being alone) but this showed no sign of any attempt to think through how alternative versions of the account would look (eg, a white woman contributing to a forum on racism to complain discrimination by non-white men protected by the patriarchy, also they were on cocaine?) The last-minute drug reference was weird (doesn't tha complicate the question of who the victim is, not reinforce it? why cocaine - except that "crack" would sit in a field adjacent to both "abuse" and" damage" and "drug" next to "doctor")? Other men had shared experiences, but not like this. Feed the stories into GPT and ask which is most likely AI and it "carefully and respectfully" identifies 239 for different reasons (symmetry, etc) and offered to rewrite in a more human way. Nobody has contributed since.
https://www.survivinginscrubs.co.uk/your-stories/
yeah I'm in an industry that for no good reason still writes instructions using a language that hasn't been spoken since about 800AD. Ai scribes might not change how we practice but it is having real-world effects.
ampachick 10 hours ago [-]
I meant - "crack" is near "damage" but also near "drug" which is near both "abuse and "doctor" and AI does love synonyms/homonyms, but if we are talking about racism - we know AI tools also favours medical research by native English speakers? https://www.bmj.com/content/392/bmj-2025-087581/rr-8 - so now a non-white written "accent" can silence non-white doctors.
As a doctor navigating AI scribes (which I don't use) it feels like we are being distracted with toys while tech companies figure out how to become what pharma was to the last generation. Interested in non-medical perspectives
bilekas 23 hours ago [-]
I was applying recently to a role that was pretty interesting and so I wrote the email on the train on my way home, didn't have my laptop with me.
In the email I wrote out everything myself, absolutely no use of AI, but after I hit send I realised there was a pretty silly typo, nothing grave but it irked me.
I decided out of boredom to see would my email be considered AI as it was probably going to go through a million filters these days, I popped it into an online checker (I don't know the quality of these so who knows) and it told me with 75% certainty it was written by AI.
It was not at all. It was written overly hastily on a phone on public transport.
So I wonder how someone who might be grammar orientated and particular with the semantics would prove otherwise.
I can see a company needing to find any excuse to let people go saying "well theAI says you used the AI to do your work, we're letting you go"
This is just about not using free/public AI tools.
beatthatflight 1 days ago [-]
that's mentioned in OP article.
Heidi is frustratingly consistent at hallucinating stuff. I've seen it in almost all of the dozen or so summaries I've had from medical people recently (surgeon, physio, consultant). A GP I know tried for a month and then was like 'it's not worth the risk exposure to me or my patients'.
alliao 1 days ago [-]
AI doesn't forget and soon all of new zealanders will have their health histories internalised by AI so it can individually calculate insurance premiums without knowing why....
bpodgursky 1 days ago [-]
This is a ridiculous sentence. Of course inference-only AI forgets. You can literally just program it that way.
In fact, it's human transcribers who chose whether to forget the details of a case or whether to share the details of an especially funny patient with their buddies at the bar.
simmerup 22 hours ago [-]
But in this case they’re actively being told by the hospital not to use AI due to privacy concerns so it seems unlikely that it’s been programmed to respect privacy doesn’t it
dpoloncsak 16 hours ago [-]
Sending data to OpenAI to train a new model on does not feel like it constitutes 'AI doesn't forget'. The AI has nothing to do with the thousands of other companies storing your data for various reasons.
You can program a harness to always send a MEMORY.md file like OpenClaw, or use Vector Stores like OpenAI does, or find some other implementation of 'memory', but these are not an inherent feature of 'AI'. Quite the opposite...the LLMs we currently see will never learn or adapt by themselves, they don't touch their own weights
yunnpp 1 days ago [-]
This is a blatant violation of patient privacy. That the output is often hallucinated doesn't even matter here. If the hospital wants to use LLMs, better deploy them on-premise or a trusted network at least.
simianwords 1 days ago [-]
Noticing that people in the west would rather speak about privacy than more efficient or cheaper healthcare.
Enterprises are ok sharing their code base with OpenAI. I think it should be okay for patients.
peterashford 1 days ago [-]
Enterprises do that if they choose. Patients can choose as well. And it's their choice, not anyone else's
theshackleford 24 hours ago [-]
Some AI bullshitting into my medical history isn’t making my healthcare cheaper or more efficient, it’s just making it worse.
Patients are guilted into allowing the doctors to use it. I have gotten pushback when asked to have it turned off.
The messaging is that it all stays local. In reality it’s not and when I last looked it was running on Azure OpenAI in Australia.
I spoke to a practice nurse a few days ago to discuss this.
She said she didn’t think patients would care if they knew the data would be shipped off site. She said people’s problems are not that confidential and their heath data is probably online anyway so who cares.
If you're why doesn't this guy just check the AI scribe notes? Well, probably because with the amount of detail it writes, he'd be better off writing a quick soap note.
It's funny how the assumption is always that LLMs are very useful in an industry other than your own.
For all the whinging about bugs and errors around here the software industry in general (some niche sub-fields excepted) long ago decided 80% is good enough to ship and we will figure the rest out later. This entire site is based on startup culture which largely prided itself on MVP moonshots.
Plus plenty of places are perfectly fine with tech dept and the AI fire hose is effectively tech debt on steroids but while it creates it at scale it can also help in understanding it.
It is is own panacea in a way.
I think it is gonna be a while before the industry figures out how to handle this better so might as well just ride the wave and not worry too much about it in software.
Still software is not medicine even if software is required in basically every industry now. It should more conservative and wait till things settle down before jumping in.
But engineers have these same problems. The key is that the content creator (engineers for codegen, doctors for medicine) is still responsible for the output of the AI, as if they wrote it themselves. If they make a mistake with an AI (eg, include false data - hallucinations), they should be held accountable in the same way they would if they made a mistake without it.
Are you willing to put your life in the hands of these people fully using the machines to do everything?
Acting like that smart people aren't getting one shot'ed by these machines is very dangerous. Even worse is how quickly your skills actual degrade. If knew my doctor was using anything LLM related, I would switch doctors.
I was updating some gitlab pipelines and some simple testing scripts and it created 3 separate 300+ line README type metadata files (I think even the QUCIKSTART.md was 300 lines).
That's funny. I would have said the same thing about your field prior to reading your comment.
https://news.ycombinator.com/item?id=44564349
[0] https://developers.openai.com/api/docs/guides/your-data#whic...
[1] https://developers.openai.com/api/docs/guides/your-data#stor...
[2] https://platform.claude.com/docs/en/build-with-claude/zero-d...
The models are licensed to Microsoft, and you pay them for the inference.
[0] https://github.com/openai/openai-python/issues/2300
I've had similar experiences in Australia. I emailed one of my docs' practices asking if they use Heidi AI (or anything similar) and that I do not consent. They were using it without my consent.
In the consultation, he tried to give me the schpiel, including the 'it stays local' thing. The Heidi AI website has the scripts for clinicians; he ran through them all.
Oh, their documents for clinicians also mention every two sentences that patient/client consent is not required at all. I wonder why they keep saying that? Hmm.
This doctor knows I am a developer. When I asked him to explain what he meant by 'local data', he said the servers were in Australia. I almost flipped the desk. Aside from the fact that it is mandatory (it's the law! they do not have a choice!), it's ...kind of meaningless where the servers are, especially when he (on behalf of Heidi AI) was trying to sell it as a security or privacy feature. When I pointed that out, he just couldn't wrap his head around it. Of course he can't, he doesn't understand.
AHPRA's "Meeting your professional obligations when using Artificial Intelligence in healthcare" guideline[0] (not any kind of enforceable requirement, unfortunately) has great stuff in it. It encourages using it with the informed consent of patients. Even if my doctor read it and agreed with it, and cared about getting consent, how the hell can he inform patients sufficiently when he has absolutely no idea about, well, anything?
He keeps pushing it and asking me about whether I've changed my mind about allowing him to use it. No! He keeps asking me questions that only confirm he hasn't even done a perfunctory web search about why some people hate LLMs, especially in the context of PII and PHI.
I really do feel for clinicians, but these products are not the answer.
[0] https://www.ahpra.gov.au/Resources/Artificial-Intelligence-i...
I’d go as far as saying she’s right. And we’re in a tiny minority for even thinking about it.
That memo is how you make staff hide things instead of asking for help.
The scarier part though is that LLM-written clinical notes probably look fine. That's the whole problem. I built a system where one AI was scoring another AI's work, and it kept giving high marks because the output read well. I had to make the scorer blind to the original coaching text before it started catching real issues. Now imagine that "reads well, isn't right" failure mode in clinical documentation.
Nobody's re-reading the phrasing until a patient outcome goes wrong.
I wouldn't go as far as "no place in medicine" though. The Heidi scribe tool mentioned in the article is a good example, because in the end it's the doctor who reviews and signs off.
IMO the problem is AI doing the work with no human verification step, but I can 100% agree I don't want to have vibe-doctor for my next surgery/consult :D
I think any person using 'AI' knows it makes mistakes. In a medical note, there are often errors at present. A consumer of a medical note has to decide what makes sense and what to ignore, and AI isn't meaningfully changing that. If something matters, it's asked again in follow up.
You think wrong. I’m now encountering people on a regular basis arguing “those days are behind us” and it’s “old news.”
I think there is a chance that these systems will lead to a change where the note isn't the fundamental record of the encounter. Instead different artifacts are created specifically for each entity that needs it. Billing gets their view, and scheduling gets theirs, and, etc etc... It will, hopefully, give the practitioners a chance to get back to focusing on the patient and not ensuring their note quality captured one more billable code. Of course the negative is also likely to happen here too. As practitioners spend less time on the note they will likely not get that back in time with individual patients, but instead on seeing more patients. It will also likely lead to higher bills as the health systems do start squeezing more out of every encounter. There is no perfect here when profit is the driving motivator but with this much change happening I can only hope that it causes the industry as a whole to shake up enough to maybe find a new better optimum to land in.
This is what an EHR does somewhat. The discrete data elements in the DB and the way they are displayed in the system are a better record than free text notes.
The problem is creating standards so this data is easily exchanged. Anyone can read and parse a free text note - but if we had standards this would be less necessary.
We had to correct them at the end of the consultation.
https://en.wiktionary.org/wiki/Gell-Mann_Amnesia_effect
In the email I wrote out everything myself, absolutely no use of AI, but after I hit send I realised there was a pretty silly typo, nothing grave but it irked me.
I decided out of boredom to see would my email be considered AI as it was probably going to go through a million filters these days, I popped it into an online checker (I don't know the quality of these so who knows) and it told me with 75% certainty it was written by AI.
It was not at all. It was written overly hastily on a phone on public transport. So I wonder how someone who might be grammar orientated and particular with the semantics would prove otherwise.
I can see a company needing to find any excuse to let people go saying "well theAI says you used the AI to do your work, we're letting you go"
This is just about not using free/public AI tools.
Heidi is frustratingly consistent at hallucinating stuff. I've seen it in almost all of the dozen or so summaries I've had from medical people recently (surgeon, physio, consultant). A GP I know tried for a month and then was like 'it's not worth the risk exposure to me or my patients'.
In fact, it's human transcribers who chose whether to forget the details of a case or whether to share the details of an especially funny patient with their buddies at the bar.
You can program a harness to always send a MEMORY.md file like OpenClaw, or use Vector Stores like OpenAI does, or find some other implementation of 'memory', but these are not an inherent feature of 'AI'. Quite the opposite...the LLMs we currently see will never learn or adapt by themselves, they don't touch their own weights
Enterprises are ok sharing their code base with OpenAI. I think it should be okay for patients.