I want to highlight this passage: "It’s a classic parallel to what we experienced back in the early days of…
Curbside Consult with Dr. Jayne 11/18/24
The practices in which I’ve spent the majority of my clinical time over the past few years don’t use AI-assisted or ambient transcription technologies. One uses human scribes, while the other leaves physicians to their own devices for finding ways to become more efficient with their documentation.
In the urgent care setting, my scribes have always been cross trained. They started out as patient care technicians or medical assistants, and if they had excellent performance and a desire to learn, they could request to enter the in-house scribe training program. During that multi-month period, they received additional training in medical terminology, clinical documentation, regulations and requirements, and understanding the physician thought process for history-taking, creating a differential diagnosis, and ultimately creating and documenting a care plan.
Many of our human scribes had the goal of attending medical school or PA school, so they had a strong drive to learn as much as possible while doing their job. As they learned our habits for seeing patients and describing our findings, they would sometimes prompt us for something that we might have forgotten to mention or might not have performed during the exam. Because of the level of cross training, they could also assist us with minor procedures during the visit rather than just standing there and waiting for us to describe some findings.
Towards the end of the visit, when the physician typically summarizes the findings for the patient and describes the plan of care, the scribes would review and clean up the notes so that they were ready for our signature as soon as the patient disposition was complete. I would often be able to sign my notes in real time, and even if I had to wait until the end of the day, it might take me less than a minute to review each note because of the diligence they used capturing the visit.
Human scribes are also helpful when conducting sensitive visits, which often happen in the urgent care environment as we discuss a patient’s sexual history or perform sensitive portions of the physical exam. In those situations, our scribes served as both chaperones and assistants, providing support to patients when needed and assisting with specimen collection – uncapping and capping jars and tubes, ensuring accurate labeling, etc. I’ve had scribes help patients take their shoes and socks off and assist them in getting on the exam table and returning to a chair. When contrasting a visit that uses a human scribe to one where the physician has to perform their own documentation, there’s a substantial difference in the time that it takes to complete the visit, and not just from a documentation standpoint.
In speaking with my colleagues who have transitioned from human scribes to either virtual scribes or AI-assisted technologies in similar practice environments, they note that they miss the physical assistance of the scribe. No one is in the room with them who can step out and grab supplies or equipment when a situation occurs where it would be more efficient to do that instead of the physician stepping out to get what they need. There are also flow issues when chaperones are needed or when assistance is needed during a procedure, which can make the day bumpier.
Some colleagues with whom I recently discussed this mentioned that their organizations didn’t consider these workflow changes when moving to non-human documentation assistance strategies. One said that he felt that everyone thought it would be so much cheaper to not pay a person that they forgot to calculate in the time physicians would now be spending doing things that they didn’t have to do in the past.
It’s a classic parallel to what we experienced back in the early days of EHR implementation, when there were constant encounters with unintended consequences. One example: in a paper-based workflow where no one reconciled medications, implementing an EHR that requires medication reconciliation is going to increase visit duration, whether it’s done by an assistant or the physician. They should have been doing medication reconciliation in the first place because it’s a patient safety issue, but the EHR took the blame as forcing them to do something they didn’t think was important. Now we have different unintended consequences when we layer on more sophisticated technologies such as AI-assisted documentation.
One colleague described the problem of excessive summarization, where his organization’s AI documentation solution took a lengthy physician / patient discussion that included detailed risks and benefits of treatment or lack thereof and condensed it down into two sentences. When that happens, one has to consider the downstream ramifications. Will a physician even see that it’s been condensed in that way, or are they just signing notes without reviewing them to keep their inbox clear? That situation happens more than many would think. If a physician catches the issue, will they spend the time editing the note or will they just move on because they’re pressed for time? And if they do take the time to edit the visit note, will they capture all the nuances of the discussion exactly as it had occurred with that particular patient?
Another colleague, who is also a clinical informaticist, mentioned that having AI documentation solutions doesn’t fix underlying physician behavior challenges. The physician who never finished his notes at the end of the day and instead left them for Saturday mornings still leaves them for Saturday mornings, which means that he’s reviewing documentation that’s up to five days old and for visits that are no longer fresh in his mind. It’s creating issues with the technology platform, since recordings have to be kept until the notes are signed, and it’s skewing metrics for chart closure that were important to measure the success of the project.
The team that implemented the solution could have anticipated this had they looked at baseline chart closure rates, but they were in such a hurry to get the solution rolled out that now they’re having to go back and examine that data retrospectively. They also missed the opportunity to coach those physicians during the implementation phase about the patient safety value of closing notes in a timely manner.
Others have noted issues with using AI solutions to examine documentation after the fact, such as only using data from structured fields. This is great when you have a specialty that does a lot of structured documentation, but doesn’t work well in one where the subtleties of the patient’s story are largely captured via free text.
I recently attended a lecture where they discussed the hazards of using AI tools in the pediatric population, since so much of the language used in capturing a child’s status varies based on the age of the patient. For example, saying a patient is “increasingly fussy” has a meaning that goes beyond the words themselves and has a different impact when treating an infant versus an older child or a teenager.
The pediatricians also mentioned the difficulty in obtaining consent for use of AI tools during visits, especially when only one parent is present or when the child might be brought to the office by a caregiver such as a nanny or sitter. Although those individuals may have capacity to consent to treatment, they may not have specific ability to consent to the use of AI tools. There is also the issue of the child’s consent to being recorded. Although the laws generally allow parents to consent on behalf of their children, obtaining the permission of an adolescent patient is an ethical issue as well, and one which physicians may not have the time to address appropriately due to packed schedules.
The dialogue around use of AI solutions has certainly changed over the last year, and we’ve gone beyond talking about how cool it is to addressing the questions it has raised with expanding use. It’s great to see people asking thoughtful questions and even better to see vendors incorporating ethical discussions into their implementation processes. We’ll have to see what this landscape looks like in another year or two. I suspect that we will have found many other areas that need to be addressed.
How is your organization balancing the addition of AI solutions with the need for human assistants and the need to respect patient decisions? Leave a comment or email me.
Email Dr. Jayne.
I want to highlight this passage: “It’s a classic parallel to what we experienced back in the early days of EHR implementation, when there were constant encounters with unintended consequences. One example: in a paper-based workflow where no one reconciled medications, implementing an EHR that requires medication reconciliation is going to increase visit duration, whether it’s done by an assistant or the physician. They should have been doing medication reconciliation in the first place because it’s a patient safety issue, but the EHR took the blame as forcing them to do something they didn’t think was important. Now we have different unintended consequences when we layer on more sophisticated technologies such as AI-assisted documentation.”
This is exactly my concern. There are SO many unintended consequences here. We haven’t even begun to tease all of them out. You mentioned quite a few, but there will be more. Governance and safety means thoroughly examining and checking for these unintended consequences, and that takes time and money on its own.