<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Byte to Bedside]]></title><description><![CDATA[An exploration of how emerging health tech can be brought to the bedside. ]]></description><link>https://www.byte2bedside.com</link><image><url>https://www.byte2bedside.com/img/substack.png</url><title>Byte to Bedside</title><link>https://www.byte2bedside.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 04 May 2026 17:37:05 GMT</lastBuildDate><atom:link href="https://www.byte2bedside.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Ron]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[byte2bedside@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[byte2bedside@substack.com]]></itunes:email><itunes:name><![CDATA[Ron Li]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ron Li]]></itunes:author><googleplay:owner><![CDATA[byte2bedside@substack.com]]></googleplay:owner><googleplay:email><![CDATA[byte2bedside@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ron Li]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[How AI has changed my clinical teaching as an attending physician ]]></title><description><![CDATA[How do we train new physicians when medical knowledge and clinical reasoning are no longer scarce resources?]]></description><link>https://www.byte2bedside.com/p/how-ai-has-changed-my-clinical-teaching</link><guid isPermaLink="false">https://www.byte2bedside.com/p/how-ai-has-changed-my-clinical-teaching</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Tue, 17 Mar 2026 05:51:26 GMT</pubDate><content:encoded><![CDATA[<p>I start every attending block with medical trainees the same way now. After we finish introductions, I pull out my phone, open an AI application like OpenEvidence, and tell the team: <em>I&#8217;m going to be looking things up today. I expect you to do the same.</em></p><p>We all recognize the reflex: that moment of hesitation, even embarrassment, when someone asks you a question and you don&#8217;t know the answer. In medicine, this reflex runs deep. We&#8217;re trained in a culture where knowledge is currency, where the physician who can recall the landmark trial or the rare diagnosis from memory commands a particular kind of authority. Pulling out your phone to look something up, especially in front of a patient, feels like a crack in that aura. It signals that maybe you don&#8217;t know as much as you&#8217;re supposed to. So instead, we hedge and deflect, and often attempt to instead answer a different question to showcase what we <em>do</em> know, a tactic that is easily recognized, including by patients.</p><p>I want the medical trainees to feel that it&#8217;s safe to not know. In fact, it&#8217;s expected. What&#8217;s not acceptable is pretending you know when you don&#8217;t, or letting an unanswered question quietly disappear.</p><p>One of my main jobs on rounds is to listen for those moments when there is the lingering question, the thing that doesn&#8217;t quite fit, or the detail someone glossed over because they weren&#8217;t sure. I try to catch it, name it, and then answer it together in real time. Now that the barrier to finding out is drastically lower with AI, there really isn&#8217;t any excuse to not look up something we don&#8217;t know. </p><h2>Teaching the cognitive work that AI won&#8217;t replace (yet)</h2><p>Medicine is traditionally taught through pattern recognition. You learn an illness script &#8212; a constellation of symptoms, a differential diagnosis, a treatment algorithm &#8212; and your job is to match the patient in front of you to the right script. This produces a particular kind of associative thinking that is common among physicians, and those who are best at making these associations are lauded as &#8220;master diagnosticians&#8221; (think House, MD). </p><p>This is precisely the type of cognitive task that will be commoditized by AI. </p><p>Take a patient who presents with a possible cellulitis, a common skin infection. Traditional teaching rounds may consist of discussions (and pop quizzes) on the key words associated with cellulitis (e.g. erythema, tenderness, well demarcated borders, leukocytosis), the indications for IV antibiotics, and red flag symptoms for more severe infections such as osteomyelitis or necrotizing fasciitis that should trigger subsequent steps to also memorize (MRI, surgical consult, etc.). A quick search on any LLM would yield this checklist instantly, which is why being able to recite it on rounds no longer constitutes expertise.</p><p>What I try to teach instead is to think more like a curious scientist rather than a pattern-matcher.</p><p>Take the same clinical scenario: a patient comes in with a red, painful area on their skin. I ask the team to forget all the words they may have memorized and instead think about the concepts that underly cellulitis: <em>picture the anatomy. Skin, subcutaneous tissue, fascia, muscle, bone. Now, based on the clinical presentation, what is your hypothesis for which layer the pathology is in? </em></p><p>Then we stress-test it. What exam findings would increase or decrease the likelihood of that hypothesis? What&#8217;s your level of certainty &#8212; and I make them commit to a number. Are you 50% confident? 80%? 95%? The number matters, because the next question is: <em>is that enough certainty to act, or do you need more information?</em> If you&#8217;re only 60% sure the infection is superficial, maybe you need an ultrasound to rule out an abscess or a MRI to look more closely within the soft tissue and bone. If there&#8217;s even a 10% chance the process involves the deep fascia, that may be enough for a surgical consult, or at least additional testing. And the test you order has its own accuracy, so now you&#8217;re reasoning about uncertainty on top of uncertainty.</p><p>This is how scientists and engineers think. Observe, assess the fidelity of your data, construct a hypothesis about the underlying mechanism, quantify your uncertainty, decide whether to act or gather more information, and then build a framework for how you think the patient gets better. That last part is important to be honest about: your treatment plan is, at some level, your best guess. It&#8217;s an informed guess, grounded in evidence, but it&#8217;s still a prediction about a complex biological system under uncertainty.</p><p>The culture in medicine often treats uncertainty as something to hide or resolve as quickly as possible, not something to name and manage. When I ask a trainee to commit to a probability, I'm doing something that feels unnatural to many, but is what I believe to be a distinct quality that remains within the purview of a skilled physician in a time when AI will beat out any master diagnostician at pattern matching: the ability to name uncertainty, hold it transparently, and still earn the trust of the patient whose life depends on our decisions. </p><h2>The skill that won&#8217;t be automated: taking accountability </h2><p>The physician&#8217;s job was never really to be the sole source of clinical knowledge even though our training system was built as if it were. <strong>The real job is to own the outcome</strong><em><strong>.</strong></em> Someone has to integrate the evidence, the patient&#8217;s context, the team&#8217;s capabilities, and the system&#8217;s constraints into a plan that results in high-quality care and a safe discharge. And when things don&#8217;t go as expected, they have to take responsibility, escalate to the right resources, and course-correct.</p><p>As physicians, AI will help us construct our reasoning, challenge our hypotheses, surface evidence we miss, and flag when our plans seem inconsistent with the data. But we need to architect and own the reasoning framework, which empowers us to take the accountability needed to sit at the patient&#8217;s bedside with honesty and confidence to say <em>I&#8217;m not sure what the best course of action is yet, but I am here to figure this out with you, because I care about what happens to you.</em> </p><h2>What I&#8217;m evaluating for in trainees </h2><p>When I think about who is going to thrive in this new world, I&#8217;m not primarily looking at the medical trainee who knows the most. The knowledge playing field is leveling fast, and that&#8217;s not a bad thing in my opinion.</p><p>Instead, I look for the individual who chases down the loose thread and follows up on the thing that didn&#8217;t quite make sense. Who stays with the patient&#8217;s problem even after the initial plan is made, watching for the moment when the hypothesis starts to break down? Who takes ownership of the hospital stay &#8212; not just making the diagnosis, but the whole arc of getting a patient safely home?</p><p>Curiosity, drive, and accountability have always been important in medicine, but they used to be masked by the knowledge hierarchy; the physician who could recall the most obscure fact earned a certain kind of credibility, regardless of whether they followed through on the plan. Now that AI is democratizing access to knowledge, the traits that were always the real differentiators are becoming impossible to hide behind anything else.</p><p>This is what the future of medicine will be built on.</p><p></p>]]></content:encoded></item><item><title><![CDATA[Building new tech-enabled care models within a health system ]]></title><description><![CDATA[AI and virtual care are the building blocks for the next generation of care models and health systems can lead in this transformation.]]></description><link>https://www.byte2bedside.com/p/from-ai-models-to-care-models-how</link><guid isPermaLink="false">https://www.byte2bedside.com/p/from-ai-models-to-care-models-how</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Sat, 08 Nov 2025 18:26:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!3A_L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3A_L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3A_L!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!3A_L!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!3A_L!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!3A_L!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3A_L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2951283,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.byte2bedside.com/i/178303950?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3A_L!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!3A_L!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!3A_L!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!3A_L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b0b65af-b93d-4cc5-a1e3-60bd8d4bfdeb_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I&#8217;m often struck by how little the fundamental architecture of care has changed in over a century &#8212; we still ask patients to come see a doctor in a clinic when they have a problem, and in hospitals, we still round room to room much as we did a hundred years ago.  The Covid pandemic accelerated digital modalities such as telehealth, but the underlying architecture of care delivery &#8212; how we organize work, teams, engagement, and follow-up &#8212; remains largely the same.</p><p>In an earlier post &#8212; <em><a href="https://www.byte2bedside.com/p/what-if-we-had-to-care-for-twice">What if we had to care for twice as many patients?</a></em> &#8212; I posed what I believe is an existential question for healthcare in our country: what would it take to double our capacity for care, without doubling our workforce or costs?</p><p>When I look at new capabilities such as AI or virtual care, I always go back to this core question.  We can deploy AI models into workflows or increase the number of telehealth encounters, but without re-architecting the underlying system, we risk perpetuating the same legacy inefficiencies, except with new technologies that may actually add to costs.  </p><p>Nevertheless, these technologies are critical building blocks for the transformations that we need.  If we are to upgrade the underlying chassis of care delivery, we have to leverage capabilities such as AI and virtual care as <strong>enabling functions</strong> for the care models of the future. These technologies are not solutions in themselves; they are <strong>materials</strong> from which new systems can be built.</p><h3><strong>Designing Systems for Better Care</strong></h3><p>Imagine the care models of the future: care would be informed by <strong>better, more accurate information</strong>; delivered through <strong>processes unconstrained by time and place</strong>; and experienced by <strong>patients who are informed, empowered, and connected</strong>.<br>It would be a learning system that captures experiences from each interaction to improve the next.  It would also be scalable &#8212; able to expand access and capacity without a proportional rise in cost or burden.</p><p>How would we engineer such a system?</p><p>We might think of it as having two fundamental layers: <strong>an intelligence layer</strong> that makes sense of information and guides action, and a <strong>delivery layer</strong> that executes those actions efficiently across people, time, and space.</p><p>AI provides the foundation for the <strong>intelligence layer</strong> &#8212; organizing and interpreting data, surfacing the right information, and supporting decisions in context. Virtual care (coupled with AI) supports a lower cost, scalable <strong>delivery layer</strong> &#8212; the channels, workflows, and human touchpoints through which care can be delivered independent of physical proximity and constraints.  Virtual care also provides a more seamless digital medium for AI to be integrated into the care delivery experience.  </p><p>In person care will continue to be necessary, both as standalone models for higher acuity care delivery and complements to virtual first care.  However, the future of care delivery will depend on how seamlessly we integrate in-person and virtual layers into a unified, intelligent system &#8212; one that delivers the right level of care, in the right place, at the right time.</p><p>We are already beginning to see early examples of this architecture take shape within health systems. At my institution, Stanford Health Care, two new digital health programs &#8212; <em>eConsult 2.0</em> and <em>Stanford Health Care at Home</em> &#8212; illustrate how the combination of AI and virtualization can enable entirely new systems of care delivery. Each begins with the following question: <strong>how might intelligence and virtualization of care allow us to scale expertise and capacity without scaling cost?</strong></p><h3><em><strong>eConsult 2.0</strong></em><strong> &#8212; An AI-Enabled Specialty Consult Service</strong></h3><p>At Stanford Health Care, our eConsult program has long improved access to specialty expertise. An eConsult is an asynchronous virtual care model where a provider &#8212; often a primary care clinician &#8212; consults a specialist digitally rather than referring their patient to a specialty clinic. This model enables patients to receive specialist input within days instead of waiting months for a traditional appointment. But the real opportunity lies in reimagining eConsult as a tech enabled specialty care service &#8212; one in which AI is built into the architecture of how expertise is delivered, not simply layered on top of it.</p><p>Yet scaling this model across a health system has proved challenging. Specialists face high cognitive burden, limited compensation, and outdated consultation templates that required constant manual upkeep. The program&#8217;s success has depended not only on clinical expertise but on labor-intensive processes &#8212; gathering data, reviewing charts, and drafting responses &#8212; that don&#8217;t scale linearly with demand.</p><p>Our first experiments with AI focused narrowly on chart review, using large language models to surface relevant clinical data and reduce time to completion. The early success of these efforts prompted a deeper question: what if AI could do more than streamline a task? What if it could become a <strong>core design element</strong> of the care model itself?</p><p>That question led to the creation of <em>eConsult 2.0</em> &#8212; a re-architected, AI-enabled clinical service built from the ground up as a tech-enabled system of care, not a digital add-on. Rather than deploying AI just to help with chart review, we began designing around AI as an <strong>enabling function</strong>&#8212;the intelligence layer that supports a continuously learning and scalable specialty consultation service that consists of three layers: </p><ol><li><p><strong>Knowledge Base</strong> &#8212; continuously updated, AI-assisted consultation templates that serve as the shared memory of the service.</p></li><li><p><strong>Application Layer</strong> &#8212; a &#8220;one-stop-shop&#8221; interface, integrated into Epic, that uses large language models to summarize relevant chart data for the specialist and draft responses while continuously learning and updating the knowledge base.</p></li><li><p><strong>Workforce Layer</strong> &#8212; specialists whose expertise is enhanced by AI, allowing more consults in less time and with higher consistency.</p></li></ol><p>The redesign of the eConsult program demonstrates how intelligence can be embedded into the fabric of a clinical service &#8212; turning what was once a high-friction workflow into a scalable system for distributing expertise. </p><h3><strong>Stanford Health Care at Home &#8212; Expanding Hospital Capacity Through Virtual Care</strong></h3><p>If <em>eConsult 2.0</em> reimagines how expertise is delivered, <em>Stanford Health Care at Home</em> reimagines where hospital care is delivered. The program was born from a pressing organizational need to expand hospital bed capacity, leading to the following question: <strong>what if patients who no longer require hospital-level intensity care could continue recovery safely at home, supported virtually by their care team?</strong></p><p>In today&#8217;s system, inpatient providers often face a binary choice &#8212; keep the patient in the hospital, or discharge them into a full ambulatory environment that may not be sufficient to support their immediate post discharge care needs.  Patients and providers often describe hospital discharge as &#8220;walking off a cliff&#8221; &#8212; a sudden drop in support that can lead to unnecessarily prolonged hospitalizations or, conversely, premature discharges that result in readmissions and other adverse outcomes.</p><p><em>Stanford Health Care at Home</em> creates a third option: a virtual-first, home-based model that bridges the transition between inpatient and ambulatory care. Patients who are clinically stable but still need close monitoring &#8212; for example, those resolving sepsis, restarting medications, or awaiting culture results &#8212; can be discharged earlier and managed at home through a hybrid model of virtual physician visits, nursing support, and coordinated follow-up after hospitalization.</p><p>This design expands hospital capacity without building new beds. By virtualizing the &#8220;tail end&#8221; of hospitalization &#8212; the final days often defined by low-acuity, high-cost care &#8212; it converts fixed inpatient infrastructure into flexible, distributed capacity. Over 350 patients have been enrolled since the program&#8217;s launch six months ago, with overwhelmingly positive feedback from patients and families who describe it as a safe, high quality, and patient centered care experience. </p><h3>A virtual care chassis for deploying AI</h3><p>After demonstrating the feasibility of this virtual care model in providing high acuity care at home, the next step is to layer intelligence onto its virtual chassis. As the model grows in size and reach, it must find ways to scale care safely without inflating cost or staff burden &#8212; a challenge well suited to AI as the intelligence layer that learns from operations, anticipates patient needs, and orchestrates care across settings.</p><p>Because the program already operates through digital workflows and virtual touchpoints, it provides a natural substrate for integrating AI &#8212; data flows are continuous, interfaces are digital by default, and feedback loops can be instrumented directly into the care process.</p><p>In practice, this intelligence layer could take many forms. AI models could continuously synthesize remote monitoring data and patient messages to identify early signs of deterioration or unmet needs, prompting timely virtual interventions.  AI agents can support care coordination tasks based on these needs. Generative models could summarize longitudinal data and encounters for the care team, reducing documentation burden while improving situational awareness. Predictive algorithms could help determine which patients are suitable for home-based care and dynamically adjust visit frequency based on evolving risk. </p><p>Together, these capabilities furthers <em>Stanford Health Care at Home&#8217;s </em>ability to become a re-engineered virtual extension of inpatient care at scale.  </p><h3><strong>Why Health Systems Are Poised to Lead</strong></h3><p>I have always thought of health systems as the builders, rather than just consumers and implementers, of the tech enabled care delivery products of the future.  </p><p>Both <em>eConsult 2.0</em> and <em>Stanford Health Care at Home</em> reveal a deeper truth: the most transformative applications of AI in healthcare are not just technology deployments, but <strong>system designs</strong>. They require not just the technology, but the infrastructure, workflows, and broader care delivery systems of a health system to make them real. This is why health systems are uniquely positioned to lead. They already possess three key elements &#8212; the <strong>delivery environment</strong>, the <strong>clinical expertise</strong>, and the <strong>trust of patients and providers</strong> that enable AI to be developed, tested, and refined to redesign the ecosystems where care happens.</p><p>To build these systems, health systems must begin to see themselves not only as providers of care, but as <strong>engineers of care models</strong>. The next era of innovation will not be defined by who has the most sophisticated AI model, but by who can integrate intelligence into new architectures of care delivery.</p><h3><strong>Engineering the Care Models of the Future </strong></h3><p>We might imagine a new discipline: &#8220;<strong>Care Model Engineering&#8221;</strong> that fuses delivery science, clinical informatics, and product design into a unified practice. It treats care models the way technology organizations treat products: something that can be designed, tested, and iterated, with clear hypotheses about how structure and process drive outcomes. In this framework, AI becomes part of the design language of healthcare &#8212; a material for building systems that are more adaptive, equitable, and scalable.</p><p>For health systems, this represents an opportunity to truly bend the cost curve for healthcare by leaning into what they do well. With the right infrastructure &#8212; cross-functional product teams and a culture of iterative design that elevate the expertise inherent in all the frontline clinicians and operators who know where care breaks down &#8212; <strong>health systems can transform from slow adopters into the architects of new models of care delivery.</strong>  </p><p>AI and virtual care are the tools that make reinvention possible. What we build with them &#8212; the new systems, the new architectures of care &#8212; will determine whether we finally make healthcare scalable, sustainable, and capable of caring for the people who most need it. </p><p>More than a century after the birth of modern hospitals and clinics, we find ourselves at a moment to define what the next phase of care delivery should be.</p>]]></content:encoded></item><item><title><![CDATA[What if we had to care for twice as many people at half the cost? ]]></title><description><![CDATA[Scarcity-driven thinking may be what is needed to unlock needed innovation.]]></description><link>https://www.byte2bedside.com/p/what-if-we-had-to-care-for-twice</link><guid isPermaLink="false">https://www.byte2bedside.com/p/what-if-we-had-to-care-for-twice</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Sat, 18 Jan 2025 19:17:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/649cd92a-7b46-42b3-9807-ab079ac2b785_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In March 2020, I watched our hospital transform overnight.  </p><p>I remember sitting in a conference room listening to one of our ICU physicians describe a patient who had just been transferred: a middle age, otherwise healthy man with respiratory failure, intubated, with a positive COVID PCR test and classic findings on chest imaging consistent with COVID pneumonia.  He was the first known case of severe COVID-19 infection at our hospital.  </p><p>Those of us in the room &#8212; a group of clinicians and administrators &#8212; shared uneasy glances at each other, but continued the meeting with the level of equipoise healthcare professionals strive to maintain despite facing life or death crises.  However, beneath the calm exterior, we all knew this was just the beginning.  </p><p>Over the subsequent months as the COVID pandemic unfolded, our hospital transformed and innovated in an all hands on deck manner that was equally stressful and inspiring.  Traditional boundaries dissolved as we mobilized every resource and idea we had, knowing there was no playbook for what we were facing.  The constraints and sense of scarcity we encountered - limited PPE, hospital beds, staff shortages - didn&#8217;t paralyze us, but rather forced us to innovate that resulted in lasting transformations.  For example, telehealth adoption increased from less than 10% to 90%; while this number has since equilibrated, telehealth and virtual care have evolved from being regarded as novelties to necessary tools for care delivery.  </p><p>Looking back, the scarcity we faced didn&#8217;t just challenge us; it gave us a mandate to freely innovate, driven by an unprecedented sense of urgency that bound us together around a common mission.  </p><h1>The case for scarcity-driven innovation </h1><p>The American healthcare system is fortunate to have survived through the worst parts of the COVID pandemic, but is slowly inching towards a crisis of greater magnitude.  Ballooning healthcare costs are indebting households as well as our nation.  Providers are burning out and as a population, we are not healthier.  </p><p>Unlike COVID, this crisis is manifesting over years and decades rather than weeks and months, but shares the same accelerating, existential feel.  And just as with COVID, addressing this crisis will require us to operate under hard constraints: <strong>our healthcare system needs to deliver better care for more people at a significantly lower cost.</strong>  </p><p>Also unlike COVID, we have the benefit of time that should enable us to be more intentional and thoughtful at innovating without being forced to constantly react to immediate threats.  The scarcity we face may not be as obvious or immediate, but are very real.  We will need to decide whether to embrace this burning platform now, or wait until we have no choice. </p><p><em>How would we deliver high quality care to twice as many people at half the cost?</em>  </p><p>This is not just a thought experiment - it is a lens that forces us to fundamentally rethink care delivery and challenge assumptions and perceived barriers that systems of abundance often afford. </p><h1>Scarcity forces us to challenge core assumptions about care delivery </h1><p>When resources feel abundant, we rarely question basic assumptions about care delivery, especially if we have more to lose than to gain from changing the status quo.  During the early days of the COVID pandemic, we initially continued the practice of having the entire medical team enter every patient&#8217;s room each morning for rounds, consuming precious N95 masks and gowns to maintain a ritual that added questionable clinical value on top of what could otherwise be monitored virtually.  The looming PPE shortage forced us to reconsider when a physician truly needs to evaluate a patient at the bedside, and what instead could be substituted by a virtual interaction and asynchronous monitoring.  </p><p>We eventually created a virtual rounding system where the team would evaluate patients via video while reviewing vital trends and other clinical data.  The physician would enter the room when physical examination was truly necessary.  COVID specific clinical teams were then developed that leveraged virtual interactions, asynchronous monitoring, and standardized care protocols which enabled a single physician to efficiently care for large numbers of COVID patients.  The ICU proactively identified patients in the hospital who may require care escalation to avoid unanticipated transfers and intubations, which were risky for both the patient and providers involved.  We did explore using AI to perform these predictions, although it turned out that simply following the rate of increase in the amount of oxygen needed provided an immediately effective solution at no additional cost.  </p><p>This shift, born from necessity, revealed how many of our 'essential' practices were more ritual than requirement. The quality of care didn't suffer; in fact, many aspects improved. Teams spent more time discussing care plans and teaching. Patients weren't awakened repeatedly by large groups trooping through their rooms. What started as a response to PPE scarcity evolved into a more thoughtful approach to patient care.  </p><p>Now on the other side of the pandemic, this experience has taught us to more broadly question fundamental assumptions about care delivery: does a given patient encounter need to be in-person with a physician versus delivered virtually?  Does it require synchronous communication, or could it be handled asynchronously? Does this level of care truly require a hospital setting, or could it be safely delivered at home with the right monitoring and support?  When we strip away historical assumptions and focus on what creates real therapeutic value, we often find that traditional care models are built more on convention than necessity, creating opportunities for transformation. </p><h1>Shifting from a mindset of technology deployment to technology enabled care model design</h1><p>Achieving this type of transformation, however, requires a shift in mindset.  The current discourse around technology in healthcare suffers from a sense of abundance: we layer expensive new tools onto existing expensive care models.  The net result is more cost, more complexity, attention grabbing headlines, and ritzy conferences, but not necessarily fundamentally better care.  </p><p>Scarcity-driven thinking would force us to flip this approach.  For example, instead of asking &#8220;How can we deploy AI into cardiovascular disease care,&#8221; we would ask &#8220;what would we need to do if we had to treat twice as many patients with congestive heart failure with half as much clinic space and staffing?&#8221;  </p><p>Consider how these different mindsets play out in practice. The abundance approach might result in the addition of a fancy (and likely expensive) AI monitoring tool to existing heart failure clinic visits that may generate additional clinical encounters and interventions: all adding to cost while preserving the incumbent care model.  The scarcity approach would force us to reimagine the entire care model since there simply would not be enough staff and space for the number patients we need to treat.  We may design a system of using AI to continuously monitor patients at home, automatically adjusting medications within safe protocols or with asynchronous input from the clinical team, and reserving precious clinic time only for complex decisions that truly require in-person physician expertise. Same technology, but fundamentally different impact on care delivery and cost.</p><h1>Embracing scarcity is more often a reality at the local level than for the healthcare system at large </h1><p>Change is difficult when the scarcity is hidden.  Healthcare can feel like the proverbial frog in slowly boiling water; despite mounting evidence of crisis &#8212; rising costs, debt, workforce shortages, burnout &#8212; our system continues to operate as if resources are infinite. Current incentives reinforce this mindset: fee-for-service revenue models, technology vendors promising additive solutions, and headlines celebrating the latest expensive innovations all suggest the pie will keep growing. It's hard to embrace scarcity thinking when the water feels comfortably warm.</p><p>Yet at the local level, scarcity is already a daily reality. Individual clinics struggle with staff shortages. Hospitals face space constraints. Primary care practices buckle under growing patient panels. These frontline challenges, while less glamorous than the latest AI breakthrough, are where real innovation often happens out of necessity.</p><p>This is why frontline teams need to be empowered to lead innovation at the local level.  They intimately understand both the constraints and opportunities in their local systems, and are more likely to be driven by the problems they face rather than grand visions of technology transformation.  Yet, there are challenges - we all naturally want to preserve or expand our resources, and the case for change is difficult when there isn&#8217;t a pressing crisis.  But when given a clear mandate and permission to redesign care delivery, those who are closest to the work and experience the scarcity first hand often generate the most practical and impactful innovations.  </p><h1>We are in an all hands on deck moment</h1><p>When working in the hospital, memories from Covid sometimes trigger me to wonder, what if we suddenly have to care for twice as many patients in our medical ward?  What fundamental changes would need to occur in our workflows and care models that would make me confident our system would hold?  What if one day in the future, we are faced with the real challenge of needing to care for twice as many patients at half the cost?   </p><p>The tools and talent exist to tackle this question; what we need is the will to act before crisis forces our hand.  My bet is that in doing so, we will also generate the best, most innovative ideas to improve healthcare.  </p><p></p>]]></content:encoded></item><item><title><![CDATA[When AI outshines doctors ]]></title><description><![CDATA[AI might be becoming better than doctors at diagnostic reasoning. What now?]]></description><link>https://www.byte2bedside.com/p/when-ai-outshines-doctors</link><guid isPermaLink="false">https://www.byte2bedside.com/p/when-ai-outshines-doctors</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Sat, 23 Nov 2024 17:41:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!YqRA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In 2009, Charles Friedman proposed a &#8220;<a href="https://www.sciencedirect.com/science/article/pii/S1067502708002417">Fundamental Theorem of Informatics</a>&#8221; that stated &#8220;A person working in partnership with an information resource is &#8216;better&#8217; than that same person unassisted.&#8221;  This was the underlying principle for the value of informatics in medicine: clinicians armed with the right data driven tools will always be better at taking of patients than those without.  </p><p>What if this is no longer true with AI? </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YqRA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YqRA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 424w, https://substackcdn.com/image/fetch/$s_!YqRA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 848w, https://substackcdn.com/image/fetch/$s_!YqRA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 1272w, https://substackcdn.com/image/fetch/$s_!YqRA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YqRA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png" width="1034" height="621" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:621,&quot;width&quot;:1034,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:108599,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YqRA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 424w, https://substackcdn.com/image/fetch/$s_!YqRA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 848w, https://substackcdn.com/image/fetch/$s_!YqRA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 1272w, https://substackcdn.com/image/fetch/$s_!YqRA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15191ac0-495b-4cf3-8215-8496b9bfe9b0_1034x621.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A <a href="https://jamanetwork.com/journals/jamanetworkopen/fullarticle/2825395">recently published paper </a>from colleagues at Stanford reports a single blinded randomized controlled trial assessing whether access to a large language model (LLM) chatbot improved physicians' diagnostic reasoning. 50 physicians were randomized to an LLM-assisted group (the physician had access to a LLM chatbot such as ChatGPT in addition to conventional reference tools) or a control group (no LLM but access to conventional reference tools) and asked to assess written clinical vignettes and generate lists of diagnoses that were then later scored by a group of expert physicians.</p><p>There was no significant difference in overall diagnostic performance scores between the physician + LLM group (76%) and control group (74%). <strong>However, the LLM alone (without any physician input) was by far the most accurate of the three groups with a median score of 92%.</strong>  </p><p>The results were surprising to many in the field as it seemed to challenge the prevailing assumption that human + AI is better than human or AI alone.  In fact, physicians interacting with the AI seemed to have had a negative effect on the AI&#8217;s ability to make accurate diagnoses!  </p><h1>What then, differentiates physicians from AI, and how do we teach it? </h1><p>As an internal medicine physician, I have always taken great pride in our specialty's diagnostic reasoning skills.  It is important to recognize that we cannot draw definitive conclusions from a single study. Further, generating a differential diagnosis from a pre-written clinical vignette is also an artificial and narrow task that does not fully capture the complexities and nuances of real-world diagnostic reasoning, which requires skill in collecting and synthesizing relevant information from disparate sources over time.  </p><p><strong>However, I believe it is time to acknowledge that superior diagnostic reasoning is no longer what differentiates our value in the era of AI.  </strong>If the <a href="https://arxiv.org/abs/2001.08361">scaling laws for neural networks</a> continue to hold true, it is highly probable that in the near future, AI will consistently outperform physicians.  </p><p>I would also argue that superiority in this narrow cognitive task <em><strong>should not</strong></em> be the defining characteristic of a physician&#8217;s value.  </p><p>A physician&#8217;s job is to improve the health of our patients by preventing, treating, and managing disease.  While accurately identifying the correct diagnosis is undoubtedly a crucial step in this process, it is just one of many essential components.</p><p>What truly sets the best physicians apart from others is their capacity to deliver comprehensive, patient-centered care that goes beyond mere diagnosis.  They cultivate trusting relationships with their patients, deeply understand individual circumstances, beliefs, and preferences, and guide their patients through complex medical decisions with empathy and compassion.  They tailor their medical knowledge and expertise to the unique needs of each patient, ensuring that the diagnostic and treatment strategies are not only technically correct but also aligned with the patients&#8217; values and goals and successfully executed within a complex healthcare system.  </p><p>While some may label these qualities as the "art" of medicine, I believe this term undermines the fact that these qualities can be systematically identified, nurtured, and taught as core competencies in medical education.</p><p>As we train the next generation of physicians, we need to intentionally think about how to cultivate and prioritize these essential human skills alongside traditional medical knowledge. This means redesigning medical education to place greater emphasis on effective communication, empathy, shared decision-making, and managing complex systems. We must provide aspiring physicians with the tools and guidance to navigate complex patient interactions, build trusting relationships, and deliver personalized care that respects each patient's unique needs and values.</p><p>Nowadays, when I attend on the medical wards working with trainees, I start the week by telling them that it is not important for them to memorize any data or medical guidelines; they are encouraged to use the computer to look up any information they need, including medical references or even a LLM chatbot, when they discuss patients on rounds.  Instead, I emphasize that the primary focus is for them to leverage these tools to confidently make an assessment, reach a decision that the patient is also comfortable with, and successfully carry out the care plan as a team.  The ultimate goal is not simply getting the right diagnosis, but providing the best possible treatment to improve the patient&#8217;s health. </p><h1>We need a new paradigm for building &#8220;human + AI&#8221; teams</h1><p>Giving doctors access to an LLM chatbot is a start, but does not constitute a true human-AI team.  Designing such teams and its enabling technologies is a wide open space for exploration.  In a <a href="https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners">previous </a><em><a href="https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners">Byte to Bedside</a></em><a href="https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners"> post</a>, I discussed the concept of &#8220;cognitive integration" between AI and physicians, where the AI's capabilities are seamlessly integrated with the physician's skills, creating a complementary partnership that enhances the strengths of both.  </p><p>As AI becomes more consistently proficient at diagnosis and reasoning tasks, we must consider how these capabilities can be leveraged to enable new healthcare delivery models that can provide high-quality, accessible care to a larger population at a reduced cost.  For example, imagine a primary care practice where each physician is supported by a team of AI agents that can triage patient concerns, provide personalized health education, monitor chronic conditions, and even suggest evidence-based treatment plans. This AI-augmented approach could allow a single physician to effectively manage a much larger patient panel while maintaining high-quality, individualized care. Patients would benefit from improved access, reduced waiting times, and proactive, data-driven health management, all at a lower cost compared to how care is delivered today.  </p><p>These new AI-enabled care models will bring drastically different unit economics and scalability. It is time to move past the current model that relies on a physician personally seeing every patient to render a diagnosis and treatment plan - an approach that is not meaningfully different from 100 years ago.</p><h1>The fundamental theorem of informatics still holds true, but needs an update</h1><p><strong>Humans + computers &gt; humans or computers alone, </strong><em><strong>sometimes</strong></em><strong>.</strong>  </p><p>We need to better understand the following: 1) what the "human + AI" team looks like, and 2) what are the specific tasks for which this collaboration outperforms humans or AIs alone.  </p><p>For the task of diagnostic reasoning from text vignettes where the human + AI team is just giving physicians access to a LLM chatbot, AI may be giving physicians a run for our money.  But for the task of taking care of patients, I am betting on the human + AI team.  The work now is to figure out how to be build the team.  </p><h1>References</h1><p><em>Goh E, Gallo R, Hom J, et al. Large Language Model Influence on Diagnostic Reasoning: A Randomized Clinical Trial. JAMA Netw Open. 2024;7(10):e2440969. doi:10.1001/jamanetworkopen.2024.40969</em></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Empowering individual health systems to validate AI models ]]></title><description><![CDATA[Few would have expected Epic to be a first mover in the decentralization and democratization of AI in healthcare.]]></description><link>https://www.byte2bedside.com/p/empowering-individual-health-systems</link><guid isPermaLink="false">https://www.byte2bedside.com/p/empowering-individual-health-systems</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Sun, 02 Jun 2024 19:24:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/35c72efa-8fc6-4331-923f-68dded925733_295x171.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Epic <a href="https://www.fiercehealthcare.com/ai-and-machine-learning/epic-releases-ai-validation-software-health-systems">recently announced an open source &#8220;AI trust and assurance software suite&#8221;</a> that health systems can use to validate and monitor ML models integrated into the EHR.  This tool automates the most time-consuming and replicable steps of model validation, such as collecting, aggregating, and mapping data and metrics.  By streamlining these processes, it enables health systems without robust data science teams to perform necessary validations and ensure they can use ML models safely and effectively for their patients.</p><p><strong>What does model validation entail, and why is this software significant?</strong></p><p>Consider the following example: a hospital interested in implementing a new readmission risk model must first ensure that the model accurately predicts readmissions for its patient population without significant biases. This process requires building a clean dataset of a sufficiently large cohort of patients and identifying the "true positives"&#8212;the patients who were actually readmitted&#8212;using a standard definition (i.e., inpatient readmission at 30 days). The ML model is then used to generate predictions for each patient from that cohort (typically a numerical score classified into "likely or unlikely to be readmitted" based on some pre-specified cutoff). The hospital must compare these predictions against the actual outcomes to assess the model's recall and precision (measures of accuracy), and potential biases across different patient subgroups.</p><p>This process is resource-intensive, requiring significant data management and analytical expertise. Epic's software automates key aspects of this process, making it more accessible for health systems that lack extensive data science resources, but can now leverage knowledge of local patient populations, clinical workflows, and organizational needs to better assess the model&#8217;s safety and effectiveness.  For instance, local clinicians might know that certain socioeconomic factors or comorbidities prevalent in their community significantly influence readmission rates, and that workflow constraints require a certain level of model precision for the model to be effectively used.  Real time monitoring of model performance also enables model and workflows to be quickly adjusted based on local needs.  </p><p>The ability to assess the model locally empowers the users of these models &#8212; the clinicians and staff caring for the patients the AI intends to benefit &#8212; to have the necessary input in the application of AI into their work.  Their input also enables the industry to gather valuable real-world experience needed to inform standards and frameworks for governing the use of AI in healthcare.  </p><p><strong>What&#8217;s in it for Epic? </strong></p><p>It would be naive to assume that Epic does not have strategic motives behind this release.  It has been interesting to see <a href="https://www.byte2bedside.com/p/clinical-trials-matching-how-the">Epic&#8217;s evolution from a EHR company into a platform company</a>, and I believe this move further solidifies their positioning as the dominant healthcare technology platform.  Epic acknowledges that they may not be the sole producer of AI for healthcare, but they would like all producers and users of AI in healthcare to depend on Epic (or a standard that Epic establishes).  I am interested to see the extent to which this validation tool will be adopted; given Epic&#8217;s level of scale, depth of workflow integration, and customer stickiness among health system, I would bet in their favor.  </p><p></p><p></p><p></p><p></p><p></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Monthly Roundup: Summarizing clinical information with LLMs, digital care models bypassing traditional healthcare providers, user centered design for designing AI tools ]]></title><description><![CDATA[Takeaways from articles about research, case study spotlights, and industry trends]]></description><link>https://www.byte2bedside.com/p/monthly-roundup-summarizing-clinical</link><guid isPermaLink="false">https://www.byte2bedside.com/p/monthly-roundup-summarizing-clinical</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Thu, 07 Mar 2024 18:34:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!H2VJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>Large language models perform better than medical experts in summarizing clinical information </h1><p>Many outside of healthcare underestimate how many tasks performed by highly trained experts simply involve summarizing and communication information to another individual.  Offloading this chunk of work to AI will have a transformative impact on healthcare.  </p><p>This <a href="https://www.nature.com/articles/s41591-024-02855-5">study</a> from Stanford describes an analysis comparing the performance of LLMs against medical experts in performing a set of text summarization tasks from the following data sources: radiology reports, progress notes, patient questions, and patient-provider conversations. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H2VJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H2VJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 424w, https://substackcdn.com/image/fetch/$s_!H2VJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 848w, https://substackcdn.com/image/fetch/$s_!H2VJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 1272w, https://substackcdn.com/image/fetch/$s_!H2VJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H2VJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png" width="1364" height="993" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:993,&quot;width&quot;:1364,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:519939,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H2VJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 424w, https://substackcdn.com/image/fetch/$s_!H2VJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 848w, https://substackcdn.com/image/fetch/$s_!H2VJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 1272w, https://substackcdn.com/image/fetch/$s_!H2VJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F297ab2e2-9544-4d20-9ad9-98948ecc9755_1364x993.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The performance metrics included completeness, correctness, conciseness, and prevalence of fabricated information.  In 81% cases, the LLM generated summaries were rated either equivalent (45%) or superior (36%) than those generated by their human counterparts. </p><p>While more studies will need to be done to validate this result, I think this is an important milestone in establishing the viability of LLMs for safely and effectively summarizing clinical text, which is a foundational and extensible task that underlies many healthcare workflows.  Importantly, the study highlights that the benchmark for LLM performance is not perfection, but rather human experts.  As a practicing physician, I can say that we are far from perfect! </p><h1>Amazon and Eli Lilly launching new programs that directly identify and provide care to patients with chronic diseases</h1><p>The landscape of chronic disease management is changing.  While primary care providers (PCPs) remain the primary frontline access points for healthcare, nontraditional industry players such as <a href="https://health.amazon.com/health-condition-programs?ref_=abt_amzn_plrs">Amazon</a> and <a href="https://lillydirect.lilly.com/?gclid=CjwKCAiAiP2tBhBXEiwACslfntStZfjXHuss8aPWQ1uCU_scGmf5kuSiIn-aneGTAm5dh5Cf-4q1iBoCJJkQAvD_BwE">Eli Lilly</a> are creating their own direct channels to patients.  Strategically, these companies are betting that they can deliver care more efficiently by bypassing traditional healthcare delivery channels and employ a more digital first, consumer directed approach.  These tend to focus on simplicity and efficiency of access rather than solving for complex medical issues.  For Eli Lilly and potentially other biopharma companies, this is also an extension of their sales and marketing strategy for blockbuster medications (eg. semaglutide) by removing barriers to patients being prescribed those drugs.  </p><p>These efforts are currently limited and only represent a small sliver of healthcare delivery in the US.  However, incumbent health systems need to take notice, if only to observe and assess the demand for such services, as it will likely come from patients disillusioned with the inefficiencies of existing systems and searching for alternatives. I see a likely rise in the number of strategic partnerships between incumbent health systems looking to outsource certain parts of care pathways to these digital first companies in exchange for maintaining referral channels to complex specialty care.   </p><p><strong>Yet, the key to success for these efforts is clinical integration.</strong>  These &#8220;direct to consumer&#8221; digital first care models often fall short in serving patients with more complex medical histories and needs.  Even for the patient who may be just looking for Ozempic to lose weight, they likely have other comorbid medical problems that need to be assessed and managed.  How these digital first companies integrate their care with the rest of the healthcare ecosystem will be a critical factor towards their success.  Further, it will be interesting to see how they manage healthcare data.  Will they join nationwide <a href="https://www.hhs.gov/about/news/2023/12/12/hhs-marks-major-milestone-nationwide-health-data-exchange.html">efforts around interoperability and data exchange</a> that have created significant progress towards enabling transparency and sharing of medical data among healthcare providers, or will new &#8220;walled gardens&#8221; be formed where medical data collected by Amazon clinics will be sequestered within the &#8220;Amazon health ecosystem&#8221;?  </p><h1>A framework for designing an AI guided tool for communicating prognosis </h1><p>Physicians often navigate conversations with patients laden with complexity, uncertainty, and deep emotions that require more than simply knowing the correct information.  For example, communicating prognosis for a patient diagnosed with cancer involves more than just conveying an accurate survival rate; it requires translating and framing statistical outcomes into a narrative that resonates with the broader context of the patient&#8217;s life, hopes, and fears.  </p><p>There exist many machine learning models that predict survival with the intention of eventually being integrated into clinical workflows, such as communicating prognosis, but simply presenting a survival probability often falls short of what is actually useful to a physician.  <strong>The key lies in designing software that harmoniously blends ML predictions with additional information and design features that help providers create the narratives needed to drive effective conversations.</strong>  </p><p>The authors of this <a href="https://academic.oup.com/jamia/article/31/1/174/7320060">study</a> from the University of Utah developed a tool aimed at assisting oncologists in discussing prognosis with patients facing advanced solid tumors using a user centered design approach to integrate 6 month survival predictions into a tool that addresses these broader needs.  The team underwent several rounds of iterative design sessions with oncologists.  <strong>What stood out to me was how they intentionally sequenced their design steps in partnership with the end users to build towards the final product:</strong> </p><ol><li><p><strong>Use of Initial Interfaces</strong>: In rounds 1 and 2, initial interfaces provided a common visual and terminology for facilitating discussions between the team and clinician input.</p></li><li><p><strong>Content Determination and Model Clarification</strong>: These interfaces helped in deciding what content should be presented, clarified the function of the model, and established a suitable threshold for classifying survival risk.</p></li><li><p><strong>Exposure of Misassumptions</strong>: Misassumptions among clinical and technical experts were exposed, such as the belief that the model output represented expected survival rather than a risk classification.</p></li><li><p><strong>Triggering of Questions and Additional Information</strong>: The discussions prompted questions and led to the presentation of additional information by technical experts, enabling further exploration of the model and its applications.</p></li><li><p><strong>Clinical Relevance and Feature Assessment</strong>: Clinicians on the study team assessed the clinical relevance, examined the features of the model, and reviewed the gold standard mortality data.</p></li><li><p><strong>Advice on Data Pre-processing</strong>: Clinicians advised on data pre-processing, highlighting the importance of clinician involvement in AI system design.</p></li><li><p><strong>Application of Design Decisions</strong>: Based on the findings from rounds 1 and 2, design decisions were applied to enhance the model and interface, focusing on trust and transparency and ensuring a match between the system and the real world.</p></li><li><p><strong>Progression of Interim Interfaces</strong>: The evolution of interim interfaces through these rounds was documented, showcasing the iterative design process and improvements made.</p></li></ol><p>This iterative, dyadic partnership model facilitates the optimization of critical design and engineering decisions early on, based on what is genuinely useful to oncologists. This approach is crucial as many insights from end users often emerge only when they are presented with a visual representation, allowing spontaneous comments to be transformed into specific technical requirements.</p><p>Here is an example of a comment from an oncologist during a design session after viewing an early user interface: </p><blockquote><p>&#8220;As oncologists we sometimes have rose-colored glasses on and like to overestimate the benefits of second, third, fourth line treatment. And so, I think that this can maybe ground you and bring you back to reality like, hey, look it&#8217;s probably not such a good idea. Let&#8217;s think about alternatives.&#8221; (#7) &#8220;I think this would be useful for family members.&#8221; (#3) &#8220;&#8230;helpful in&#8230;patients who&#8230;see cancer as like a battle that they have to fight&#8230;often unwilling to stop treatment despite all evidence that treatment might harm them rather than help them.&#8221;</p></blockquote><p>A key insight here is how this prognosis communication tool can add value by helping oncologists recalibrate expectations and confront optimistic biases towards aggressive treatment courses, which is a specific use case that may require tailored design choices.  </p><p>Communicating with patients in high stakes situations is complex and requires the art of marrying data with human connection so that the patient not only receives the correct information, but also feels genuinely heard and supported.  Digital and AI enabled tools that support these interactions need to be designed with a deep understanding of the nuances of patient-provider communication that can only come from a solid dyadic partnership between the end-users and technical teams established from the outset.  </p><h1>References </h1><p>Van Veen, D., Van Uden, C., Blankemeier, L. <em>et al.</em> Adapted large language models can outperform medical experts in clinical text summarization. <em>Nat Med</em> (2024).</p><p>Catherine J Staes, Anna C Beck, George Chalkidis, Carolyn H Scheese, Teresa Taft, Jia-Wen Guo, Michael G Newman, Kensaku Kawamoto, Elizabeth A Sloss, Jordan P McPherson, Design of an interface to communicate artificial intelligence-based prognosis for patients with advanced solid tumors: a user-centered approach, <em>Journal of the American Medical Informatics Association</em>, Volume 31, Issue 1, January 2024, Pages 174&#8211;187</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.byte2bedside.com/subscribe?"><span>Subscribe now</span></a></p><p></p><p></p><p></p><p></p><p></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Monthly Roundup: ChatGPT and CDS, virtual specialty care and payers, unified financial platforms]]></title><description><![CDATA[Takeaways from articles about research, case study spotlights, and industry trends this month.]]></description><link>https://www.byte2bedside.com/p/monthly-roundup-chatgpt-and-cds-virtual</link><guid isPermaLink="false">https://www.byte2bedside.com/p/monthly-roundup-chatgpt-and-cds-virtual</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Wed, 05 Jul 2023 07:12:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong><a href="https://academic.oup.com/jamia/article/30/7/1237/7136722">Using AI-generated suggestions from ChatGPT to optimize clinical decision support</a></strong></h2><p>I found this paper interesting because it offers a fresh perspective on applying AI and large language models (LLMs) to clinical decision support (CDS). Led by Siru Liu at Vanderbilt University Medical Center (VUMC), the study uses ChatGPT to generate suggestions for refining the <em>logic</em> (rather than the content) of CDS alerts. This approach is intriguing because it integrates AI into an existing human-driven process, making it a practical and potentially beneficial application in the near term for healthcare provider organizations.</p><p>To understand why this approach is significant, let&#8217;s first consider the &#8220;anatomy&#8221; of traditional CDS systems. These systems have a &#8220;backend&#8221; comprising clinical data, a knowledge base, and logic, and a &#8220;frontend&#8221; that includes the user interface and subsequent action. When we talk about integrating AI into CDS systems, it&#8217;s crucial to specify which part of this anatomy we&#8217;re optimizing. Often, the focus is on training the AI, but the intentional design of the CDS system is equally important.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0aFl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0aFl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 424w, https://substackcdn.com/image/fetch/$s_!0aFl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 848w, https://substackcdn.com/image/fetch/$s_!0aFl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 1272w, https://substackcdn.com/image/fetch/$s_!0aFl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0aFl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png" width="532" height="252" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:252,&quot;width&quot;:532,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!0aFl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 424w, https://substackcdn.com/image/fetch/$s_!0aFl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 848w, https://substackcdn.com/image/fetch/$s_!0aFl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 1272w, https://substackcdn.com/image/fetch/$s_!0aFl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1fbd0b89-007f-424e-81e3-9976f3459923_532x252.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>An example of a set of logic for a CDS system to alert providers to avoid live virus immunizations for immunocompromised patients</em>.</p><p><a href="https://academic.oup.com/jamia/article/30/7/1237/7136722">Source: Liu et al. JAMIA 2023</a></p><p>While AI is often touted as a potential replacement for knowledge bases and logic, the reality is that current machine learning models fall short in being able to be applied to CDS at scale. Yet, there remains a need to change how CDS systems are built; the process of creating and maintaining these knowledge bases is time consuming, involving synthesis and translation of expert knowledge into discrete knowledge artifacts and rules. This is a task that seems well-suited for a large language model.</p><p><a href="https://academic.oup.com/jamia/article/30/7/1237/7136722">The study at VUMC</a> provides a great example of this intentional design led approach to applying AI to CDS. The researchers used seven alerts from the Epic EHR system and transformed the documented alert logic into ChatGPT prompts. They then mixed LLM-generated suggestions with those previously generated by clinical informaticians, randomizing the order of suggestions. The suggestions were rated by a group of physicians and pharmacists on eight different perspectives, including understanding, relevance, and usefulness.</p><p>The results of the study were promising. LLM-generated suggestions scored high in understanding and relevance, suggesting that AI can provide valuable insights that can enhance the effectiveness of CDS alerts. However, the study also highlighted some challenges with AI-generated suggestions, including lack of knowledge management and implementation understanding. These findings underscore the need for further research and refinement of LLMs for CDS optimization.</p><p>Looking ahead, the integration of AI into CDS optimization could pave the way for more dynamic, responsive, and effective CDS systems. As we continue to explore this opportunity, it is clear that the journey is just as important as the destination.</p><p>Reference: https://academic.oup.com/jamia/article/30/7/1237/7136722</p><h2><strong><a href="https://www.cvshealth.com/news/virtual-care/aetna-and-oshi-collaborate-to-provide-virtual-care-for-digestive-disorders.html">Aetna and Oshi collaborate to provide virtual care for digestive disorders</a></strong></h2><p>This is news back from April, but it recently caught my eye. In a move towards vertical integration into specialty care, Aetna announced a value based collaboration with <a href="https://oshihealth.com/">Oshi Health</a>, a virtual digestive health provider. Oshi&#8217;s digital care model in involves connecting patients with specialized providers via telehealth visits, and continuously adapting and refining care plans based on monitoring of symptoms, food intake, and bowel movements with the support of health coaches and nutritionists. Early results presented in a <a href="https://www.prweb.com/releases/clinical_trial_results_demonstrate_oshi_healths_multidisciplinary_gi_care_improves_patient_outcomes_with_savings_greater_than_10_000_per_patient_within_six_months/prweb19142794.htm">clinical study</a> look promising in terms of member satisfaction and cost savings. The partnership will provide Aetna commercial members with in-network access to Oshi&#8217;s services. It is unclear whether/how Aetna is building in any features to &#8220;navigate&#8221; its members towards these virtual care options versus traditional specialty referrals.</p><p>This partnership follows its recent acquisitions of <a href="https://www.signifyhealth.com/">Signify Health</a> (home health), and <a href="https://www.oakstreethealth.com/">Oak Street Health</a>. At $8 billion and $10.6 billion, respectively, these are much more aggressive and higher conviction moves than the partnership with Oshi. However, the trend and strategy is clear: CVS/Aetna is betting on vertical integration of its payer business with virtual first provider capabilities, which now includes not just supportive and primary care, but specialty care services.</p><p>To those in the health technology space, this presents unique opportunities and problems to be solved, ranging from referral optimization, care coordination, and even clinical decision support at a very different scale than with traditional provider organizations. The breadth and depth of products (and the data that will drive these products) will be different. I&#8217;ll be keeping my eye out for products and companies that emerge in this space.</p><p>Reference: https://www.cvshealth.com/news/virtual-care/aetna-and-oshi-collaborate-to-provide-virtual-care-for-digestive-disorders.html</p><h2><strong><a href="https://www.pymnts.com/healthcare/2023/79percent-of-consumers-want-to-pay-all-medical-bills-from-one-digital-platform/">79% of Consumers Want to Pay All Medical Bills From Single Digital Platform</a></strong></h2><p>A survey of 2,034 consumers in the United States revealed that participants by and large prefer a unified digital platform to help with financial transactions related to healthcare, including payment of medical bills, financing options, and HSA investment options.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_8D_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_8D_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 424w, https://substackcdn.com/image/fetch/$s_!_8D_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 848w, https://substackcdn.com/image/fetch/$s_!_8D_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 1272w, https://substackcdn.com/image/fetch/$s_!_8D_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_8D_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png" width="580" height="498" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:498,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!_8D_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 424w, https://substackcdn.com/image/fetch/$s_!_8D_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 848w, https://substackcdn.com/image/fetch/$s_!_8D_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 1272w, https://substackcdn.com/image/fetch/$s_!_8D_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b5e943b-b718-44a4-a6e9-b366cad0ea46_580x498.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Of note, this study was conducted by a healthcare fintech company <a href="https://www.lynx-fh.com/">Lynx</a> and <a href="https://www.pymnts.com/">Pymnts</a>, a fintech news outlet, so I cannot speak to its degree of bias and validity. Nevertheless, it highlights two interesting trends that I believe those of us in health IT should continue to track closely: 1) the experience of paying for healthcare is a major pain point, and 2) patients are looking for more control and convenience over the process, rather than deferring it to the provider organizations.</p><p>Managing their own healthcare becomes an arduous labyrinth for patients juggling multiple providers, health systems, and pharmacies, each with its unique policies, paperwork, and billing processes; a mounting frustration that only exacerbates the stress of navigating their medical problems. I think the sentiment from this survey represents a desire for a future state towards a model where patients manage healthcare transactions on their terms &#8212; not just the payment process, but also financing, access to tax deductible investment accounts, and I would also guess price transparency (which is not explicitly mentioned in the survey).</p><p>This may also imply a shift in preferences in the tradeoff between privacy and convenience. Traditionally, privacy has been a paramount concern in healthcare. Still, the growing demand for a simplified and unified payment system may cause consumers to compromise on this front.</p><p>Patients are expressing their frustration with legacy health systems and processes, and as health IT professionals, we need to be listening. The desire for control and convenience is redefining the way we must think about our systems, processes, and the products that we are building for the future. Perhaps patients are telling us something a bit more nuanced about the tradeoffs that they are willing to make: while they undoubtedly value their privacy, their increasing desire for convenience, control, and an integrated healthcare experience suggests that they might be open to a more flexible model of data management where transparency, security, and user-friendliness coexist.</p><p>Reference: https://www.pymnts.com/healthcare/2023/79percent-of-consumers-want-to-pay-all-medical-bills-from-one-digital-platform/</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.byte2bedside.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Monthly roundup: price transparency integration, large language model pilots, explainable AI for clinicians]]></title><description><![CDATA[Takeaways from articles about research, case study spotlights, and industry trends this month.]]></description><link>https://www.byte2bedside.com/p/monthly-roundup-price-transparency</link><guid isPermaLink="false">https://www.byte2bedside.com/p/monthly-roundup-price-transparency</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Sat, 27 May 2023 07:08:00 GMT</pubDate><content:encoded><![CDATA[<p></p><h2><a href="https://www.mobihealthnews.com/news/amazons-one-medical-partners-rightway-expand-primary-care-access?mc_cid=1170ee1897&amp;mc_eid=be4e5cbad0">Amazon&#8217;s One Medical partners with Rightway to expand primary care access</a></h2><p>Similar to many of my physician colleagues, I&#8217;ve developed a level of learned helplessness when attempting to navigate cost of care for my patients. Healthcare costs are often just as opaque to us as they are to our patients. While we are trained to ignore the costs and just focus on what is clinically indicated when recommending treatment plans, that is not practical in the real world.</p><p>Price transparency efforts in healthcare have had mixed success &#8212; mostly due to misaligned incentives as well as real data and technical barriers. I&#8217;m curious what this partnership between Amazon&#8217;s One Medical, a tech-forward primary care provider, and Rightway, a care navigation/price transparency platform as well as a pharmacy benefits manager. Amazon is all about capturing the customer value chain; I wonder how much of that ethos is reflected in this strategy, which, in the best case scenario, would provide a level of vertical integration across care navigation, care delivery, drug benefits, and price transparency centered on the patient user.</p><p>Successful technical/process integration with sufficiently clean data is also harder than it seems; perhaps a tech forward provider organization like Amazon One Medical will have a leg up over legacy health systems and primary care clinics. I look forward to seeing what comes out of this partnership.</p><h2><a href="https://www.beckershospitalreview.com/ehrs/4-health-systems-piloting-microsoft-epics-generative-ai.html">4 health systems piloting Microsoft, Epic&#8217;s generative AI</a></h2><p>There are now four large health systems piloting <a href="https://medcitynews.com/2023/04/epic-to-integrate-gpt-4-into-its-ehr-through-expanded-microsoft-partnership/">Epic&#8217;s GPT-4 enabled large language model product</a>, which currently is limited to a use case with in-basket messages. However, Epic has a few other products in the pipeline: a large language model (LLM) powered database querying tool that enables self service analytics (a big game changer for clinicians who otherwise would have to wait in a queue to get data, a significant bottleneck to quality improvement and research), as well as other more experimental features like chart summarization.</p><p>I would not underestimate Epic in the healthcare AI race. Even though they may not have the most advanced in-house data science/AI capabilities, they have several significant advantages:</p><ul><li><p>Their deep understanding of healthcare processes and relationships with provider organizations will allow them to build AI products that are not necessarily the most cutting edge, but will be exactly what the folks with the checkbook are willing to pay for.</p></li><li><p>I would not discount Epic&#8217;s product vision. The way they&#8217;ve designed the in-basket LLM product is clever; there is built-in backend prompt engineering that generates standard types of message responses, which reduces variability, risk, and the need for physicians to fiddle around with different LLM prompts &#8212; all very attractive features. The SlicerDicer database query tool is also a very practical application of LLMs for health systems. Both represent Epic&#8217;s deep understanding of what their customers want, as well as sufficient expertise in AI to build those products.</p></li><li><p>Foundational models like GPT-4 are leveling the playing field for AI companies. Having a better AI model is no longer a moat. Epic historically had been at a disadvantage because their in-house AI models were indeed subpar; however, with GPT-4 integration and their strategic partnership with Microsoft Azure, this is no longer the case (at least with LLMs). They can now apply the most advanced AI capabilities into what is their moat, which is data access, integration into health systems, and scale. Their early moves so far signal that they are well aware of this advantage and are willing to act on it.</p><p></p></li></ul><h2><a href="https://www.nature.com/articles/s41746-023-00837-4">Solving the explainable AI conundrum by bridging clinicians&#8217; needs and developers&#8217; goals</a></h2><p>There has been a ton of work put into developing &#8220;explainable AI&#8221; for healthcare, but &#8220;explainability&#8221; means something different depending on whether you talk to an AI engineer or a clinician end user. Here is a <a href="https://link.springer.com/chapter/10.1007/978-3-031-09108-7_8">book chapter</a> I wrote with my colleagues that provides some background.</p><p>This research paper reports a qualitative study that explores the ways AI developers and physician users differ in what they find important in terms of explainability. I found this article valuable because 1) it endorses the model of developers and clinician users collaboratively building together, which I think is critical to creating useful products in healthcare, and 2) presents a concrete framework for how to design for explainable AI systems. A few takeaways from the study:</p><ul><li><p>While developers focused on computational measures for interpretability (eg. <a href="https://shap.readthedocs.io/en/latest/example_notebooks/overviews/An%20introduction%20to%20explainable%20AI%20with%20Shapley%20values.html">Shapely values</a> to describe feature contributions of models), clinicians found those measure unhelpful, and instead just wanted to know whether/how the machine learning predictions make sense clinically. For clinicians, model explainability means providing clinical context for the model prediction.</p></li><li><p>AI developers and clinicians often diverge in their perceptions of the most reliable source of information. Developers tend to concentrate solely on the data used for training the model, viewing it as the ultimate source of truth. On the other hand, clinicians perceive this dataset as just one component of a broader narrative, which is significantly influenced by patient interactions and data not incorporated into the dataset. This underscores the previous observation that clinicians generally prefer context and real-world application over precise mathematical justifications when interpreting model predictions.</p></li><li><p>Developers and clinicians have different mindsets when it comes to the purpose of ML models. Developers are focused on exploring new knowledge and discovering unknown patterns in data, while clinicians more typically see these models as helping them retrieve and apply established knowledge and evidence. This difference in mindset needs to be recognized when designing AI systems, as developers and clinicians may have different priorities and goals.</p></li></ul><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.byte2bedside.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI and physicians as “thought partners” with large language models]]></title><description><![CDATA[Our work in medicine is driven by narratives.]]></description><link>https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners-e3c</link><guid isPermaLink="false">https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners-e3c</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Wed, 15 Feb 2023 07:58:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!uYJp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Our work in medicine is driven by narratives.</p><p>Whether it is figuring out a diagnosis, formulating a treatment plan, or simply learning about a patient&#8217;s life, we construct narratives to understand and exchange ideas in order to provide care. At the core of this interaction is human language, which enables the rich and high dimensional information exchange that we have evolved to be naturally good at.</p><p>Amidst all the buzz about large language models (LLMs) and system such as ChatGPT, what stands out to me the most is not the their information retrieval capabilities (which is impressive albeit with limitations) or that it even represents a new form of &#8220;intelligence.&#8221; <a href="https://twitter.com/ylecun">Yann LeCun</a>, Chief AI Scientist at Meta and an early pioneer of convolutional neural nets, articulates these limitations well:</p><p>Rather, it is that LLMs now enable humans to exchange complex ideas with AI systems in a way that is more like how we talk to each other. This is a big change from other machine learning (ML) models that only give quantitative predictions or discrete classifications.</p><p>In other words, LLMs provide a new &#8220;high fidelity API&#8221; between humans and AI.</p><h2><strong>&#8220;Cognitive integration&#8221; and why it is difficult with current AI systems</strong></h2><p>In a <a href="https://www.byte2bedside.com/2022/08/ai-and-physicians-as-thought-partners-a-use-case-in-endoscopic-colorectal-cancer-screening/">previous </a><em><a href="https://www.byte2bedside.com/2022/08/ai-and-physicians-as-thought-partners-a-use-case-in-endoscopic-colorectal-cancer-screening/">Byte to Bedside </a></em><a href="https://www.byte2bedside.com/2022/08/ai-and-physicians-as-thought-partners-a-use-case-in-endoscopic-colorectal-cancer-screening/">post</a>, I wrote about the concept of &#8220;thought partnership&#8221; and &#8220;cognitive integration&#8221; of AI into physician workflows.</p><p>&#8220;Traditional workflow integration strategies focus on deploying technologies based on how they can streamline physician tasks. However, this task oriented approach misses a critical source of value for AI solutions, which is the integration of AI into how physicians think (and vice versa). This &#8220;cognitive integration&#8221; will be increasingly relevant as AI systems become more intelligent. The use case described above is a good example of how AI and physicians can team up as &#8220;thought partners&#8221; to enhance each other&#8217;s performance.&nbsp;&#8220;</p><p>True &#8220;thought partnership&#8221; requires a bidirectional, collaborative relationship where each party brings a unique perspective to tackle a complex problem. The output is greater than the sum of its parts because of the dynamic &#8220;back and forth&#8221; that leads to new and sometimes unexpected insights. This is what makes being a physician challenging and fun &#8211; we are surrounded by thought partners (our fellow clinicians, support staff, and patients and their families), which allows us to continuously learn and tackle the complex problems we encounter in our practice.</p><p>It is difficult to achieve this type of thought partnership with ML models that only produce predictions and classifications. The predictions may be highly accurate and come from complex, high dimensional models, but the output is still one dimensional: a single number. Regardless of how accurate that number is, it is far from able to enable the complex, dynamic high dimensional information exchange that human beings expect and need in order to collaboratively solve difficult problems.</p><p>This limitation, in my view, is a key reason why many clinician-facing AI systems in healthcare have not gained adoption &#8211; the unidirectional deployment of a number (regardless of how accurate it is) does not fit into the complexity of patient care. It also does not seem very &#8220;intelligent.&#8221;</p><h2><strong>Language as a &#8220;high bandwidth&#8221; interface between humans and AI systems</strong></h2><p>Soon after its launch, ChatGPT spread like wildfire among physician circles. This was something special &#8211; after long years of AI innovators struggling to find use cases and gain traction in healthcare, physicians instead were excitedly sharing self discovered use cases in their practice with ChatGPT. <a href="https://tillthecavalryarrive.substack.com/p/the-time-saving-magic-of-chat-gpt">This post</a> by a urologist, <a href="https://twitter.com/CanesDavid">David Canes</a>, is one of my favorites &#8211; a great mix of real world clinical and operational uses inspired by the needs of a practicing physician.</p><p>What is different this time? ChatGPT is indeed impressively accurate, but no more accurate than a high quality web search. It does not generate new information or predictions (even though it seems like it does sometimes). The difference is that we can now interact with an AI system with human language. It feels more accessible and attractive because our brains are wired that way &#8212; we evolved to communicate, learn, and innovate through language. ChatGPT unlocked an exponentially higher bandwidth link between humans and AI systems.</p><h2><strong>An example of &#8220;thought partnership&#8221; with Chat GPT</strong></h2><p>Consider the following example of how I, as a physician, may use a LLM application such as ChatGPT. I also look forward to trying this out with Google&#8217;s <a href="https://arxiv.org/abs/2212.13138">Med-PaLM</a>!</p><p>I start by consulting ChatGPT about a hypothetical patient I am caring for with heart failure.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uYJp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uYJp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 424w, https://substackcdn.com/image/fetch/$s_!uYJp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 848w, https://substackcdn.com/image/fetch/$s_!uYJp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 1272w, https://substackcdn.com/image/fetch/$s_!uYJp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uYJp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png" width="580" height="550" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:550,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!uYJp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 424w, https://substackcdn.com/image/fetch/$s_!uYJp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 848w, https://substackcdn.com/image/fetch/$s_!uYJp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 1272w, https://substackcdn.com/image/fetch/$s_!uYJp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a518fbd-c020-40e8-bcb0-692e9ab215fa_580x550.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The output is fairly accurate, although reads like a laundry list of generic recommendations. However, I focus on step #5 given the patient&#8217;s complaint of chest pain, which reminds me that I may be missing a myocardial infarction (heart attack).</p><p>The exchange continues:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tLRP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tLRP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 424w, https://substackcdn.com/image/fetch/$s_!tLRP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 848w, https://substackcdn.com/image/fetch/$s_!tLRP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 1272w, https://substackcdn.com/image/fetch/$s_!tLRP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tLRP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png" width="580" height="539" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:539,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!tLRP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 424w, https://substackcdn.com/image/fetch/$s_!tLRP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 848w, https://substackcdn.com/image/fetch/$s_!tLRP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 1272w, https://substackcdn.com/image/fetch/$s_!tLRP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F450fb6df-e463-4158-bfe3-7b9748cef8c3_580x539.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The conversation takes a turn. I know how to treat an MI so do not necessarily need ChatGPT to tell me specifically which medications to administer. However, its output reminded me that beta blockers may need to be used with caution in a patient with both heart failure and myocardial infarction. I am unsure about the evidence, so ask it to synthesize the latest research. Here, it would have been great if I could directly navigate to those papers (or see snippets of summaries) &#8211; a feature I am sure is in the works by some company out there.</p><p>Based on my assessment, I do decide to administer beta blockers to the patient, but I remember that his son had expressed anxiety and hesitancy to me when I had mentioned beta blockers earlier.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0rBK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0rBK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 424w, https://substackcdn.com/image/fetch/$s_!0rBK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 848w, https://substackcdn.com/image/fetch/$s_!0rBK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 1272w, https://substackcdn.com/image/fetch/$s_!0rBK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0rBK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png" width="580" height="524" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:524,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!0rBK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 424w, https://substackcdn.com/image/fetch/$s_!0rBK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 848w, https://substackcdn.com/image/fetch/$s_!0rBK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 1272w, https://substackcdn.com/image/fetch/$s_!0rBK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c4ee8ca-8e70-4985-9e60-79a45774c129_580x524.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>These explanations are pretty good and incorporate the use of patient friendly language that many physicians are not good at applying.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kMDl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kMDl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 424w, https://substackcdn.com/image/fetch/$s_!kMDl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 848w, https://substackcdn.com/image/fetch/$s_!kMDl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 1272w, https://substackcdn.com/image/fetch/$s_!kMDl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kMDl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png" width="580" height="476" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:476,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!kMDl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 424w, https://substackcdn.com/image/fetch/$s_!kMDl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 848w, https://substackcdn.com/image/fetch/$s_!kMDl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 1272w, https://substackcdn.com/image/fetch/$s_!kMDl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13163ea6-70b3-474e-bcd9-5acf61af97ec_580x476.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>ChatGPT then helps me and the patient&#8217;s son remember that the patient was previously on a medication called labetalol. It also reminded me that there is a beta blocker named levobunolol (as an internist, I often forget the names of eye drops&#8230;).</p><p>The purpose of this example is to illustrate three points:</p><p>1) a patient encounter can consistent of a complex, non linear set of tasks and decisions,</p><p>2) back and forth dialogue between a physician and an AI system can help navigate through these tasks and decisions, and</p><p>3) the richness of information conveyed through language that far exceeds what would otherwise be possible if the AI system was only capable of putting out a discrete classification or quantitative prediction.</p><h2><strong>A call to designers and builders: create AI enabled products that can facilitate thought partnership with language.</strong></h2><p>We will like using it because this is how our brains are wired. The accuracy of the output from the AI system is of course important and sources, bias, errors will need to be transparently conveyed while preserving usability. However, the capability to actively exchange information via human language is, in my view, what is transformative.</p><p>What features, designs, and UIs will enable true &#8220;thought partnership&#8221; between physicians and AI systems? I look forward to finding out.</p>]]></content:encoded></item><item><title><![CDATA[Clinical trials matching: how the Epic EHR is becoming a platform]]></title><description><![CDATA[Epic is ramping up its platform play.]]></description><link>https://www.byte2bedside.com/p/clinical-trials-matching-how-the</link><guid isPermaLink="false">https://www.byte2bedside.com/p/clinical-trials-matching-how-the</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Tue, 27 Sep 2022 06:55:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ENeh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Epic is ramping up its platform play.</p><p>Having just announced the <a href="https://www.prnewswire.com/news-releases/epic-launches-life-sciences-program-unifying-clinical-research-with-care-delivery-301624287.html">launch of the Life Sciences Program</a>, its new clinical trials matching solution, the electronic health record (EHR) company is making aggressive moves to capitalize on three of their core assets to become a platform business:</p><ol><li><p>healthcare data</p></li><li><p>sticky health system relationships</p></li><li><p>the ecosystem of digitized transactions and interactions that drive patient care.</p></li></ol><p>Epic describes the following key features (quoted from the announcement)</p><ul><li><p><em>&#8220;<strong>Matching participating providers with clinical trial opportunities</strong>&nbsp;suited to the makeup of their patient populations.</em></p></li><li><p><em><strong>Sending participating providers purpose-built&nbsp;<a href="https://c212.net/c/link/?t=0&amp;l=en&amp;o=3647409-1&amp;h=1435610996&amp;u=https%3A%2F%2Fcosmos.epic.com%2F&amp;a=Cosmos">Cosmos</a>&nbsp;searches</strong>&nbsp;to help them validate whether a trial is right for them without the need to develop their own queries.</em></p></li><li><p><em><strong>Making clinical trials accessible</strong>&nbsp;to more provider groups by lowering the technical and staffing barriers to study activation.</em></p></li><li><p><em><strong>Increasing clinical trial efficiency</strong>&nbsp;by eliminating duplicative workflows and connecting researchers, care teams, patients, and sponsors through a single system.</em></p></li><li><p><em><strong>Supporting clinicians with point-of-care insights</strong>&nbsp;into when their patients might qualify for a clinical trial and applying predictive models to assist with the timing of therapy administration.&#8221;</em></p></li></ul><p>I wrote in <a href="https://www.byte2bedside.com/2021/12/how-electronic-health-record-systems-are-becoming-healthcare-platforms/">a post last year</a> about how EHR systems are beginning to look like platform businesses because of three core characteristics:</p><ol><li><p><em>A high percentage of <strong>mission critical processes </strong>in healthcare delivery involve the EHR, making it an incredibly sticky product.</em></p></li><li><p><em>EHRs <strong>accumulate rich data</strong> with each additional interaction occurring on its system.</em></p></li><li><p><em>&#8220;EHR integration&#8221; is a favorite word among healthcare technology vendors. What this means is if <strong>the EHR, rather than being a barrier, becomes a facilitator for this integration</strong> (and not just with the healthcare provider, but also between vendors), it could unlock a lot of value.</em></p></li></ol><p>Epic&#8217;s Life Sciences Program is an example of how an EHR system can stitch together these components into a powerful product if executed well. For example, Epic can aggregate data from documented clinical encounters (sourced from <strong>mission critical care delivery processes</strong>) and create large cohorts across health systems (<strong>data aggregation</strong>). Pharmaceutical companies running clinical trials can partner with Epic to access these cohorts and cross reference with their trial eligibility criteria to look for potential trial participants. A provider using Epic can also proactively look up potential clinical trials for their patient at the point of care while using Epic during the clinical encounter (a <strong>mission critical process</strong>). The addition of the patient (and associated clinical documentation) also adds to Epic&#8217;s database to facilitate future clinical trial matching opportunities (<strong>data accumulation</strong>). All of this <strong>facilitates interactions</strong> between the patient, provider, and pharmaceutical company recruiting for a clinical trial that otherwise may not have occurred.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ENeh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ENeh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 424w, https://substackcdn.com/image/fetch/$s_!ENeh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 848w, https://substackcdn.com/image/fetch/$s_!ENeh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 1272w, https://substackcdn.com/image/fetch/$s_!ENeh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ENeh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png" width="580" height="319" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:319,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!ENeh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 424w, https://substackcdn.com/image/fetch/$s_!ENeh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 848w, https://substackcdn.com/image/fetch/$s_!ENeh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 1272w, https://substackcdn.com/image/fetch/$s_!ENeh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b635e71-1ff7-40bd-a94f-f97b5f1af5b1_580x319.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><strong>A framework for how EHRs can provide value via a platform model. (Source: Byte to Bedside)</strong></figcaption></figure></div><p>Many in the industry emphasize Epic&#8217;s advantage in aggregating healthcare data (some have predicted that Epic may eventually become just a healthcare database company), but often underestimate the importance of sticky health system relationships and mission critical healthcare transactions and interactions that Epic owns. A successful platform needs to not only facilitate data exchange, but enable value added transactions between different entities that would otherwise not occur.</p><p>Epic&#8217;s announcement includes the following quote from Alan Hutchinson, Vice President at Epic: &#8220;The Life Sciences program is designed to <strong>create a seamless connection </strong>between participant patients, healthcare providers, and research sponsors through the use of a single system. Unifying clinical research with care delivery and <strong>building a direct connection </strong>to study sponsors will help accelerate the development of new therapies by making studies more efficient, more accessible, and more effective.&#8221;</p><p>Sounds pretty platform-y to me.</p>]]></content:encoded></item><item><title><![CDATA[AI and physicians as “thought partners”: a use case in endoscopic colorectal cancer screening]]></title><description><![CDATA[&#8220;Integrating AI into clinical practice&#8221; has been the holy grail in the field of AI in healthcare.]]></description><link>https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners</link><guid isPermaLink="false">https://www.byte2bedside.com/p/ai-and-physicians-as-thought-partners</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Sat, 20 Aug 2022 06:51:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!alnM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&#8220;Integrating AI into clinical practice&#8221; has been the holy grail in the field of AI in healthcare. With ML models becoming increasingly ubiquitous and accurate, solving the &#8220;integration&#8221; problem would substantially move the field forward towards real impact on patient care. But what does &#8220;integration&#8221; truly entail?</p><p>I discuss in this post an example of<a href="https://www.nature.com/articles/s41746-022-00633-6#ref-CR13"> a recently published AI system for colonic polyp classification </a>that approaches &#8220;integration&#8221; in what I believe to be a clever way: not just into physician workflow tasks, but also into how the physician thinks. I refer to this concept as &#8220;cognitive integration.&#8221;</p><h2><strong>The problem: reduce errors in routine, but high stakes procedures</strong></h2><p>Colorectal cancer screening using optical colonoscopy (OC) is one of the most effective and highest yield preventative clinical interventions in modern medicine. Unlike most other screening tests, OC can also be therapeutic; if a precancerous lesion such as polyp is identified, it can also be removed during the same session. Therefore, it is critical for polyps to be accurately detected and classified by the physician during a colonoscopy procedure.</p><p>For a well trained gastroenterologist, performing a screening colonoscopy is a relatively routine procedure, and some may even think of it as a rote task. However, the stakes are still very high, as a missed polyp can turn into metastatic colon cancer. This combination of high stakes and routine repetition can be a dangerous pair in medicine. Human error and cognitive biases will afflict even the most seasoned physicians, and the relatively routine and repetitive nature of the task may actually amplify some of these biases and potential for error.</p><h2><strong>An AI system that leverages Bayesian thinking to integrate into physician workflow</strong></h2><p>Current AI capabilities are still far from being able to completely automate screening colonoscopies, which can often be complex and require human expertise. Rather than attempting full automation, <a href="https://www.nature.com/articles/s41746-022-00633-6#ref-CR13">Biffi et al in their article in NPJ Digital Medicine</a> describe an AI enabled computer aided detection (CAD) system that functions as a &#8220;thought partner&#8221; with a human gastroenterologist to reduce errors in polyp detection and resection. The system, which has two components, functions in the following steps (Figure 1):</p><ol><li><p>A continuous video stream is fed into the CAD system as the endoscopist performs the colonoscopy</p></li><li><p>The first component of the CAD system initially marks potential areas of interest in the colonic mucosa where there could be a polyp to guide the endoscopist</p></li><li><p>The endoscopist navigates the scope towards areas of interest (does not necessarily have to be where the CAD system marked) and exposes the colonic musoca to better visualize the polyp</p></li><li><p>The second component of the CAD system then automatically activates when a polyp is consistently brought into frame by the endoscopist</p></li><li><p>The CAD system then generates in real time a classification for the polyp in the frame as adenomatous (possible cancer) versus benign, and displays the results of the classification on the screen next to the framed polyp</p></li><li><p>The CAD system then disengages if the endoscopic resects the polyp or navigates away from it</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!alnM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!alnM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 424w, https://substackcdn.com/image/fetch/$s_!alnM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 848w, https://substackcdn.com/image/fetch/$s_!alnM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 1272w, https://substackcdn.com/image/fetch/$s_!alnM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!alnM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png" width="580" height="394" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:394,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!alnM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 424w, https://substackcdn.com/image/fetch/$s_!alnM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 848w, https://substackcdn.com/image/fetch/$s_!alnM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 1272w, https://substackcdn.com/image/fetch/$s_!alnM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff75085ba-1c7f-4aca-b872-f7104291d39a_580x394.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><strong>Figure 1: The AI enabled computer aided detection (CAD) system for polyp detection and classification. Source: <a href="http://biffi,%20c.,%20salvagnini,%20p.,%20dinh,%20n.n.%20et%20al.%20a%20novel%20ai%20device%20for%20real-time%20optical%20characterization%20of%20colorectal%20polyps.%20npj%20digit.%20med.%205,%2084%20(2022)./">Biffi, C et al. </a></strong><em><strong><a href="http://biffi,%20c.,%20salvagnini,%20p.,%20dinh,%20n.n.%20et%20al.%20a%20novel%20ai%20device%20for%20real-time%20optical%20characterization%20of%20colorectal%20polyps.%20npj%20digit.%20med.%205,%2084%20(2022)./">npj Digit. Med.</a></strong></em><strong><a href="http://biffi,%20c.,%20salvagnini,%20p.,%20dinh,%20n.n.%20et%20al.%20a%20novel%20ai%20device%20for%20real-time%20optical%20characterization%20of%20colorectal%20polyps.%20npj%20digit.%20med.%205,%2084%20(2022)./">&nbsp;5, 84 (2022)</a>.</strong></figcaption></figure></div><p>The AI system includes two sets of machine learning (ML) models: the first is a convolutional neural network (CNN) that classifies each detected polyp within a single video frame as &#8220;adenoma&#8221; or &#8220;non-adenoma&#8221; and the second is a CNN that generates an image quality score. A key milestone this AI system achieves is the ability to track a polyp across multiple frames in a colonoscopy video and produce a temporally weighted decision (frames that are in focus for longer periods of time are given a higher weight), which is important for real world usability, as images of polyps dynamically change as the endoscopist navigates the scope.</p><p>The authors employ a clever strategy of pairing the complementary capabilities and limitations of the AI system and the human physician. The system leverages a Bayesian-guided approach where the AI system, rather than generate predictions independent of any insights that come from the physician and then shows the results to the physician (which is how most AI systems in healthcare operate), actually incorporates prior physician insights into its predictions.</p><p>In Bayesian statistics, probabilities are sequentially &#8220;updated&#8221; with information (priors) to calculate new probabilities that better reflect reality. This is best highlighted in step #4: the CAD system leverages the <em>prior probability</em> generated by the human endoscopist when they navigate to and focus on a suspected lesion, and then deploys the AI classification model on that chosen polyp to generate a<em> conditional probability</em> of whether the lesion is adenomatous or benign. In principle, this conditional probability should more accurately reflect whether a polyp is truly adenomatous than if the AI system naively classified all lesions visualized in the colonoscopy, because it is conditional on the endoscopist focusing on the polyp in the first place, which incorporates their expertise and human instincts.</p><p>The study included 513 prospectively acquired polyps from 165 human subjects that were examined histologically to generate ground truth labels. In their primary analysis, the CAD system was non-inferior to expert endoscopists in accurately detecting adenomatous lesions (OR 1.211 [0.766-1.915]) and superior to the accuracy of non-expert endoscopists (OR 1.875 [1.191-2.052]).</p><h2><strong>&#8220;Cognitive integration&#8221; of AI into physician workflows</strong></h2><p>Traditional workflow integration strategies focus on deploying technologies based on how they can streamline physician tasks. However, this task oriented approach misses a critical source of value for AI solutions, which is the integration of AI into how physicians think (and vice versa). This &#8220;cognitive integration&#8221; will be increasingly relevant as AI systems become more intelligent. The use case described above is a good example of how AI and physicians can team up as &#8220;thought partners&#8221; to enhance each other&#8217;s performance. The CAD system &#8220;interprets&#8221; the endoscopists&#8217; focus on a lesion as a desire to learn more, and then deploys its AI capabilities on top of insights that the endoscopist already surfaced, which in principle increases its classification accuracy since it incorporates the priors generated by the endoscopist. The endoscopist then uses the AI generated prediction to further hone in on their assessment of the polyp, which also increases their diagnostic accuracy.</p><p>When implemented well, this form of cognitive integration between AI and human experts could in theory synergistically enhance the performance of both entities in order to a produce a superior shared result. However, if designed poorly, it could very well amplify biases and lead to poorer performance and other downstream consequences.</p><p>The takeaway: when designing AI systems for physicians, consider a cognitive integration strategy that allow the AI and the physician to &#8220;make each other smarter.&#8221;</p><h2><strong>References</strong></h2><p><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">Biffi, C., Salvagnini, P., Dinh, N.N.&nbsp;</a><em><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">et al.</a></em><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">&nbsp;A novel AI device for real-time optical characterization of colorectal polyps.&nbsp;</a><em><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">npj Digit. Med.</a></em><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">&nbsp;</a><strong><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">5</a></strong><a href="https://www.nature.com/articles/s41746-022-00633-6#citeas">, 84 (2022). https://doi.org/10.1038/s41746-022-00633-6</a></p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.byte2bedside.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI enabling new workflows to expand access to care (part 1)]]></title><description><![CDATA[Originally published on 4/25/2022 on byte2bedside.com Examples of AI in healthcare typically involve machine learning (ML) models that can perform the same task a human expert performs manually (eg. diagnosing pneumonia on a chest X-ray). The hope is that these systems can be useful for augmenting (or even replacing) existing workflows for clinicians, but more or less preserve the same structure and approach towards diagnosing and treating disease.]]></description><link>https://www.byte2bedside.com/p/ai-enabling-new-workflows-to-expand</link><guid isPermaLink="false">https://www.byte2bedside.com/p/ai-enabling-new-workflows-to-expand</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Mon, 25 Apr 2022 17:23:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5b6649df-3aaf-4383-a259-b32143756ae2_580x189.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Originally published on 4/25/2022 on byte2bedside.com </em></p><p>Examples of AI in healthcare typically involve machine learning (ML) models that can perform the same task a human expert performs manually (eg. diagnosing pneumonia on a chest X-ray). The hope is that these systems can be useful for augmenting (or even replacing) existing workflows for clinicians, but more or less preserve the same structure and approach towards diagnosing and treating disease.</p><p>I am intrigued by the next generation of how AI can impact healthcare: not by just performing a task that a human already is doing, but enabling new tasks that previously did not exist and fundamentally changing the structure of how healthcare can be delivered to expand access to high quality care.</p><p>Over the next few posts, I will be highlighting several examples of ML in published literature that I believe represents this potential opportunity. This post will describe an example of a use case in medical imaging.</p><h2><strong>First generation of AI in imaging: augment existing diagnostic tasks</strong></h2><p>Early examples of ML image classifiers in healthcare demonstrated promise in performing tasks that could potentially augment or even replace the workflows of human expert clinicians, such as identifying <a href="https://stanfordmlgroup.github.io/projects/chexnet/">pneumonia on chest X-rays</a>, <a href="https://www.nature.com/articles/s41598-021-95249-3">pulmonary embolism on CT scans</a>, or <a href="https://www.nature.com/articles/nature21056">skin cancer from photos of skin lesions</a>. The idea is that an AI system could help radiologists and dermatologists perform these existing diagnostic tasks more quickly and efficiently. There are already examples of these capabilities being operationalized into clinical workflows: <a href="https://www.viz.ai/">Viz.ai </a>developed ML classifiers that detect ischemic stroke and large vessel occlusions on head CT images and has even been successful at attaining a <a href="https://jnis.bmj.com/content/13/5/406">Medicare reimbursement code </a>for their technology when used on hospitalized patients.</p><h2><strong>Next generation: AI enabling new diagnostic tasks</strong></h2><p>The earlier examples of AI all performed tasks that already existed in clinical medicine (eg. radiologists already routinely diagnose pulmonary emboli and strokes on CT scans, the AI is trained to perform the same task). Yet, we may only be scratching the surface in understanding the complex, hidden relationships between the information from imaging studies and an individual&#8217;s physiological and disease state.</p><p>The disease phenotypes that we traditionally diagnose from these studies are simply from the accumulated experiences of expert physicians who learned to interpret these images under existing mental models and knowledge structures (as well as the limitations of the human eye). What if we allow an AI system to take a second look, which has the benefit of accessing and interpreting every pixel, voxel, and frame of an imaging study, as well the ability to learn without these pre-existing mental models?</p><p>The following study illustrates one example of attempting to apply AI to enable a new diagnostic task on a CT scan: <a href="https://www.nature.com/articles/s41746-021-00460-1">Automated coronary calcium scoring using deep learning with multicenter external validation</a>.</p><h2><strong>A new way of detecting coronary artery calcification on routine non gated chest CTs</strong></h2><p>Measuring coronary artery calcium (CAC) can be a useful non-invasive method for detecting coronary artery disease and help risk stratify for treatment decisions. Calcium is radio-opaque, so can be visualized on CT scans. However, because the heart is a dynamically beating organ, the coronary arteries and surrounding muscle and tissue are not stationary, which makes it difficult for radiologists to precisely quantify CAC on regular CTs. To solve this problem, &#8220;cardiac gating&#8221; is added to the CT image acquisition process, which means that the CT scan is triggered during a specific portion of the cardiac cycle (ie. a point in time during the heartbeat) using attached electrocardiogram leads. Medications to slow down the patient&#8217;s heart rate are also typically given during the scanning process.</p><p>This specialized CT scan, known as a &#8220;gated coronary CT,&#8221; is currently required for patients who need to measure their CAC, but are much less widely available (and more expensive) compared to regular CT scans, which do capture coronary calcium but cannot be formally interpreted by radiologists to measure CAC. The result is a significant underutilization of CAC screening to detect latent coronary artery disease, representing a missed opportunity in the early detection and prevention of cardiovascular disease.</p><h4><strong>What if AI can calculate CAC from non gated routine CT scans?</strong></h4><p>This is what the research team that published this <a href="https://www.nature.com/articles/s41746-021-00460-1">paper</a> demonstrated with some early data. They trained a deep learning model that automatically calculates the CAC score from regular non gated chest CTs using data from Stanford Health Care, a tertiary academic medical center, and the Multi-Ethnic Study of Atherosclerosis (MESA) study, a multi-center prospective cohort. In a validation set of 42 patients from Stanford (which is also where the data for the training set came from) who also received gated coronary CTs, there was almost perfect agreement between the AI generated CAC score from regular CTs and the radiologist interpreted score from the gated coronary CTs. Agreement was moderate in a validation set of 46 patients from the MESA study. External validation at other sites also showed moderate to substantial agreement. The models also performed well in a binary classification task of detecting CAC greater or equal to 100.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3KDa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3KDa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 424w, https://substackcdn.com/image/fetch/$s_!3KDa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 848w, https://substackcdn.com/image/fetch/$s_!3KDa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 1272w, https://substackcdn.com/image/fetch/$s_!3KDa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3KDa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png" width="580" height="189" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:189,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!3KDa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 424w, https://substackcdn.com/image/fetch/$s_!3KDa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 848w, https://substackcdn.com/image/fetch/$s_!3KDa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 1272w, https://substackcdn.com/image/fetch/$s_!3KDa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870100a1-2b44-45ee-9fa5-672fa2bd2d98_580x189.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><strong>Example of AI detection of CAC on routine chest CT. Left: reference image; Middle: ML model detection of CAC; Right: manual detection of CAC. Source: <a href="https://www.nature.com/articles/s41746-021-00460-1">Eng et al. NPJ Digital Medicine 2021.</a></strong></figcaption></figure></div><p>The authors also reported the development and validation of a ML model that automatically calculates CAC from gated coronary CTs with high concordance with radiologist interpretations, which is exciting work and represents the opportunity to augment an existing diagnostic task.</p><p>I am, however, particularly excited about the results from the non gated CT model, which indicate the potential for AI to <strong>enable new diagnostic tasks and workflows, rather than just augment existing ones.</strong> Under the current standard of care, access to CAC screening is fairly limited, especially among more vulnerable populations (who are also often at higher risk of cardiovascular disease). Relying on only gated coronary CTs also misses all the opportunities for CAC detection from regular non coronary CT scans that people receive for other reasons. An AI system that could capture diagnostic and screening information from these regular CT scans could fundamentally change how we screen for CAC: <strong>by shifting from a centralized system gated (no pun intended) by a limited number of specialized centers to a more distributed system that can spread much more widely in the community.</strong></p><p>There is much more work to do beyond this proof of concept paper. Models have to be developed on larger, more representative datasets and assessed for generalizability. Implementation feasibility, ranging from technical and clinical integration, will need to be explored and tested. It may even be possible that it turns out the ML is not as good as expected when stress tested under other conditions. However, I believe this foundational concept of AI extracting insights from previously untapped diagnostic tests to enable new workflows that broadly expand access to care is will drive important future work in AI.</p><h2><strong>Reference</strong></h2><p>Eng, D. et al. (2021). Automated coronary calcium scoring using deep learning with multicenter external validation. <em>Npj Digital Medicine</em>, <em>4</em>(1). https://doi.org/10.1038/s41746-021-00460-1</p>]]></content:encoded></item><item><title><![CDATA[How electronic health record systems are becoming healthcare platforms]]></title><description><![CDATA[Two seemingly unrelated headlines caught my attention recently: one is the widely discussed Oracle acquisition of Cerner, and the other is an article in NPJ Digital Medicine describing key challenges and opportunities for leveraging the electronic health record (EHR) to enhance the]]></description><link>https://www.byte2bedside.com/p/how-electronic-health-record-systems</link><guid isPermaLink="false">https://www.byte2bedside.com/p/how-electronic-health-record-systems</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Tue, 21 Dec 2021 18:13:00 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" width="5184" height="3456" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:3456,&quot;width&quot;:5184,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;person sitting while using laptop computer and green stethoscope near&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="person sitting while using laptop computer and green stethoscope near" title="person sitting while using laptop computer and green stethoscope near" srcset="https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1576091160550-2173dba999ef?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHwyfHxtZWRpY2FsJTIwcmVjb3JkfGVufDB8fHx8MTcwNDQ0MDQ1Nnww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@nci">National Cancer Institute</a> on <a href="https://unsplash.com">Unsplash</a></figcaption></figure></div><p>Two seemingly unrelated headlines caught my attention recently: one is the widely discussed <a href="https://www.cnbc.com/2021/12/20/oracle-to-buy-medical-records-company-cerner.html">Oracle acquisition of Cerner</a>, and the other is an <a href="https://www.nature.com/articles/s41746-021-00542-0">article</a> in <em>NPJ Digital Medicine </em>describing key challenges and opportunities for leveraging the electronic health record (EHR) to enhance the <a href="https://www.fda.gov/safety/fdas-sentinel-initiative">Sentinel Initiative</a>, the FDA&#8217;s medical product safety surveillance program. To me, these are both chapters of the larger story of how the EHRs are transforming from software for medical records into a platforms for healthcare.</p><h2><strong>Platforms are both technical products and business models</strong></h2><p>As a technical product, a platform can be thought of as a system of hardware and software on which different applications and processes can be deployed. Historically, platforms involved on-premises architecture that supported the myriad of enterprise applications needed for a business. Much of this has since migrated to the cloud as platform-as-a-service (PaaS) solutions. In healthcare, large cloud players such as Microsoft Azure, Amazon AWS, and now increasingly Google Cloud offer PaaS products in addition to their infrastructure services that allow healthcare customers to deploy applications and analytics on its technology stack. Oracle has been a PaaS leader in many industries, but not healthcare.</p><p>The broader way of thinking about platforms is as a type of business model. Platform businesses enable exchanges between different entities by facilitating connections and transactions. Value creation therefore does not rely on a linear supply chain (build, ship, and sell a product to the customer), but from the cumulative connections and interactions that occur on the platform. Revenue is generated per interaction rather than per product unit, which is a powerful driver for economics of scale. Successful platform businesses can create network effects, meaning each additional entity joining the platform creates more value for the rest of the platform. For example, each additional driver and passenger who joins the Uber platform increases the total number of available drivers and customers, which then increases the chance of a value generating match.</p><h2><strong>EHR systems are beginning to look like platforms</strong></h2><p>The EHR is traditionally thought of as a piece of software for documenting, organizing, and accessing patient health records. The interaction is mainly limited between the user (clinician) and the EHR software. However, the scope of transactions that EHRs support have rapidly expanded to other mission critical services, such as <a href="https://www.cerner.com/solutions/revenue-cycle-management">revenue cycle management</a>, clinical research, regulatory reporting, <a href="https://www.orpca.org/Epic%20Secure%20Chat%203-26-2020.pdf">clinical communication</a>, and <a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6625217/">patient engagement</a>. These expanded services also come with an exponential increase in both the volume and diversity of interactions occurring through the EHR. Clinicians are no longer the only users of EHRs; patients, researchers, government regulatory agencies, non clinical healthcare staff, and medical device and digital health vendors are all now part of a growing list.</p><p>However, EHRs are not platforms yet. EHR companies still typically follow a linear SaaS business model, deriving most of their revenue from enterprise contracts with healthcare provide customers. While there are indeed a large number of interactions with the EHR system by different types of entities, <em>these interactions typically occur directly with the EHR system, rather than between the entities themselves.</em> Plenty of healthcare software vendors try to &#8220;integrate with the EHR&#8221; to reach users, but the EHR is often seen as a barrier to get through, rather than a facilitator for the interaction, which is what a true platform is supposed to offer. EHR systems are now starting to integrate different services and applications, such as clinical care, patient engagement, and revenue cycle in a &#8220;platform-like&#8221; manner; that is, they provide the technical back-end for these applications to interact with each other, but they ultimately still occur within the walled gardens of individual enterprise customers. I am interested in seeing what comes out of the <a href="https://epicresearch.org/">Epic Research Network</a> and the <a href="https://ehrintelligence.com/news/epics-faulkner-has-high-hopes-for-forthcoming-cosmos-technology">Cosmos </a>program, which is working towards aggregating patient data from different health system customers to support research and discovery, which can potentially facilitate value generating interactions between health systems, vendors, and researchers.</p><p>Nevertheless, I believe EHRs are powerfully positioned to become platform businesses because:</p><ol><li><p>A high percentage of mission critical processes in healthcare delivery involve the EHR, making it an incredibly sticky product.</p></li><li><p>EHRs accumulate rich data with each additional interaction occurring on its system.</p></li><li><p>&#8220;EHR integration&#8221; is a favorite word among healthcare technology vendors. What this means is if the EHR, rather than being a barrier, becomes a facilitator for this integration (and not just with the healthcare provider, but also between vendors), it could unlock a lot of value.</p></li></ol><h2><strong>What does this have to with Oracle&#8217;s Cerner acquisition and the FDA Sentinel Initiative?</strong></h2><p>My view is that they are examples of growing supply and demand for EHRs to become platforms.</p><p>I can only speculate, but one version of the strategy behind Oracle&#8217;s acquisition of Cerner may be to create a true healthcare technology platform that facilitates interactions between entities within the healthcare ecosystem, including patients, providers, healthcare provider organizations, payers, third party software vendors, and government organizations. Oracle has the cloud, middleware, and analytics stack &#8212; all building blocks of a platform solution &#8212; and Cerner has the data and reach into healthcare services to cultivate a customer base and ecosystem.</p><p>On a parallel note, this <a href="https://www.nature.com/articles/s41746-021-00542-0">article</a> lays out a set of barriers to scaling the FDA Sentinel Initiative for medical product surveillance. Currently, there does not exist the infrastructure and analytics stack needed to leverage data from multiple EHR systems, payers, and medical devices fragmented across the industry in order support a surveillance program at the necessary scale. A healthcare platform that weaves together these entities can make it happen.</p><p>There are elements of healthcare that will present barriers to allowing platform businesses to grow in the same way as they have in other industries. However, I am excited to see the growth of both supply and demand, which makes me optimistic that the secular trend is there.</p>]]></content:encoded></item><item><title><![CDATA[Remote patient monitoring: a technology or a care model?]]></title><description><![CDATA[I recently came across a nice publication describing an example of a real world implementation of remote patient monitoring (RPM) for patients with Covid-19 at the Mayo Clinic.]]></description><link>https://www.byte2bedside.com/p/remote-patient-monitoring-a-technology</link><guid isPermaLink="false">https://www.byte2bedside.com/p/remote-patient-monitoring-a-technology</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Thu, 19 Aug 2021 17:08:00 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" width="4720" height="3146" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:3146,&quot;width&quot;:4720,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;white and black digital device&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="white and black digital device" title="white and black digital device" srcset="https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1615486511484-92e172cc4fe0?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxwdWxzZSUyMG94aW1ldGVyfGVufDB8fHx8MTcwNDQ0MDE5MXww&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@mockupgraphics">Mockup Graphics</a> on <a href="https://unsplash.com">Unsplash</a></figcaption></figure></div><p>I recently came across a nice publication describing an example of a real world implementation of remote patient monitoring (RPM) for patients with Covid-19 at the Mayo Clinic. There are a few aspects of this paper that I particularly liked that I will discuss, but it also led me to think about a more fundamental question of whether provider organizations should approach RPM as a set of technologies, or a care model?</p><p>RPM refers to digital technologies that allow for the capture of health data from patients remotely (eg. at home) and the transmission of that data to providers in order to guide the diagnosis and management of health conditions. At a high level, the technology is broken down into the following components: hardware that detect the health data (eg. glucometer, blood pressure cuff, scale, pulse oximeter, ECG tracing), software that organizes and at times analyzes the health data, and the capability to transmit the health data and insights to the provider, often via an online portal or integration with an existing system such as the electronic health record (EHR). The value proposition for RPM has mostly centered around disease prevention through early detection and decreasing healthcare utilization, although the Covid-19 pandemic has surfaced the additional benefit of decreasing unnecessary in-person patient/provider interactions, which has greatly accelerated the interest and adoption of RPM.</p><p>A frequently noted challenge and barrier to adoption for RPM companies is &#8220;integration&#8221; into care delivery. The barrier is often seen as technical, with a common one being integration of data from the RPM system into the provider&#8217;s EHR. This problem, in fact, is often more easily solved than expected, as there are a variety of methods for integrating external data into EHRs. I believe that a more important question is how RPM integrates with the care model it is meant to improve.</p><p>Coffrey et al. describes in the latest issue of npj Digital Medicine a RPM program for managing patients with Covid-19 at the Mayo Clinic. The technology component of the program is fairly simple: a set of equipment provided by Mayo to patients for monitoring vitals signs at home (thermometer, blood pressure cuff, pulse oximeter) that deliver patient reported data to providers via either manual entry or passive collection from bluetooth connected devices.</p><p>What caught my attention about this paper was not its RPM technology, but how they presented the RPM implementation as a comprehensive care model. The components of the RPM care models are summarized in this table. The care models were centered on a specific patient population (Covid-19) and detailed in a granular fashion the value added steps needed to deliver care for those patients. The technological component was only a piece of an overall tech enabled process that includes care pathways and care team engagement methods specific to this care model.</p><p>This way of presenting a RPM implementation is important because it highlights that the implementation is actually that of a tech-enabled care model rather than just a technology. The endpoints they propose to evaluate the RPM program were also appropriately chosen to capture the processes of this care model.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ycyF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ycyF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 424w, https://substackcdn.com/image/fetch/$s_!ycyF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 848w, https://substackcdn.com/image/fetch/$s_!ycyF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 1272w, https://substackcdn.com/image/fetch/$s_!ycyF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ycyF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp" width="768" height="214" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:214,&quot;width&quot;:768,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:73010,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ycyF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 424w, https://substackcdn.com/image/fetch/$s_!ycyF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 848w, https://substackcdn.com/image/fetch/$s_!ycyF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 1272w, https://substackcdn.com/image/fetch/$s_!ycyF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f759734-4331-4751-bbd7-0b5391ef8df1_768x214.webp 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><p><em>RPM evaluation endpoints (from https://www.nature.com/articles/s41746-021-00490-9/tables/4)</em></p><p>Recognizing what it is that we are actually implementing is critical for understanding the barriers and facilitators to implementation and the expected impact on care delivery. Often, what may initially appear as an implementation of a technology such as RPM is actually an implementation of a new care model. And implementation barriers, such as data integration challenges many RPM companies face with provider organization clients may appear to be technical on the surface, but in reality arise from the lack of a clear RPM enabled care model.</p><p>Getting the data in is easy, knowing how to use it is harder.</p>]]></content:encoded></item><item><title><![CDATA[“API-Fying” healthcare services]]></title><description><![CDATA[*Reposted* &#8212; originally published on 7/12/2021 on byte2bedside.com]]></description><link>https://www.byte2bedside.com/p/api-fying-healthcare-services</link><guid isPermaLink="false">https://www.byte2bedside.com/p/api-fying-healthcare-services</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Mon, 12 Jul 2021 17:04:00 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" width="6016" height="4016" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:4016,&quot;width&quot;:6016,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;white monitor&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="white monitor" title="white monitor" srcset="https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1559137781-875af01c14bc?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw2fHxyZWNlcHRpb25pc3R8ZW58MHx8fHwxNzA0NDM5ODY2fDA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@proxyclick">Proxyclick Visitor Management System</a> on <a href="https://unsplash.com">Unsplash</a></figcaption></figure></div><p><em>*Reposted* &#8212; originally published on 7/12/2021 on byte2bedside.com</em> </p><p>One of the revolutionary advances in software and internet services is the emergence of application programming interfaces (APIs) that allow for processes to occur between software applications in a consistent manner that can be easily scaled.</p><p>APIs make it possible for consumers to conduct and receive all sorts of services such as online banking and ordering merchandise with a consistent experience that is high quality and low cost. At the foundation of these services lies a series of data exchanges and processes carried out between different software applications that must be consistently executed and easily scaled. APIs provide the interaction points between software applications to allow this to happen.</p><p>Imagine if next time you order an item from Amazon, after clicking the &#8220;order&#8221; button, you are instead instructed to wait for a phone call from the merchant to confirm your order. The phone call comes in three days (although there is a 20% chance that it may never come), which ends up going to your voicemail. You check your voicemail and call the merchant back, only to be greeted by someone who was not aware that you had even placed the order, and asks for you to go through the order information again. You describe to the representative the item that you want to order, but it turns out that the item originally listed on the website is no longer available, but there is an alternative brand that can ship in three weeks, but they are unable to tell you anything more about the product, or even how much it costs. And it turns out that if you want to buy a different item, you will have to go through a different lengthy and complicated process that changes depending on what item you order.</p><p>The above scenario is not far from the realities of what patients have to go through to receive services in healthcare, such as scheduling an appointment, getting a referral, or getting medical records. Similar to software processes, healthcare services also involve interaction points between people, software, and organizations that require multiple exchanges of information that results in some action (eg. an appointment getting scheduled). There are of course some key differences between getting a referral for an orthopedic surgeon for a knee replacement and ordering trash bags from Amazon, but the underlying need for consistency, efficiency, and scalability applies to healthcare services just as much (if not more so) as online shopping. Countless medical errors and unnecessary waste occurs in healthcare due to the lack of these qualities.</p><p>When thinking about the potential for informatics in healthcare, I am particularly excited about the opportunities to &#8220;API-fy&#8221; healthcare services; that is, to leverage data, software, IT infrastructure, and processes to create &#8220;APIs&#8221; (both literally and figuratively) that can make healthcare interaction points more consistent, efficient, and scalable. What if a patient with a worrisome skin lesion could secure a referral with an available dermatologist within minutes, rather than having to play a months long game of back and forth phone tag with insurance companies and schedulers that may result in delays in care that could turn out to be deadly? There are many interactions and processes in the back end that would need to take place for this to occur, many of which could be addressed by actual software-software APIs, and others by human and organizational driven processes that can still follow the principles of software APIs to deliver the needed consistency, efficiency, and scalability.</p><p>Just as how software APIs have evolved from just code into products, I envision these healthcare &#8220;APIs&#8221; to also emerge from disparate processes into packaged products and services that involve software, data, and organizational mechanisms.</p>]]></content:encoded></item><item><title><![CDATA[What can the history of cloud computing teach us about the future of AI?]]></title><description><![CDATA[*Reposted* &#8212; originally published on 4/21/2021 on byte2bedside.com There is much excitement about AI in healthcare, but we still don&#8217;t quite understand how it can produce impact at a meaningful scale. Discussions about AI often jump to examples of machine learning (ML) models that perform eye catching tasks such as interpreting X-rays and diagnosing disease. However, although there is no shortage of published papers describing highly accurate ML models that predict all sorts of clinical outcomes, there is a dearth of actual AI products that are widely adopted and deliver value.]]></description><link>https://www.byte2bedside.com/p/what-can-the-history-of-cloud-computing</link><guid isPermaLink="false">https://www.byte2bedside.com/p/what-can-the-history-of-cloud-computing</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Wed, 21 Apr 2021 17:03:00 GMT</pubDate><enclosure url="https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw"><img src="https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080" width="5120" height="2880" data-attrs="{&quot;src&quot;:&quot;https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2880,&quot;width&quot;:5120,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;a computer circuit board with a brain on it&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="a computer circuit board with a brain on it" title="a computer circuit board with a brain on it" srcset="https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 424w, https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 848w, https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1272w, https://images.unsplash.com/photo-1677442135703-1787eea5ce01?crop=entropy&amp;cs=tinysrgb&amp;fit=max&amp;fm=jpg&amp;ixid=M3wzMDAzMzh8MHwxfHNlYXJjaHw1fHxhaXxlbnwwfHx8fDE3MDQ0Mzk2MDV8MA&amp;ixlib=rb-4.0.3&amp;q=80&amp;w=1080 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Photo by <a href="https://unsplash.com/@steve_j">Steve Johnson</a> on <a href="https://unsplash.com">Unsplash</a></figcaption></figure></div><p>There is much excitement about AI in healthcare, but we still don&#8217;t quite understand how it can produce impact at a meaningful scale.</p><p>Discussions about AI often jump to examples of machine learning (ML) models that perform eye catching tasks such as interpreting X-rays and diagnosing disease. However, although there is no shortage of published papers describing highly accurate ML models that predict all sorts of clinical outcomes, there is a dearth of actual AI products that are widely adopted and deliver value.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Byte to Bedside! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Common explanations for this gap between research and real world application include low risk tolerance in healthcare, complexity and messiness of healthcare data that limit the ability to train ML models, lack of IT infrastructure for deploying ML, and unclear ways to integrate ML models into clinical workflows. In essence, the problem, as described, centers around limitations in being able to train and deploy ML models and understanding what to do with them.</p><p>These explanations tell part of the story, but I think there is a deeper gulf of understanding that precludes the use of AI in healthcare in a way that delivers meaningful impact. To explain further, I find it useful to draw from the history of cloud computing.</p><p>The concept of the cloud was born out of a technology known as virtualization, which for a non-IT professional like myself, can be simplified as the ability to duplicate the functionality of a physical computer into multiple entities beyond the original hardware. So if I have a computer with 500GB of hard drive space, 2.3GHz CPU, and 32GB RAM that contains a set of files and software, virtualization would allow me to distribute the storage space, processing speed, and memory across multiple other &#8220;guest&#8221; computers via virtual machines that can run the same software and access the same files located on the original host physical computer. This functionality, which was first developed decades ago and iterated over time, is foundational to the ability to flexibly distribute and scale IT resources and construct what is commonly referred to as the &#8220;cloud.&#8221;</p><p>Cloud computing has now made its way well beyond concept into being a core component of the modern economy. To appreciate its full impact, we need to look beyond just the traditional cloud service providers (eg. AWS, Azure, GCP) to the vast array of businesses and products <strong>that are made possible </strong>by cloud architectures such as infrastructure as a service (IAAS), platform as a service (PAAS), and software as a service (SAAS). While a company like Netflix may incorporate cloud computing into its tech stack to build its products, <strong>its business value is generated from what is enabled by the cloud, rather than directly from the cloud itself</strong>. Netflix did not start its business sitting on a bunch of virtual machines and wondering what to do with them; rather, they identified a set of unaddressed needs in the market and realized that cloud computing would enable a suite a products that fulfilled those needs.</p><p>I find it useful to compare AI to cloud computing in that both ML and virtualization are significant methodological advances that can deliver new functionalities and enable products and business models that previously were not possible. But just as the early developers of virtualization software probably did not envision the creation of Netflix, it is also difficult for the AI community to fully appreciate what future products and business models can be made possible by AI. We need to look beyond the immediate applications of ML models to the higher order derivatives that may emerge from functionalities that AI enables. By re-framing the question from &#8220;what can we do with a more accurate ML model&#8221; to <strong>&#8220;what do we need to change in healthcare, and what future state products, care processes, and business models needed to deliver these changes can be made possible by AI,&#8221;</strong> we can greatly expand the opportunity space for AI in healthcare.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Byte to Bedside! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Innovation is greater than the sum of its parts]]></title><description><![CDATA[Why we probably don't yet see the true innovative potential of AI in healthcare]]></description><link>https://www.byte2bedside.com/p/innovation-is-greater-than-the-sum</link><guid isPermaLink="false">https://www.byte2bedside.com/p/innovation-is-greater-than-the-sum</guid><dc:creator><![CDATA[Ron Li]]></dc:creator><pubDate>Thu, 11 Feb 2021 00:12:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!fgqG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What is innovative about the Apple Watch ECG app? The answer to the above question may differ based on who you ask.</p><p>AI enthusiasts will point to the novel deep learning classifier that <a href="http://jama%20cardiology/">detects atrial fibrillation</a> from captured ECG waveforms, while others will speak to the sensing capabilities of the hardware or the usability of the app.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Byte to Bedside! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fgqG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fgqG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 424w, https://substackcdn.com/image/fetch/$s_!fgqG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 848w, https://substackcdn.com/image/fetch/$s_!fgqG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 1272w, https://substackcdn.com/image/fetch/$s_!fgqG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fgqG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp" width="445" height="603" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:603,&quot;width&quot;:445,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:72286,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fgqG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 424w, https://substackcdn.com/image/fetch/$s_!fgqG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 848w, https://substackcdn.com/image/fetch/$s_!fgqG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 1272w, https://substackcdn.com/image/fetch/$s_!fgqG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd20fc4a9-0b0d-4ade-8b9d-a20a4b4fc1b8_445x603.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><em>The ECG app on the Apple Watch Series 4 detects signals from your pulse that is fed into a deep learning classifier and detects the presence of atrial fibrillation. In a <a href="https://www.nejm.org/doi/full/10.1056/NEJMoa1901183">study</a> of 419,297 participants over a median duration of 117 days, the app was found to have a positive predictive value of 84% for detecting atrial fibrillation.</em></p><p><em>https://www.apple.com/healthcare/apple-watch/</em></p><h2>The iPhone and Apple Watch as illustrative examples </h2><p>When Steve Jobs first <a href="https://www.youtube.com/watch?v=wGoM_wVrwng">unveiled</a> the iPhone in 2007, he cryptically began his introduction by describing three separate products: &#8220;a widescreen iPod with touch controls, &#8220;a revolutionary mobile phone&#8221;, and &#8220;a breakthrough internet communications device.&#8221; Although the applause from the audience grew with the announcement of each separate &#8220;product,&#8221; the punchline was of course when he revealed that they are actually features of one product: the iPhone. Jobs then moves on to describe three enabling technologies that made the iPhone revolutionary at the time: the multi-touch screen, a powerful operating system and suite of software, and syncing capabilities with other products in the Apple ecosystem.</p><p>The purpose of this post is not to promote Apple products, but to illustrate a concept that I think Apple understands very well: <strong>breakthrough innovation does not come from advances in just one type of technology, but from the integration of multiple streams of technologies that creates something greater than the sum of its parts</strong><em>.</em> The same concept applies to the Apple ECG app. The AI, digital health, and hardware components are of course all advanced in their own right, but it is the <strong>integration </strong>of these capabilities into a product that allows us to think about how these individual technologies can impact health. This is not an easy task, and is part of the great filter that sits between new technologies and great products.</p><p>In the case of the Apple ECG app, integration needed to occur both in technology and design. On the technical end, the AI needed to be integrated into the software and hardware systems that ingest and derive the right data into model features at runtime, and enable the delivery of the predictions in a way that creates a viable user experience. A model prediction, however accurate it may be, must also be integrated into a user experience that drives desired behavior changes in order to be impactful on health. The Apple ECG app is innovative precisely because it is not simply just a smartwatch, an AI model, and a sensor. The integration of these three technologies created something that is greater than the sum of its parts.</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5Mom!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5Mom!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 424w, https://substackcdn.com/image/fetch/$s_!5Mom!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 848w, https://substackcdn.com/image/fetch/$s_!5Mom!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 1272w, https://substackcdn.com/image/fetch/$s_!5Mom!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5Mom!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp" width="768" height="544" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:544,&quot;width&quot;:768,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:98988,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5Mom!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 424w, https://substackcdn.com/image/fetch/$s_!5Mom!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 848w, https://substackcdn.com/image/fetch/$s_!5Mom!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 1272w, https://substackcdn.com/image/fetch/$s_!5Mom!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F470e2970-9727-46d6-8789-e9fa9f32e55a_768x544.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><em>The AI model predictions generate snippets and clinical interpretations of ECG waveforms that can be saved into the Apple Health Record, which is a digital personal health record that aggregates data from electronic health records from APIs. A notable gap here is the ability to upload data from Apple Health Records into the EHR systems used by physicians (as indicated here, a PDF has be generated and emailed or printed for the physician).</em></p><p><em>https://support.apple.com/en-us/HT208955</em></p><h2>Innovation as an emergent property </h2><p>In systems science, an emergent property is a property of an entity that cannot be derived from just the sum of its individual parts. For example, a neuron in a human brain is comprised of individual units such as a lipid membrane, organelles, DNA, and neurotransmitters, each with its set of properties and functions. However, as the neuron fires and ultimately leads to a complex emotion such as love or sadness, this &#8220;emergent&#8221; phenomenon cannot be explained simply be adding up the functions and properties of the individual units.</p><p>I like to think of innovation as an emergent property that comes from the interactions between people, processes, and technology. This is important to keep in mind in technology driven fields such as health IT, where it can be easy to fall into the trap of equating innovation with the novelty of a particular technology. Organizations that are looking to update their tech stack with the latest AI, cloud, and digital capabilities need to think about how these technologies interact with each other, and with the people and processes that they are looking to service and improve. Rather than just assessing each technology separately, innovators need to take a step back and look at big picture: what emerges when these components of the tech stack interact with each other and integrate within the ecosystem of people, technologies, and processes where the problems they aim to solve exist?</p><p>Apple and Jobs understood this concept well. Otherwise, they may as well have ended up separately developing a phone, music player, and mobile internet browser rather than the iPhone.</p><h2>Why this is relevant now with AI in healthcare </h2><p>The recent advances in AI and large language models are impressive, but represent only one of the building blocks for the types of innovation that will truly change healthcare.  These capabilities need to be integrated with the rest of the healthcare ecosystem in ways that will allow these innovations to emerge.  What they will look like can be anyone&#8217;s guess &#8212; but likely not just an AI chatbot that answers medical questions. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.byte2bedside.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Byte to Bedside! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>