Malcolm Gladwell: Tell People What It's Really Like To Be A Doctor | Forbes

Gladwell discussing the clerical side of being a physician:

You don’t train someone for all of those years of medical school and residency, particularly people who want to help others optimize their physical and psychological health, and then have them run a claims-processing operation for insurance companies.

Such a profound insight from someone who is not a doctor.

I think in the future, well-designed electronic medical records will help cut down on some of these clerical burdens. There’s absolutely no reason why the bulk of data insurance companies want can’t be automatically generated from electronic records. Unfortunately, nobody seems to have done this yet (and current EMRs are poorly designed).

If we continue along our current trajectory, two things will happen:

  • More and more doctors will take salaried positions within large practices/hospitals in order to avoid the backbreaking work of running an independent small business in addition to practicing medicine.
  • Direct primary care will become more and more popular so that neither doctors nor patients will have to deal with insurance bureaucracies.

✚ Why bad research makes it into good medical journals—a critique of the Ontario surgical checklist study

This past week, a study in the New England Journal of Medicine called into question the effectiveness of surgical checklists for preventing harm. Atul Gawande—one of the original researchers demonstrating the effectiveness of such checklists and author of a book on the subject—quickly wrote a rebuttal on the The Incidental Economist. He writes, “I wish the Ontario study were better,” and I join him in that assessment, but want to take it a step further.

Gawande first criticizes the study for being underpowered. I had a hard time swallowing this argument given they looked at over 200,000 cases from 100 hospitals. I had to do the math. A quick calculation shows that given the rates of death in their sample, they only had about 40% power [1]. Then I became curious about Gawande’s original study. They achieved better than 80% power with just over 7,500 cases. How is this possible?!?

The most important thing I keep in mind when I think about statistical significance—other than the importance of clinical significance [2]—is that not only does it depend on the sample size, but also the baseline prevalence and the magnitude of the difference you are looking for. In Gawande’s original study, the baseline prevalence of death was 1.5%. This is substantially higher than the 0.7% in the Ontario study. When your baseline prevalence approaches the extremes (i.e.—0% or 50%) you have to pump up the sample size to achieve statistical significance.

So, Gawande’s study achieved adequate power because their baseline rate was higher and the difference they found was bigger. The Ontario study would have needed a little over twice as many cases to achieve 80% power.

This raises an important question: why didn’t the Ontario study look at more cases?

The number of cases in a study is dictated by limitations in data collection. Studies are generally limited by the manpower they can afford to hire and the realistic time limitations of conducting a study. However, studies that use existing databases are usually not subject to these constraints. While creating queries to extract data is often tricky, once you have setup your extraction methodology it simply dumps the data into your study database. You can extend or contract the time period for data collection by simply changing the parameters of your query. Modern computing power means there are few limitations on the sizes of these study databases and the statistical methodologies we can employ. Simply put, the Ontario study (which relied on ‘administrative health data,’ read: ‘existing data’) easily could have doubled the number of cases in their study.

Exactly how did they define their study group? As Gawande points out in his critique, the Ontario study relied on this bizarre 3-month window before and after checklist implementation at individual hospitals. Why 3 months? Why not 6 or 12 or 18? They even write in their methods:

We conducted sensitivity analyses using different periods for comparison. [3]

They never give the results of these sensitivity analyses or provide sound justification for the choice of a 3-month period. Three months not only keeps their power low, but it fails to account for secular trends. Maybe something like influenza was particularly bad in the post-checklist period, leading to more deaths despite effective checklist use. Maybe a new surgical technique or tool was introduced, like DaVinci robots, or many new, inexperienced surgeons were hired that increased mortality. In discussing their limitations, they address this:

Since surgical outcomes tend to improve over time, it is highly unlikely that confounding due to time-dependent factors prevented us from identifying a significant improvement after implementation of a surgical checklist.

I will leave it to you to decide if you think this is an adequate explanation. I’m not buying it.

Gawande concludes that this study reflects a failure of implementation of using checklists, rather than a failure of checklists themselves. I’m inclined to agree.

Ultimately, I don’t wonder why this study was published; bad studies are published all the time (hence the work of John Ioannidis). I wonder why this study was published in the New England Journal of Medicine. NEJM is supposed to be the gold-standard for academic medical research. If they print it, you should be confident in the results and conclusions. Their editors and peer reviewers are supposed to be the best in the world. The Ontario study seems to be far below the standards I expect for NEJM.

I think their decision to accept the paper hinged on the fact that this was a large study that showed a negative finding on a subject that has been particularly hot over the past few years [4]. Nobody seemed to care that this was not a particularly well-conducted study; this is the sadness that plagues the medical research community. Be a critical reader.


  1. Remember, we conventionally aim for a power of 80% (or better).  ↩

  2. Clinical significance refers to the importance of a finding in terms of its impact on something clinically meaningful. To use data for the Ontario study as an example, they show a statistically significant drop in the length of hospital stays from 5.11 days to 5.07 days. Despite this finding’s statistical significance, who cares?! You’re still in the hospital 5 days roughly.  ↩

  3. I am taking ‘sensitivity analysis’ to mean in this case that they actually looked at various time periods—maybe 6 or 12 or 18 months—to see how their results changed. Usually when people do this, they give some indication of the results of their sensitivity analyses and why they decided to stick with the original plan.  ↩

  4. Yes, checklists are hot. I mean, Atul Gawande wrote a best-selling book about them. Granted, he’s such a great writer that he could spend 300 pages expounding upon why the sky is blue and it would sell.  ↩

Ignorance is not bliss when it comes to health literacy | Healthcare Leadership Blog

Colin Hung:

As I looked at the screen filled with lab results, I realized that I didn’t have a clue what the information was telling me. I had no idea whether the Complete Blood Counts (CBC) or Electrolytes meant the fictitious patient needed an immediate trip to the Emergency Room or a high-five for being so healthy. 

Giving patients access to their medical data is not a complex problem. Developing strategies for helping them make sense of their medical data is incredibly complex. Patients should absolutely have easy access to their medical data, but keep in mind that access will not automatically produce meaningful action.

Hospital charges shown to vary widely | Pittsburgh Post-Gazette

Not really news if you pay even casual attention to health policy or any of the debate surround the Affordable Care Act. However, the numbers in stories like this are always staggering:

For a 200 mg gemcitabine injection, Hopkins collected $143. UPMC got $1,051.

That's over 7 times as much.

Guidelines have consequences – intended and unintended | Med Rants

We must stop insisting on calling expert opinions guidelines.  We should only call something a guideline when the data are very clear and we really have consensus on those data.

Guidelines fail when there is little or poor evidence; guidelines themselves are not inherently bad. Where we lack good evidence, we should invest in conducting sound research to fill-in our gaps in knowledge.

Perhaps when professional societies and government institutions develop guidelines they should include within them a roadmap for future research to answer looming questions...

Penn Medicine to co-develop antibiotics recommendation app for MDs | MobiHealthNews

Better information systems will revolutionize the practice of infectious diseases. The pieces [1] are already in place, we just need to stitch them together into a smarter digital tool. Penn Medicine seems to be working on such a solution.


  1. The pieces include antibiograms detailing local resistance patterns, national data tracking the spread of infectious pathogens, patient-level microbiologic data from diagnostic tests like cultures and PCR analysis, and antimicrobial pharmacokinetic data.  ↩

Why do we have to provide an admitting diagnosis? | Med Rants

Admitting diagnoses can put blinders on to other possible diagnoses. This is a real problem and I think two possible solutions exist: (1) change the term 'admitting diagnosis' to 'preliminary diagnosis' or (2) change the admitting diagnosis to a list of the differential diagnoses based on the patient's presenting symptoms. I prefer the second solution, but also think many will do a poor job of developing a rationale differential.

The Collapse of Big Law: A Cautionary Tale for Big Med | The Atlantic

A truly chilling article drawing corollaries between the contribution of performance metrics to the legal field's demise and the growing demand for similar performance metrics in medicine.

With each passing year, Big Med is following Big Law.  Physicians, medical schools, and hospitals all proudly trumpet their standing in national rankings.  Efforts to preserve and augment revenue streams produce a less patient-centered and more business-oriented approach to organizing the practice of medicine. Physicians are more and more commonly referred to as healthcare providers.

This is why the doctor-patient relationship—and using those specific identifiers—is crucial to the field of medicine. Doctors are more than care providers and patients are more than customers.

Apple exploring cars, medical devices to reignite growth | SF Gate

[Apple] wants to develop software and sensors that can predict heart attacks by identifying the sound blood makes as it tries to move through an artery clogged with plaque, the source said.

I find it highly improbable that Apple is trying to enter the medical device market. It just seems like an awkward move for them. Apple is secretive and prefers to control as much of the customer experience as possible. As 23andMe found out, the FDA likes to exercise their control over medical products. Why would Apple move into such a highly regulated market?

Regardless, I’m excited to see what they have up their sleeve, medically-focused or not.

The Hidden Curriculum: Changing The Water In Which We Swim | Health Affairs Blog

Tim Lahey writing about a recent essay in Health Affairs discussing medical education’s ‘hidden curriculum’:

There is no doubt we need a better culture of safety in medical education. In a survey of Iowa medical students, 32 percent reported inadequate communication to families, 19 percent saw patient confidentiality breached, and 14 percent witnessed deliberate deception in the context of medical care. A New York state study called out another likely universal problem: medical students fear reprisal if they report errors to protect patient safety.

[…]

Amid such efforts, we must be mindful that there is more to culture change than talking about it, or even speaking up about medical error. Culture is constructed of words, undoubtedly, but the context in which those words occur is at least as important as the words themselves. We must remember, as McLuhan reminded us, the medium is the message. And the medium in medical education—from morning report to ward rounds and every committee meeting in between—is teamwork.

Great insights throughout this article. In current medical culture, physicians (specifically attending physicians) still function as leaders of care teams. A culture of patient safety begins with their leadership and their sense of inclusivity in terms of discussing patient care with the entire team.

Doctors debate the social-media dilemma | straight.com

A Canadian perspective on the role of social media in medicine.

Dr. Kendall Ho told the Straight that physicians need to inject their expertise into the health conversations that patients are having on social media or others will fill the void.

Exactly.

The Real World Is Not an Exam | Well Blog (NY Times)

Multiple-choice board exams may not be the best assessment modality for doctors in training:

Educators may not actually teach to the test, but students think to the test, in linear multiple choice.

We spend the first few years of medical training imbuing our bright medical students with test-taking expertise focused on obscure and rare but well-characterized diseases. We then expend the remaining years breaking them of these habits to get them thinking of horses instead of zebras.

[See this related post about what med students use to study.]

Dallas Buyers Club

This movie should be required viewing for medical students, residents, pretty much any health care professional. Though there are several themes, I think it does an excellent job portraying the gulf that exists all too often between patients and doctors. Physicians tend to take a broader, more complex view while patients just want to do whatever it takes to get better. Both perspectives are valid and necessary, but are incompatible with one another without empathy. This is why I believe we need to be more inclusive of patients in all aspects of health care and do more things like patient-centered outcomes research.

Teaching Doctors the Art of Negotiation | Well Blog (NY Times)

Something I didn't expect when I started medical school—the importance of negotiation and 'selling'. Doctors are constantly negotiating with patients, which includes 'selling' them on a particular diagnostic or treatment plan. Those who do this best meet that patient halfway. They empathize with the patient's situation while estimating their health literacy (which requires knowing your patient) and think creatively how to present options in a logical, understandable way. And that simple sentence encompasses what 7 years of training is largely about. 

There's got to be a better way... | Controversies in Hospital Infection Prevention

Great post on the bureaucratic nightmare that is the 'compassionate use' process for using unapproved drugs.

I recently had the misfortune of experiencing the compassionate use process...I did a quick Google search to see how I could obtain IV zanamivir and learned that I needed to contact the drug manufacturer, the FDA and my IRB. I soon learned there were numerous forms to complete, almost all of which required me to record the same information over and over. From start to finish it took approximately 4 hours and the best word to describe the situation was kafkaesque...Maybe I'm just a simpleton, but couldn't there be a website where information is entered once and then routed to the appropriate agencies?...All of this makes me wonder how many patients don't receive treatment with potentially lifesaving drugs because the process is so painful, duplicative, time intensive and byzantine.

Similar byzantine processes exist for things like reporting adverse drug events...you know, not important stuff we don't want to have good data about.

Australian researchers sum up risks with medical apps | MobiHealthNews

Nice summary of the potential pitfalls for medical apps intended for use by medical professionals. Their last major area of concern:

Finally, the researchers cite concerns about medical app developers. Developers may lack an understanding of healthcare contexts and standards. They may be using data mining tools as mentioned above. They may also have included very little (if any) end user input during the design of their apps, which could lead to safety risks and inaccurate information or algorithms.

We need to generate a robust developer community that includes physicians in order to create a generation of great digital tools specifically designed for doctors.

✚ Healthbook - Apple's foray into mHealth

A few days ago, rumors surfaced about Apple’s upcoming release of their next iPhone operating system (iOS 8) and the much-rumored ‘iWatch’ device. The rumors include details of a new health and fitness app with the code name ‘Healthbook’. Here are a few links:

I highly recommend each of these articles to get a broad picture; for those who are time limited, the first post from 9to5Mac gives great overall context while the analyses from Fred Wilson and MG Siegler (two tech VCs) discuss important questions.

To Fred Wilson’s point about hooking collected data into existing health infrastructure via APIs, not only does there need to be data exchange but there also has to be demonstrated value. We always think more data is better. More data is only better when it is actionable. The big elephant in the mHealth/quantified self room is that no one has quite figured out what to do with all the data. Some highly motivated quantified selfers are using it to change their habits, but what impact will it have on the rest of the world?