Week 42 – BeSt

“Clinical and Radiographic Outcomes of Four Different Treatment Strategies in Patients with Early Rheumatoid Arthritis (the BeSt Study).”

Arthritis & Rheumatism. 2005 Nov;52(11):3381-3390. [free full text]

Rheumatoid arthritis (RA) is among the most prevalent of the rheumatic diseases with a lifetime prevalence of 3.6% in women and 1.7% in men [1]. It is a chronic, systemic, inflammatory autoimmune disease of variable clinical course that can severely impact physical functional status and even mortality. Over the past 30 years, as the armamentarium of therapies for RA has exploded, there has been increased debate about the ideal initial therapy. The BeSt (Dutch: Behandel-Strategieën “treatment strategies”) trial was designed to compare, according to the authors, four of “the most frequently used and discussed strategies.” Regimens incorporating traditional disease-modifying antirheumatic drugs (DMARDs), such as methotrexate, and newer therapies, such as TNF-alpha inhibitors, were compared directly.

The trial enrolled 508 DMARD-naïve patients with early rheumatoid arthritis. Pertinent exclusion criteria included history of cancer and pre-existing laboratory abnormalities or comorbidities (e.g. elevated creatinine or ALT, alcohol abuse, pregnancy or desire to conceive, etc.) that would preclude the use of various DMARDs. Patients were randomized to one of four treatment groups. Within each regimen, the Disease Activity Score in 44 joints (DAS-44) was assessed q3 months, and, if > 2.4, the medication regimen was uptitrated to the next step within the treatment group.

Four Treatment Groups

  1. Sequential monotherapy: methotrexate (MTX) 15mg/week, uptitrated PRN to 25-30mg/week. If insufficient control, the following sequence was pursued: sulfasalazine (SSZ) monotherapy, leflunomide monotherapy, MTX + infliximab, gold with methylprednisolone, MTX + cyclosporin A (CSA) + prednisone
  2. Step-up combination therapy: MTX 15mg/week, uptitrated PRN to 25-30mg/week. If insufficient control, SSZ was added, followed by hydroxychloroquine (HCQ), followed by prednisone. If patients failed to respond to those four drugs, they were switched to MTX + infliximab, then MTX + CSA + prednisone, and finally to leflunomide.
  3. Initial combination therapy with tapered high-dose prednisone: MTX 7.5 mg/week + SSZ 2000 mg/day + prednisone 60mg/day (tapered in 7 weeks to 7.5 mg/day). If insufficient control, MTX was uptitrated to 25-30 mg/week. Next, combination would be switched to MTX + CSA + prednisone, then MTX + infliximab, then leflunomide monotherapy, gold with methylprednisolone, and finally azathioprine with prednisone.
  4. Initial combination therapy with infliximab: MTX 25-30 mg/week + infliximab 3 mg/kg at weeks 0, 2, 6, and q8 weeks thereafter. There was a protocol for infliximab-dose uptitration starting at 3 months. If insufficient control on MTX and infliximab 10 mg/kg, patients were switched to SSZ, then leflunomide, then MTX + CSA + prednisone, then gold + methylprednisolone, and finally AZA with prednisone.

Once clinical response was adequate for at least 6 months, there was a protocol for tapering the drug regimen.

The primary endpoints were: 1) functional ability per the Dutch version of the Health Assessment Questionnaire (D-HAQ), collected by a blinded research nurse q3 months and 2) radiographic joint damage per the modified Sharp/Van der Heijde score (SHS). Pertinent secondary outcomes included DAS-44 score and laboratory evidence of treatment toxicity.

At randomization, enrolled RA patients had a median duration of symptoms of 23 weeks and median duration since diagnosis of RA of 2 weeks. Mean DAS-44 was 4.4 ± 0.9. 72% of patients had erosive disease. Mean D-HAQ score at 3 months was 1.0 in groups 1 and 2 and 0.6 in groups 3 and 4 (p < 0.001 for groups 1 and 2 vs. groups 3 and 4; paired tests otherwise insignificant). Mean D-HAQ at 1 year was 0.7 in groups 1 and 2 and 0.5 in groups 3 and 4 (p = 0.010 for group 1 vs. group 3, p = 0.003 for group 1 vs. group 4; paired tests otherwise insignificant). At 1 year, patients in group 3 or 4 had less radiographic progression in joint damage per SHS than patients in group 1 or 2. Median increases in SHS were 2.0, 2.5., 1.0, and 0.5 in groups 1-4, respectively (p = 0.003 for group 1 vs. group 3, p < 0.001 for group 1 versus group 4, p = 0.007 for group 2 vs. group 3, p < 0.001 for group 2 vs. group 4). Regarding DAS-44 score: low disease activity (DAS-44 ≤ 2.4) at 1 year was reached in 53%, 64%, 71%, 74% of groups 1-4, respectively (p = 0.004 for group 1 vs. group 3, p = 0.001 for group 1 vs. group 4, p not significant for other comparisons). There were no group differences in prevalence of adverse effects.

Overall, among patients with early RA, initial combination therapy that included either prednisone (group 3) or infliximab (group 4) resulted in better functional and radiographic improvement than did initial therapy with sequential monotherapy (group 1) or step-up combination therapy (group 2). In the discussion, the authors note that given the treatment group differences in radiographic progression of disease, “starting therapy with a single DMARD would be a missed opportunity in a considerable number of patients.” Contemporary commentary by Weisman notes that “the authors describe both an argument and a counterargument arising from their observations: aggressive treatment with combinations of expensive drugs would ‘overtreat’ a large proportion of patients, yet early suppression of disease activity may have an important influence on subsequent long‐term disability and damage.”

Fourteen years later, it is a bit difficult to place the specific results of this trial in our current practice. Its trial design is absolutely byzantine and compares the 1-year experience of a variety of complex protocols that theoretically have substantial eventual potential overlap. Furthermore, it is difficult to assess if the relatively small group differences in symptom (D-HAQ) and radiographic (SHS) scales were truly clinically significant even if they were statistically significant. The American College of Rheumatology 2015 Guideline for the Treatment of Rheumatoid Arthritis synthesized the immense body of literature that came before and after the BeSt study and ultimately gave a variety of conditional statements about the “best practice” treatment of symptomatic early RA. (See Table 2 on page 8.) The recommendations emphasized DMARD monotherapy as the initial strategy but in the specific setting of a treat-to-target strategy. They also recommended escalation to combination DMARDs or biologics in patients with moderate or high disease activity despite DMARD monotherapy.

References / Additional Reading:
1. “The lifetime risk of adult-onset rheumatoid arthritis and other inflammatory autoimmune rheumatic diseases.” Arthritis Rheum. 2011 Mar;63(3):633-9. [https://www.ncbi.nlm.nih.gov/pubmed/21360492]
2. BeSt @ Wiki Journal Club
3. “Progress toward the cure of rheumatoid arthritis? The BeSt study.” Arthritis Rheum. 2005 Nov;52(11):3326-32.
4. “Review: treat to target in rheumatoid arthritis: fact, fiction, or hypothesis?” Arthritis Rheumatol. 2014 Apr;66(4):775-82. [https://www.ncbi.nlm.nih.gov/pubmed/24757129]
5. “2015 American College of Rheumatology Guideline for the Treatment of Rheumatoid Arthritis” Arthritis Rheumatol. 2016 Jan;68(1):1-26
6. RheumDAS calculator

Summary by Duncan F. Moore, MD

Image Credit: Braegel, CC BY 3.0, via Wikimedia Commons

Week 39 – POISE

“Effects of extended-release metoprolol succinate in patients undergoing non-cardiac surgery: a randomised controlled trial”

Lancet. 2008 May 31;371(9627):1839-47. [free full text]

Non-cardiac surgery is commonly associated with major cardiovascular complications. It has been hypothesized that perioperative beta blockade would reduce such events by attenuating the effects of the intraoperative increases in catecholamine levels. Prior to the 2008 POISE trial, small- and moderate-sized trials had revealed inconsistent results, alternately demonstrating benefit and non-benefit with perioperative beta blockade. The POISE trial was a large RCT designed to assess the benefit of extended-release metoprolol succinate (vs. placebo) in reducing major cardiovascular events in patients of elevated cardiovascular risk.

The trial enrolled patients age 45+ undergoing non-cardiac surgery with estimated LOS 24+ hrs and elevated risk of cardiac disease, meaning: either 1) hx of CAD, 2) peripheral vascular disease, 3) hospitalization for CHF within past 3 years, 4) undergoing major vascular surgery, 5) or any three of the following seven risk criteria: undergoing intrathoracic or intraperitoneal surgery, hx CHF, hx TIA, hx DM, Cr > 2.0, age 70+, or undergoing urgent/emergent surgery.

Notable exclusion criteria: HR < 50, 2nd or 3rd degree heart block, asthma, already on beta blocker, prior intolerance of beta blocker, hx CABG within 5 years and no cardiac ischemia since

Intervention: metoprolol succinate (extended-release) 100mg PO starting 2-4 hrs before surgery, additional 100mg at 6-12 hrs postoperatively, followed by 200mg daily for 30 days. (Patients unable to take PO meds postoperatively were given metoprolol infusion.)

Comparison: placebo PO / IV at same frequency as metoprolol arm

Outcome:
Primary – composite of cardiovascular death, non-fatal MI, and non-fatal cardiac arrest at 30 days

Secondary (at 30 days)

        • cardiovascular death
        • non-fatal MI
        • non-fatal cardiac arrest
        • all-cause mortality
        • non-cardiovascular death
        • MI
        • cardiac revascularization
        • stroke
        • non-fatal stroke
        • CHF
        • new, clinically significant atrial fibrillation
        • clinically significant hypotension
        • clinically significant bradycardia

Pre-specified subgroup analyses of primary outcome:

Results:
9298 patients were randomized. However, fraudulent activity was detected at participating sites in Iran and Colombia, and thus 947 patients from these sites were excluded from the final analyses. Ultimately, 4174 were randomized to the metoprolol group, and 4177 were randomized to the placebo group. There were no significant differences in baseline characteristics, pre-operative cardiac medications, surgery type, or anesthesia type between the two groups (see Table 1).

Regarding the primary outcome, metoprolol patients were less likely than placebo patients to experience the primary composite endpoint of cardiovascular death, non-fatal MI, and non-fatal cardiac arrest (HR 0.84, 95% CI 0.70-0.99, p = 0.0399). See Figure 2A for the relevant Kaplan-Meier curve. Note that the curves separate distinctly within the first several days.

Regarding selected secondary outcomes (see Table 3 for full list), metoprolol patients were more likely to die from any cause (HR 1.33, 95% CI 1.03-1.74, p = 0.0317). See Figure 2D for the Kaplan-Meier curve for all-cause mortality. Note that the curves start to separate around day 10. Cause of death was analyzed, and the only group difference in attributable cause was an increased number of deaths due to sepsis or infection in the metoprolol group (data not shown). Metoprolol patients were more likely to sustain a stroke (HR 2.17, 95% CI 1.26-3.74, p = 0.0053) or a non-fatal stroke (HR 1.94, 95% CI 1.01-3.69, p = 0.0450). Of all patients who sustained a non-fatal stroke, only 15-20% made a full recovery. Metoprolol patients were less likely to sustain new-onset atrial fibrillation (HR 0.76, 95% CI 0.58-0.99, p = 0.0435) and less likely to sustain a non-fatal MI (HR 0.70, 95% CI 0.57-0.86, p = 0.0008). There were no group differences in risk of cardiovascular death or non-fatal cardiac arrest. Metoprolol patients were more likely to sustain clinically significant hypotension (HR 1.55, 95% CI 1.38-1.74, P < 0.0001) and clinically significant bradycardia (HR 2.74, 95% CI 2.19-3.43, p < 0.0001).

Subgroup analysis did not reveal any significant interaction with the primary outcome by RCRI, sex, type of surgery, or anesthesia type.

Implication/Discussion:
In patients with cardiovascular risk factors undergoing non-cardiac surgery, the perioperative initiation of beta blockade decreased the composite risk of cardiovascular death, non-fatal MI, and non-fatal cardiac arrest and increased the overall mortality risk and risk of stroke.

This study affirms its central hypothesis – that blunting the catecholamine surge of surgery is beneficial from a cardiac standpoint. (Most patients in this study had an RCRI of 1 or 2.) However, the attendant increase in all-cause mortality is dramatic. The increased mortality is thought to result from delayed recognition of sepsis due to masking of tachycardia. Beta blockade may also limit the physiologic hemodynamic response necessary to successfully fight a serious infection. In retrospective analyses mentioned in the discussion, the investigators state that they cannot fully explain the increased risk of stroke in the metoprolol group. However, hypotension attributable to beta blockade explains about half of the increased number of strokes.

Overall, the authors conclude that “patients are unlikely to accept the risks associated with perioperative extended-release metoprolol.”

A major limitation of this study is the fact that 10% of enrolled patients were discarded in analysis due to fraudulent activity at selected investigation sites. In terms of generalizability, it is important to remember that POISE excluded patients who were already on beta blockers.

Currently, per expert opinion at UpToDate, it is not recommended to initiate beta blockers preoperatively in order improve perioperative outcomes. POISE is an important piece of evidence underpinning the 2014 ACC/AHA Guideline on Perioperative Cardiovascular Evaluation and Management of Patients Undergoing Noncardiac Surgery, which includes the following recommendations regarding beta blockers:

      • Beta blocker therapy should not be started on the day of surgery (Class III – Harm, Level B)
      • Continue beta blockers in patients who are on beta blockers chronically (Class I, Level B)
      • In patients with intermediate- or high-risk preoperative tests, it may be reasonable to begin beta blockers
      • In patients with ≥ 3 RCRI risk factors, it may be reasonable to begin beta blockers before surgery
      • Initiating beta blockers in the perioperative setting as an approach to reduce perioperative risk is of uncertain benefit in those with a long-term indication but no other RCRI risk factors
      • It may be reasonable to begin perioperative beta blockers long enough in advance to assess safety and tolerability, preferably > 1 day before surgery

Further Reading/References:
1. Wiki Journal Club
2. 2 Minute Medicine
3. UpToDate, “Management of cardiac risk for noncardiac surgery”
4. 2014 ACC/AHA guideline on perioperative cardiovascular evaluation and management of patients undergoing noncardiac surgery: a report of the American College of Cardiology/American Heart Association Task Force on practice guidelines.

Image Credit: Mark Oniffrey, CC BY-SA 4.0, via Wikimedia Commons

Summary by Duncan F. Moore, MD

Week 38 – Effect of Early vs. Deferred Therapy for HIV (NA-ACCORD)

“Effect of Early versus Deferred Antiretroviral Therapy for HIV on Survival”

N Engl J Med. 2009 Apr 30;360(18):1815-26 [free full text]

The optimal timing of initiation of antiretroviral therapy (ART) in asymptomatic patients with HIV has been a subject of investigation since the advent of antiretrovirals. Guidelines in 1996 recommended starting ART for all HIV-infected patients with CD4 count < 500, but over time provider concerns regarding resistance, medication nonadherence, and adverse effects of medications led to more restrictive prescribing. In the mid-2000s, guidelines recommended ART initiation in asymptomatic HIV patients with CD4 < 350. However, contemporary subgroup analysis of RCT data and other limited observational data suggested that deferring initiation of ART increased rates of progression to AIDS and mortality. Thus the NA-ACCORD authors sought to retrospectively analyze their large dataset to investigate the mortality effect of early vs. deferred ART initiation.

The study examined the cases of treatment-naïve patients with HIV and no hx of AIDS-defining illness evaluated during 1996-2005. Two subpopulations were analyzed retrospectively: CD4 count 351-500 and CD4 count 500+. No intervention was undertaken. The primary outcome was, within each CD4 sub-population, mortality in patients treated with ART within 6 months after the first CD4 count within the range of interest vs. mortality in patients for whom ART was deferred until the CD4 count fell below the range of interest.

8362 eligible patients had a CD4 count of 351-500, and of these, 2084 (25%) initiated ART within 6 months, whereas 6278 (75%) patients deferred therapy until CD4 < 351. 9155 eligible patients had a CD4 count of 500+, and of these, 2220 (24%) initiated ART within 6 months, whereas 6935 (76%) patients deferred therapy until CD4 < 500. In both CD4 subpopulations, patients in the early-ART group were older, more likely to be white, more likely to be male, less likely to have HCV, and less likely to have a history of injection drug use. Cause-of-death information was obtained in only 16% of all deceased patients. The majority of these deaths in both the early- and deferred-therapy groups were from non-AIDS-defining conditions.

In the subpopulation with CD4 351-500, there were 137 deaths in the early-therapy group vs. 238 deaths in the deferred-therapy group. Relative risk of death for deferred therapy was 1.69 (95% CI 1.26-2.26, p < 0.001) per Cox regression stratified by year. After adjustment for history of injection drug use, RR = 1.28 (95% CI 0.85-1.93, p = 0.23). In an unadjusted analysis, HCV infection was a risk factor for mortality (RR 1.85, p= 0.03). After exclusion of patients with HCV infection, RR for deferred therapy = 1.52 (95% CI 1.01-2.28, p = 0.04).

In the subpopulation with CD4 500+, there were 113 deaths in the early-therapy group vs. 198 in the deferred-therapy group. Relative risk of death for deferred therapy was 1.94 (95% CI 1.37-2.79, p < 0.001). After adjustment for history of injection drug use, RR = 1.73 (95% CI 1.08-2.78, p = 0.02). Again, HCV infection was a risk factor for mortality (RR = 2.03, p < 0.001). After exclusion of patients with HCV infection, RR for deferred therapy = 1.90 (95% CI 1.14-3.18, p = 0.01).

Thus, in a large retrospective study, the deferred initiation of antiretrovirals in asymptomatic HIV infection was associated with higher mortality.

This was the first retrospective study of early initiation of ART in HIV that was large enough to power mortality as an endpoint while controlling for covariates. However, it is limited significantly by its observational, non-randomized design that introduced substantial unmeasured confounders. A notable example is the absence of socioeconomic confounders (e.g. insurance status). Perhaps early-initiation patients were more well-off, and their economic advantage was what drove the mortality benefit rather than the early initiation of ART. This study also made no mention of the tolerability of ART or adverse reactions to it.

In the years that followed this trial, NIH and WHO consensus guidelines shifted the trend toward earlier treatment of HIV. In 2015, the INSIGHT START trial (the first large RCT of immediate vs. deferred ART) showed a definitive mortality benefit of immediate initiation of ART in patients with CD4 500+. Since that time, per UpToDate, the standard of care has been to treat “essentially all” HIV-infected patients with ART.

Further Reading/References:
1. Wiki Journal Club
2. 2 Minute Medicine
3. INSIGHT START (2015), Pubmed, NEJM PDF
4. UpToDate, “When to initiate antiretroviral therapy in HIV-infected patients”

Summary by Duncan F. Moore, MD

Image Credit: Sigve, CC0 1.0, via WikiMedia Commons

Week 37 – LOTT

“A Randomized Trial of Long-Term Oxygen for COPD with Moderate Desaturation”

by the Long-Term Oxygen Treatment Trial (LOTT) Research Group

N Engl J Med. 2016 Oct 27;375(17):1617-1627. [free full text]

The long-term treatment of severe resting hypoxemia (SpO2 < 89%) in COPD with supplemental oxygen has been a cornerstone of modern outpatient COPD management since its mortality benefit was demonstrated circa 1980. Subsequently, the utility of supplemental oxygen in COPD patients with moderate resting daytime hypoxemia (SpO2 89-93%) was investigated in trials in the 1990s; however, such trials were underpowered to assess mortality benefit. Ultimately, the LOTT trial was funded by the NIH and Centers for Medicare and Medicaid Services (CMS) primarily to determine if there was a mortality benefit to supplemental oxygen in COPD patients with moderate hypoxemia as well to analyze as numerous other secondary outcomes, such as hospitalization rates and exercise performance.

The LOTT trial was originally planned to enroll 3500 patients. However, after 7 months the trial had randomized only 34 patients, and mortality had been lower than anticipated. Thus in late 2009 the trial was redesigned to include broader inclusion criteria (now patients with exercise-induced hypoxemia could qualify) and the primary endpoint was broadened from mortality to a composite of time to first hospitalization or death.

The revised LOTT trial enrolled COPD patients with moderate resting hypoxemia (SpO2 89-93%) or moderate exercise-induced desaturation during the 6-minute walk test (SpO2 ≥ 80% for ≥ 5 minutes and < 90% for ≥ 10 seconds). Patients were randomized to either supplemental oxygen (24-hour oxygen if resting SpO2 89-93%, otherwise oxygen only during sleep and exercise if the desaturation occurred only during exercise) or to usual care without supplemental oxygen. Supplemental oxygen flow rate was 2 liters per minute and could be uptitrated by protocol among patients with exercise-induced hypoxemia. The primary outcome was time to composite of first hospitalization or death. Secondary outcomes included hospitalization rates, lung function, performance on 6-minute walk test, and quality of life.

368 patients were randomized to the supplemental-oxygen group and 370 to the no-supplemental-oxygen group. Of the supplemental-oxygen group, 220 patients were prescribed 24-hour oxygen support, and 148 were prescribed oxygen for use during exercise and sleep only. Median duration of follow-up was 18.4 months. Regarding the primary outcome, there was no group difference in time to death or first hospitalization (p = 0.52 by log-rank test). See Figure 1A. Furthermore, there were no treatment-group differences in the primary outcome among patients of the following pre-specified subgroups: type of oxygen prescription, “desaturation profile,” race, sex, smoking status, SpO2 nadir during 6-minute walk, FEV1, BODE  index, SF-36 physical-component score, BMI, or history of anemia. Patients with a COPD exacerbation in the 1-2 months prior to enrollment, age 71+ at enrollment, and those with lower Quality of Well-Being Scale score at enrollment all demonstrated benefit from supplemental O2, but none of these subgroup treatment effects were sustained when the analyses were adjusted for multiple comparisons. Regarding secondary outcomes, there were no treatment-group differences in rates of all-cause hospitalizations, COPD-related hospitalizations, or non-COPD-related hospitalizations, and there were no differences in change from baseline measures of quality of life, anxiety, depression, lung function, and distance achieved in 6-minute walk.

The LOTT trial presents compelling evidence that there is no significant benefit, mortality or otherwise, of oxygen supplementation in patients with COPD and either moderate hypoxemia at rest (SpO2 > 88%) or exercise-induced hypoxemia. Although this trial’s substantial redesign in its early course is noted, the trial still is our best evidence to date about the benefit (or lack thereof) of oxygen in this patient group. As acknowledged by the authors, the trial may have had significant selection bias in referral. (Many physicians did not refer specific patients for enrollment because “they were too ill or [were believed to have benefited] from oxygen.”) Another notable limitation of this study is that nocturnal oxygen saturation was not evaluated. The authors do note that “some patients with COPD and severe nocturnal desaturation might benefit from nocturnal oxygen supplementation.”

For further contemporary contextualization of the study, please see the excellent post at PulmCCM from 11/2016. Included in that post is a link to an overview and Q&A from the NIH regarding the LOTT study.

References / Additional Reading:
1. PulmCCM, “Long-term oxygen brought no benefits for moderate hypoxemia in COPD”
2. LOTT @ 2 Minute Medicine
3. LOTT @ ClinicalTrials.gov
4. McDonald, J.H. 2014. Handbook of Biological Statistics (3rd ed.). Sparky House Publishing, Baltimore, Maryland.
5. Centers for Medicare and Medicaid Services, “Certificate of Medical Necessity CMS-484– Oxygen”
6. Ann Am Thorac Soc. 2018 Dec;15(12):1369-1381. “Optimizing Home Oxygen Therapy. An Official American Thoracic Society Workshop Report.”

Summary by Duncan F. Moore, MD

Image Credit: Patrick McAleer, CC BY-SA 2.0 UK, via Wikimedia Commons

Week 34 – PLCO

“Mortality Results from a Randomized Prostate-Cancer Screening Trial”

by the Prostate, Lung, Colorectal, and Ovarian (PLCO) Cancer Screening Trial project team

N Engl J Med. 2009 Mar 26;360(13):1310-9. [free full text]

The use of prostate-specific-antigen (PSA) testing to screen for prostate cancer has been a contentious subject for decades. Prior to the 2009 PLCO trial, there were no high-quality prospective studies of the potential benefit of PSA testing.

The trial enrolled men ages 55-74 (excluded if hx prostate, lung, or colorectal cancer, current cancer treatment, or > 1 PSA test in the past 3 years). Patients were randomized to annual PSA testing for 6 years with annual digital rectal exam (DRE) for 4 years or to usual care. The primary outcome was the prostate-cancer-attributable death rate, and the secondary outcome was the incidence of prostate cancer.

38,343 patients were randomized to the screening group, and 38,350 were randomized to the usual-care group. Baseline characteristics were similar in both groups. Median follow-up duration was 11.5 years. Patients in the screening group were 85% compliant with PSA testing and 86% compliant with DRE. In the usual-care group, 40% of patients received a PSA test within the first year, and 52% received a PSA test by the sixth year. Cumulative DRE rates in the usual-care group were between 40-50%. By seven years, there was no significant difference in rates of death attributable to prostate cancer. There were 50 deaths in the screening group and only 44 in the usual-care group (rate ratio 1.13, 95% CI 0.75 – 1.70). At ten years, there were 92 and 82 deaths in the respective groups (rate ratio 1.11, 95% CI 0.83–1.50). By seven years, there was a higher rate of prostate cancer detection in the screening group. 2820 patients were diagnosed in the screening group, but only 2322 were diagnosed in the usual-care group (rate ratio 1.22, 95% CI 1.16–1.29). By ten years, there were 3452 and 2974 diagnoses in the respective groups (rate ratio 1.17, 95% CI 1.11–1.22). Treatment-related complications (e.g. infection, incontinence, impotence) were not reported in this study.

In summary, yearly PSA screening increased the prostate cancer diagnosis rate but did not impact prostate-cancer mortality when compared to the standard of care. However, there were relatively high rates of PSA testing in the usual-care group (40-50%). The authors cite this finding as a probable major contributor to the lack of mortality difference. Other factors that may have biased to a null result were prior PSA testing and advances in treatments for prostate cancer during the trial. Regarding the former, 44% of men in both groups had already had one or more PSA tests prior to study enrollment. Prior PSA testing likely contributed to selection bias.

PSA screening recommendations prior to this 2009 study:

      • American Urological Association and American Cancer Society – recommended annual PSA and DRE, starting at age 50 if normal risk and earlier in high-risk men
      • National Comprehensive Cancer Network: “a risk-based screening algorithm, including family history, race, and age”
      • 2008 USPSTF Guidelines: insufficient evidence to determine balance between risks/benefits of PSA testing in men younger than 75; recommended against screening in age 75+ (Grade I Recommendation)

The authors of this study conclude that their results “support the validity of the recent [2008] recommendations of the USPSTF, especially against screening all men over the age of 75.”

However, the conclusions of the European Randomized Study of Screening for Prostate Cancer (ERSPC), which was published concurrently with PLCO in NEJM, differed. In ERSPC, PSA was screened every 4 years. The authors found an increased rate of detection of prostate cancer, but, more importantly, they found that screening decreased prostate cancer mortality (adjusted rate ratio 0.80, 95% CI 0.65–0.98, p = 0.04; NNT 1410 men receiving 1.7 screening visits over 9 years). Like PLCO, this study did not report treatment harms that may have been associated with overly zealous diagnosis.

The USPSTF reexamined its PSA guidelines in 2012. Given the lack of mortality benefit in PLCO, the pitiful mortality benefit in ERSPC, and the assumed harm from over-diagnosis and excessive intervention in patients who would ultimately not succumb to prostate cancer, the USPSTF concluded that PSA-based screening for prostate cancer should not be offered (Grade D Recommendation).

In the following years, the pendulum has swung back partially toward screening. In May 2018, the USPSTF released new recommendations that encourage men ages 55-69 to have an informed discussion with their physician about potential benefits and harms of PSA-based screening (Grade C Recommendation). The USPSTF continues to recommend against screening in patients over 70 years old (Grade D).

Screening for prostate cancer remains a complex and controversial topic. Guidelines from the American Cancer Society, American Urological Association, and USPSTF vary, but ultimately all recommend shared decision-making. UpToDate has a nice summary of talking points culled from several sources.

Further Reading/References:
#. PLCO @ 2 Minute Medicine
#. ERSPC @ Wiki Journal Club
#. UpToDate, Screening for Prostate Cancer

Summary by Duncan F. Moore, MD

Image Credit: Otis Brawley, Public Domain, NIH National Cancer Institute Visuals Online

Week 33 – Varenicline vs. Bupropion and Placebo for Smoking Cessation

“Varenicline, an α2β2 Nicotinic Acetylcholine Receptor Partial Agonist, vs Sustained-Release Bupropion and Placebo for Smoking Cessation”

JAMA. 2006 Jul 5;296(1):56-63. [free full text]

Assisting our patients in smoking cessation is a fundamental aspect of outpatient internal medicine. At the time of this trial, the only approved pharmacotherapies for smoking cessation were nicotine replacement therapy and bupropion. As the α2β2 nicotinic acetylcholine receptor (nAChR) was thought to be crucial to the reinforcing effects of nicotine, it was hypothesized that a partial agonist for this receptor could yield sufficient effect to satiate cravings and minimize withdrawal symptoms but also limit the reinforcing effects of exogenous nicotine. Thus Pfizer designed this large phase 3 trial to test the efficacy of its new α2β2 nAChR partial agonist varenicline (Chantix) against the only other non-nicotine pharmacotherapy at the time (bupropion) as well as placebo.

The trial enrolled adult smokers (10+ cigarettes per day) with fewer than three months of smoking abstinence in the past year (notable exclusion criteria included numerous psychiatric and substance use comorbidities). Patients were randomized to 12 weeks of treatment with either varenicline uptitrated by day 8 to 1mg BID, bupropion SR uptitrated by day 4 to 150mg BID, or placebo BID. Patients were also given a smoking cessation self-help booklet at the index visit and encouraged to set a quit date of day 8. Patients were followed at weekly clinic visits for the first 12 weeks (treatment duration) and then a mixture of clinic and phone visits for weeks 13-52. Non-smoking status during follow-up was determined by patient self-report combined with exhaled carbon monoxide < 10ppm. The primary endpoint was the 4-week continuous abstinence rate for study weeks 9-12 (as confirmed by exhaled CO level). Secondary endpoints included the continuous abstinence rate for weeks 9-24 and for weeks 9-52.

1025 patients were randomized. Compliance was similar among the three groups and the median duration of treatment was 84 days. Loss to follow-up was similar among the three groups. CO-confirmed continuous abstinence during weeks 9-12 was 44.0% among the varenicline group vs. 17.7% among the placebo group (OR 3.85, 95% CI 2.70–5.50, p < 0.001) vs. 29.5% among the bupropion group (OR vs. varenicline group 1.93, 95% CI 1.40–2.68, p < 0.001). (OR for bupropion vs. placebo was 2.00, 95% CI 1.38–2.89, p < 0.001.)  Continuous abstinence for weeks 9-24 was 29.5% among the varenicline group vs. 10.5% among the placebo group (p < 0.001) vs. 20.7% among the bupropion group (p = 0.007). Continuous abstinence rates weeks 9-52 were 21.9% among the varenicline group vs. 8.4% among placebo group (p < 0.001) vs. 16.1% among the bupropion group (p = 0.057). Subgroup analysis of the primary outcome by sex did not yield significant differences in drug efficacy by sex.

This study demonstrated that varenicline was superior to both placebo and bupropion in facilitating smoking cessation at up to 24 weeks. At greater than 24 weeks, varenicline remained superior to placebo but was similarly efficacious as bupropion. This was a well-designed and executed large, double-blind, placebo- and active-treatment-controlled multicenter US trial. The trial was completed in April 2005 and a new drug application for varenicline (Chantix) was submitted to the FDA in November 2005. Of note, an “identically designed” (per this study’s authors), manufacturer-sponsored phase 3 trial was performed in parallel and reported very similar results in the in the same July 2006 issue of JAMA (PMID: 16820547) as the above study by Gonzales et al. These robust, positive-outcome pre-approval trials of varenicline helped the drug rapidly obtain approval in May 2006.

Per expert opinion at UpToDate, varenicline remains a preferred first-line pharmacotherapy for smoking cessation. Bupropion is a suitable, though generally less efficacious, alternative, particularly when the patient has comorbid depression. Per UpToDate, the recent (2016) EAGLES trial demonstrated that “in contrast to earlier concerns, varenicline and bupropion have no higher risk of associated adverse psychiatric effects than [nicotine replacement therapy] in smokers with comorbid psychiatric disorders.”

Further Reading/References:
1. This trial @ ClinicalTrials.gov
2. Sister trial: “Efficacy of varenicline, an alpha4beta2 nicotinic acetylcholine receptor partial agonist, vs placebo or sustained-release bupropion for smoking cessation: a randomized controlled trial.” JAMA. 2006 Jul 5;296(1):56-63.
3. Chantix FDA Approval Letter 5/10/2006
4. Rigotti NA. Pharmacotherapy for smoking cessation in adults. Post TW, ed. UpToDate. Waltham, MA: UpToDate Inc. [https://www.uptodate.com/contents/pharmacotherapy-for-smoking-cessation-in-adults] (Accessed on February 16, 2019).
5. “Neuropsychiatric safety and efficacy of varenicline, bupropion, and nicotine patch in smokers with and without psychiatric disorders (EAGLES): a double-blind, randomised, placebo-controlled clinical trial.” Lancet. 2016 Jun 18;387(10037):2507-20.
6. 2 Minute Medicine: “Varenicline and bupropion more effective than varenicline alone for tobacco abstinence”
7. 2 Minute Medicine: “Varenicline safe for smoking cessation in patients with stable major depressive disorder”

Summary by Duncan F. Moore, MD

Image Credit: Сергей Фатеев, CC BY-SA 3.0, via Wikimedia Commons

Week 32 – ARISTOTLE

“Apixaban versus Warfarin in Patients with Atrial Fibrillation”

N Engl J Med. 2011 Sep 15;365(11):981-92. [free full text]

Prior to the development of the DOACs, warfarin was the standard of care for the reduction of risk of stroke in atrial fibrillation. Drawbacks of warfarin include a narrow therapeutic range, numerous drug and dietary interactions, the need for frequent monitoring, and elevated bleeding risk. Around 2010, the definitive RCTs for the oral direct thrombin inhibitor dabigatran (RE-LY) and the oral factor Xa inhibitor rivaroxaban (ROCKET AF) showed equivalence or superiority to warfarin. Shortly afterward, the ARISTOTLE trial demonstrated the superiority of the oral factor Xa inhibitor apixaban (Eliquis).

The trial enrolled patients with atrial fibrillation or flutter with at least one additional risk factor for stroke (age 75+, prior CVA/TIA, symptomatic CHF, or reduced LVEF). Notably, patients with Cr > 2.5 were excluded. Patients were randomized to treatment with either apixaban BID + placebo warfarin daily (reduced 2.5mg apixaban dose given in patients with 2 or more of the following: age 80+, weight < 60, Cr > 1.5) or to placebo apixaban BID + warfarin daily. The primary efficacy outcome was the incidence of stroke, and the primary safety outcome was “major bleeding” (clinically overt and accompanied by Hgb drop of ≥ 2, “occurring at a critical site,” or resulting in death). Secondary outcomes included all-cause mortality and a composite of major bleeding and “clinically-relevant non-major bleeding.”

9120 patients were assigned to the apixaban group, and 9081 were assigned to the warfarin group. Mean CHADS2 score was 2.1. Fewer patients in the apixaban group discontinued their assigned study drug. Median duration of follow-up was 1.8 years. The incidence of stroke was 1.27% per year in the apixaban group vs. 1.60% per year in the warfarin group (HR 0.79, 95% CI 0.66-0.95, p<0.001). This reduction was consistent across all major subgroups (see Figure 2). Notably, the rate of hemorrhagic stroke was 49% lower in the apixaban group, and the rate of ischemic stroke was 8% lower in the apixaban group. All-cause mortality was 3.52% per year in the apixaban group vs. 3.94% per year in the warfarin group (HR 0.89, 95% CI 0.80-0.999, p=0.047). The incidence of major bleeding was 2.13% per year in the apixaban group vs. 3.09% per year in the warfarin group (HR 0.69, 95% CI 0.60-0.80, p<0.001). The rate of intracranial hemorrhage was 0.33% per year in the apixaban group vs. 0.80% per year in the warfarin group (HR 0.42, 95% CI 0.30-0.58, p<0.001). The rate of any bleeding was 18.1% per year in the apixaban group vs. 25.8% in the warfarin group (p<0.001).

In patients with non-valvular atrial fibrillation and at least one other risk factor for stroke, anticoagulation with apixaban significantly reduced the risk of stroke, major bleeding, and all-cause mortality relative to anticoagulation with warfarin. This was a large RCT that was designed and powered to demonstrate non-inferiority but in fact was able to demonstrate the superiority of apixaban. Along with ROCKET AF and RE-LY, the ARISTOTLE trial ushered in the modern era of DOACs in atrial fibrillation. Apixaban was approved by the FDA for the treatment of non-valvular atrial fibrillation in 2012. Patient prescription cost is no longer a major barrier to prescription. These three major DOACs are all preferred in the DC Medicaid formulary (see page 14). To date, no trial has compared the various DOACs directly.

Further Reading/References:
1. ARISTOTLE @ Wiki Journal Club
2. 2 Minute Medicine
3. “Oral anticoagulants for prevention of stroke in atrial fibrillation: systematic review, network meta-analysis, and cost-effectiveness analysis,” BMJ 2017

Summary by Duncan F. Moore, MD

Week 31 – Early TIPS in Cirrhosis with Variceal Bleeding

“Early Use of TIPS in Patients with Cirrhosis and Variceal Bleeding”

N Engl J Med. 2010 Jun 24;362(25):2370-9. [free full text]

Variceal bleeding is a major cause of morbidity and mortality in decompensated cirrhosis. The standard of care for an acute variceal bleed includes a combination of vasoactive drugs, prophylactic antibiotics, and endoscopic techniques (e.g. banding). Transjugular intrahepatic portosystemic shunt (TIPS) can be used to treat refractory bleeding. This 2010 trial sought to determine the utility of early TIPS during the initial bleed in high-risk patients when compared to standard therapy.

The trial enrolled cirrhotic patients (Child-Pugh class B or C with score ≤ 13) with acute esophageal variceal bleeding. All patients received endoscopic band ligation (EBL) or endoscopic injection sclerotherapy (EIS) at the time of diagnostic endoscopy. All patients also received vasoactive drugs (terlipressin, somatostatin, or octreotide). Patients were randomized to either TIPS performed within 72 hours after diagnostic endoscopy or to “standard therapy” by 1) treatment with vasoactive drugs with transition to nonselective beta blocker when patients were free of bleeding followed by 2) addition of isosorbide mononitrate to maximum tolerated dose, and 3) a second session of EBL at 7-14 days after the initial session (repeated q10-14 days until variceal eradication was achieved). The primary outcome was a composite of failure to control acute bleeding or failure to prevent “clinically significant” variceal bleeding (requiring hospital admission or transfusion) at 1 year after enrollment. Selected secondary outcomes included 1-year mortality, development of hepatic encephalopathy (HE), ICU days, and hospital LOS.

359 patients were screened for inclusion, but ultimately only 63 were randomized. Baseline characteristics were similar among the two groups except that the early TIPS group had a higher rate of patients with previous hepatic encephalopathy. The primary composite endpoint of failure to control acute bleeding or rebleeding within 1 year occurred in 14 of 31 (45%) patients in the pharmacotherapy-EBL group and in only 1 of 32 (3%) of the early TIPS group (p = 0.001). The 1-year actuarial probability of remaining free of the primary outcome was 97% in the early TIPS group vs. 50% in the pharmacotherapy-EBL group (ARR 47 percentage points, 95% CI 25-69 percentage points, NNT 2.1). Regarding mortality, at one year, 12 of 31 (39%) patients in the pharmacotherapy-EBL group had died, while only 4 of 32 (13%) in the early TIPS group had died (p = 0.001, NNT = 4.0). There were no group differences in prevalence of HE at one year (28% in the early TIPS group vs. 40% in the pharmacotherapy-EBL group, p = 0.13). Additionally, there were no group differences in 1-year actuarial probability of new or worsening ascites. There were also no differences in length of ICU stay or hospitalization duration.

Early TIPS in acute esophageal variceal bleeding, when compared to standard pharmacotherapy and endoscopic band ligation, improved control of index bleeding, reduced recurrent variceal bleeding at 1 year, and reduced all-cause mortality. Prior studies had demonstrated that TIPS reduced the rebleeding rate but increased the rate of hepatic encephalopathy without improving survival. As such, TIPS had only been recommended as a rescue therapy. Obviously, this study presents compelling data that challenge these paradigms. The authors note that in “patients with Child-Pugh class C or in class B with active variceal bleeding, failure to initially control the bleeding or early rebleeding contributes to further deterioration in liver function, which in turn worsens the prognosis and may preclude the use of rescue TIPS.” Authors at UpToDate note that, given the totality of evidence to date, the benefit of early TIPS in preventing rebleeding “is offset by its failure to consistently improve survival and increasing morbidity due to the development of liver failure and encephalopathy.” Today, TIPS remains primarily a salvage therapy for use in cases of recurrent bleeding despite standard pharmacotherapy and EBL. There may be a subset of patients in whom early TIPS is the ideal strategy, but further trials will be required to identify this subset.

Further Reading/References:
1. Wiki Journal Club []
2. 2 Minute Medicine []
3. UpToDate, “Prevention of recurrent variceal hemorrhage in patients with cirrhosis

Summary by Duncan F. Moore, MD

Week 30 – Bicarbonate and Progression of CKD

“Bicarbonate Supplementation Slows Progression of CKD and Improves Nutritional Status”

J Am Soc Nephrol. 2009 Sep;20(9):2075-84. [free full text]

Metabolic acidosis is a common complication of advanced CKD. Some animal models of CKD have suggested that worsening metabolic acidosis is associated with worsening proteinuria, tubulointerstitial fibrosis, and acceleration of decline of renal function. Short-term human studies have demonstrated that bicarbonate administration reduces protein catabolism and that metabolic acidosis is an independent risk factor for acceleration of decline of renal function. However, until this 2009 study by de Brito-Ashurst et al., there were no long-term studies demonstrating the beneficial effects of oral bicarbonate administration on CKD progression and nutritional status.

The study enrolled CKD patients with CrCl 15-30ml/min and plasma bicarbonate 16-20 mEq/L and randomized them to treatment with either sodium bicarbonate 600mg PO TID (with protocolized uptitration to achieve plasma HCO3  ≥ 23 mEq/L) for 2 years, or to routine care. The primary outcomes were: 1) the decline in CrCl at 2 years, 2) “rapid progression of renal failure” (defined as decline of CrCl > 3 ml/min per year), and 3) development of ESRD requiring dialysis. Secondary outcomes included 1) change in dietary protein intake, 2) change in normalized protein nitrogen appearance (nPNA), 3) change in serum albumin, and 4) change in mid-arm muscle circumference.

134 patients were randomized, and baseline characteristics were similar among the two groups. Serum bicarbonate levels increased significantly in the treatment arm. (See Figure 2.) At two years, CrCl decline was 1.88 ml/min in the treatment group vs. 5.93 ml/min in the control group (p < 0.01). Rapid progression of renal failure was noted in 9% of intervention group vs. 45% of the control group (RR 0.15, 95% CI 0.06–0.40, p < 0.0001, NNT = 2.8), and ESRD developed in 6.5% of the intervention group vs. 33% of the control group (RR 0.13, 95% CI 0.04–0.40, p < 0.001; NNT = 3.8). Regarding nutritional status, dietary protein intake increased in the treatment group relative to the control group (p < 0.007). Normalized protein nitrogen appearance decreased in the treatment group and increased in the control group (p < 0.002). Serum albumin increased in the treatment group but was unchanged in the control group, and mean mid-arm muscle circumference increased by 1.5 cm in the intervention group vs. no change in the control group (p < 0.03).

In conclusion, oral bicarbonate supplementation in CKD patients with metabolic acidosis reduces the rate of CrCl decline and progression to ESRD and improves nutritional status. Primarily on the basis of this study, the KDIGO 2012 guidelines for the management of CKD recommend oral bicarbonate supplementation to maintain serum bicarbonate within the normal range (23-29 mEq/L). This is a remarkably cheap and effective intervention. Importantly, the rates of adverse events, particularly worsening hypertension and increasing edema, were unchanged among the two groups. Of note, sodium bicarbonate induces much less volume expansion than a comparable sodium load of sodium chloride.

In their discussion, the authors suggest that their results support the hypothesis of Nath et al. (1985) that “compensatory changes [in the setting of metabolic acidosis] such as increased ammonia production and the resultant complement cascade activation in remnant tubules in the declining renal mass [are] injurious to the tubulointerstitium.” The hypercatabolic state of advanced CKD appears to be mitigated by bicarbonate supplementation. The authors note that “an optimum nutritional status has positive implications on the clinical outcomes of dialysis patients, whereas [protein-energy wasting] is associated with increased morbidity and mortality.”

Limitations to this trial include its open-label, no-placebo design. Also, the applicable population is limited by study exclusion criteria of morbid obesity, overt CHF, and uncontrolled HTN.

Further Reading:
1. Nath et al. “Pathophysiology of chronic tubulo-interstitial disease in rats: Interactions of dietary acid load, ammonia, and complement component-C3” (1985)
2. KDIGO 2012 Clinical Practice Guideline for the Evaluation and Management of Chronic Kidney Disease (see page 89)
3. UpToDate, “Pathogenesis, consequences, and treatment of metabolic acidosis in chronic kidney disease”

Week 28 – Symptom-Triggered Benzodiazepines in Alcohol Withdrawal

“Symptom-Triggered vs Fixed-Schedule Doses of Benzodiazepine for Alcohol Withdrawal”

Arch Intern Med. 2002 May 27;162(10):1117-21. [free full text]

Treatment of alcohol withdrawal with benzodiazepines has been the standard of care for decades. However, in the 1990s, benzodiazepine therapy for alcohol withdrawal was generally given via fixed doses. In 1994, a double-blind RCT by Saitz et al. demonstrated that symptom-triggered therapy based on responses to the CIWA-Ar scale reduced treatment duration and the amount of benzodiazepine used relative to a fixed-schedule regimen. This trial had little immediate impact in the treatment of alcohol withdrawal. The authors of the 2002 double-blind RCT sought to confirm the findings from 1994 in a larger population that did not exclude patients with a history of seizures or severe alcohol withdrawal.

The trial enrolled consecutive patients admitted to the inpatient alcohol treatment units at two European universities (excluding those with “major cognitive, psychiatric, or medical comorbidity”) and randomized them to treatment with either scheduled placebo (30mg q6hrs x4, followed by 15mg q6hrs x8) with additional PRN oxazepam 15mg for CIWA score 8-15 and 30mg for CIWA score > 15 or to treatment with scheduled oxazepam (30mg q6hrs x4, followed by 15mg q6hrs x8) with additional PRN oxazepam 15mg for CIWA score 8-15 and 30mg for CIWA score > 15.

The primary outcomes were cumulative oxazepam dose at 72 hours and duration of treatment with oxazepam. Subgroup analysis included the exclusion of symptomatic patients who did not require any oxazepam. Secondary outcomes included incidence of seizures, hallucinations, and delirium tremens at 72 hours.

Results:
117 patients completed the trial. 56 had been randomized to the symptom-triggered group, and 61 had been randomized to the fixed-schedule group. The groups were similar in all baseline characteristics except that the fixed-schedule group had on average a 5-hour longer interval since last drink prior to admission. While only 39% of the symptom-triggered group actually received oxazepam, 100% of the fixed-schedule group did (p < 0.001). Patients in the symptom-triggered group received a mean cumulative dose of 37.5mg versus 231.4mg in the fixed-schedule group (p < 0.001). The mean duration of oxazepam treatment was 20.0 hours in the symptom-triggered group versus 62.7 hours in the fixed-schedule group. The group difference in total oxazepam dose persisted even when patients who did not receive any oxazepam were excluded. Among patients who did receive oxazepam, patients in the symptom-triggered group received 95.4 ± 107.7mg versus 231.4 ± 29.4mg in the fixed-dose group (p < 0.001). Only one patient in the symptom-triggered group sustained a seizure. There were no seizures, hallucinations, or episodes of delirium tremens in any of the other 116 patients. The two treatment groups had similar quality-of-life and symptom scores aside from slightly higher physical functioning in the symptom-triggered group (p < 0.01). See Table 2.

Implication/Discussion:
Symptom-triggered administration of benzodiazepines in alcohol withdrawal led to a six-fold reduction in cumulative benzodiazepine use and a much shorter duration of pharmacotherapy than fixed-schedule administration. This more restrictive and responsive strategy did not increase the risk of major adverse outcomes such as seizure or DTs and also did not result in increased patient discomfort.

Overall, this study confirmed the findings of the landmark study by Saitz et al. from eight years prior. Additionally, this trial was larger and did not exclude patients with a prior history of withdrawal seizures or severe withdrawal. The fact that both studies took place in inpatient specialty psychiatry units limits their generalizability to our inpatient general medicine populations.

Why the initial 1994 study did not gain clinical traction remains unclear. Both studies have been well-cited over the ensuing decades, and the paradigm has shifted firmly toward symptom-triggered benzodiazepine regimens using the CIWA scale. While a 2010 Cochrane review cites only the 1994 study, Wiki Journal Club and 2 Minute Medicine have entries on this 2002 study but not on the equally impressive 1994 study.

Further Reading/References:
1. “Individualized treatment for alcohol withdrawal. A randomized double-blind controlled trial.” JAMA. 1994.
2. Clinical Institute Withdrawal Assessment of Alcohol Scale, Revised (CIWA-Ar)
3. Wiki Journal Club
4. 2 Minute Medicine
5. “Benzodiazepines for alcohol withdrawal.” Cochrane Database Syst Rev. 2010

Summary by Duncan F. Moore, MD

Image Credit: VisualBeo, CC BY-SA 3.0, via Wikimedia Commons