Treatment time limit for successful weaning from veno-arterial extracorporeal membrane oxygenation in cardiogenic shock.
Knowing the upper time limit for successful weaning from temporary mechanical circulatory support in cardiogenic shock will help with decision-making regarding advanced heart failure (HF) therapy or considering withdrawal of care. The aim of this study was to investigate the association between the support duration and successful weaning from veno-arterial extracorporeal membrane oxygenation (VA-ECMO) in patients with cardiogenic shock.
A retrospective single-centre cohort study was conducted between January 2013 and June 2023. It included 100 consecutive patients with cardiogenic shock who were treated with VA-ECMO. Patients with out-of-hospital cardiac arrest were excluded. The primary outcome was successful weaning from VA-ECMO (i.e., VA-ECMO decannulation and survival to discharge). The association between the length of support duration and the weaning success rate was analysed. Patients were divided into three groups according to ECMO support duration: Group A (≤7 days), Group B (8-14 days), and Group C (≥15 days). Multivariable logistic regression analysis was used to evaluate the impact of the length of support duration on successful weaning of VA-ECMO. The median age was 67 years, and 73% of study participants were male. The underlying aetiologies of cardiogenic shock were as follows: acute myocardial infarction, 50; fulminant myocarditis, 19; cardiomyopathy, 15; valvular heart disease, 8; and other, 8. Seventy-five patients (75%) were attempted to wean VA-ECMO, and 67 moved on to decannulation. In total, 43 (43%) patients were successfully weaned from VA-ECMO. The median length of ECMO support duration was 8 [3-15] days. Compared with those who underwent successful ECMO decannulation, those who did not had a significantly longer support duration of VA-ECMO (5 [3-9] days vs. 12 [3-22] days, P = 0.004). The weaning success rate was significantly higher in patients with short support duration; 58% (29/50), 40% (10/25), 16% (4/25) in Groups A, B, and C, respectively (P = 0.002). Overall, none of the patients supported for over 24 days (0/11) were successfully weaned from VA-ECMO. On multivariable logistic regression analysis, the length of support duration was independently associated with successful weaning after adjusting for age, sex, underlying aetiology, and left ventricular ejection fraction (odds ratio, 0.813 [per 3 days]; 95% confidence interval, 0.679-0.914; P = 0.025).
Long support duration of VA-ECMO was significantly associated with a low rate of successful weaning in patients with cardiogenic shock. Patients who require VA-ECMO for over 1 week should start considering advanced HF therapy or withdrawal of care.
Suzuki S
,Ito K
,Teraoka N
,Okuma Y
,Kimura K
,Minamisawa M
,Ebisawa S
,Motoki H
,Imamura H
,Seto T
,Kuwahara K
... -
《ESC Heart Failure》
The absolute lactate levels versus clearance for prognostication of post-cardiotomy patients on veno-arterial ECMO.
Veno-arterial extracorporeal membrane oxygenation (VA-ECMO) is a life-saving procedure for supporting patients with cardiogenic shock after cardiac surgery. This work aimed to analyse the impact of changes in blood lactate levels on the survival of patients on post-cardiotomy ECMO (PC-ECMO) and whether lactate clearance (LC) performs better than absolute lactate levels.
We retrospectively analysed the data of adult patients who received PC-ECMO at our centre between 2016 and 2022. The primary outcome was the in-hospital mortality rate. Arterial lactate levels were measured at ECMO initiation, peak and 12 and 24 h after VA-ECMO support. LC was calculated at 12 and 24 h. Out of 2368 patients who received cardiac surgeries, 152 (median age, 48 years; 57.9% of them were men) received PC-ECMO. Of them, 48 (31.6%) survived and were discharged, while 104 (68.4%) died during the index hospitalization. Non-survivors had higher frequencies of atrial fibrillation (41.35% vs. 12.5%, P < 0.001), chronic kidney disease (26.9% vs. 6.3%, P = 0.004), prolonged cardiopulmonary bypass (237 vs. 192 min, P = 0.016) and aortic cross-clamping times (160 vs. 124 min, P = 0.04) than survivors. Non-survivors had a significantly higher median Sequential Organ Failure Assessment (SOFA) score at ECMO initiation (13.5 vs. 9, P < 0.001) and a lower median Survival After Veno-arterial ECMO (SAVE) score (-3 vs. 3, P < 0.001) with higher SAVE classes (P < 0.001) than survivors. After 12 h of VA-ECMO support, the blood lactate level was negatively correlated with LC in survivors (r = -0.755, P < 0.001) and non-survivors (r = -0.601, P < 0.001). After 24 h, the same negative correlation was identified between survivors (r = -0.764, P < 0.001) and non-survivors (r = -0.847, P < 0.001). Blood lactate levels measured at 12 h to determine hospital mortality [>8.2 mmol/L, area under the receiver operating characteristic curve (AUROC): 0.868] and 24 h (>2.6 mmol/L, AUROC: 0.896) had the best performance, followed by LC-T12 (<21.94%, AUROC: 0.807), LC-T24 (<40.3%, AUROC: 0.839) and peak blood lactate (>14.35 mmol/L, AUROC: 0.828). The initial pre-ECMO blood lactate (>6.25 mmol/L, AUROC: 0.731) had an acceptable ability to discriminate mortality but was less than the following measurements and clearance. Kaplan-Meier curves demonstrated that LC of <21.94% at T12 h and <40.3% at T24 h was associated with decreased survival (log-rank P < 0.001). Cox proportional hazards regression analysis for mortality revealed that LC of <21.94% at T12 h had an adjusted hazard ratio (HR) of 2.73 [95% confidence interval (CI): 1.64-5.762, P < 0.001] and LC of <40.3% at T24 h had an adjusted HR of 1.98 (95% CI: 1.46-4.173, P < 0.001). The predictors of hospital mortality after PC-ECMO were the lactate level at 12 h [odds ratio (OR): 1.67, 95% CI: 1.121-2.181, P = 0.001], initial SOFA score (OR: 1.593, 95% CI: 1.15-2.73, P < 0.001), initial blood lactate (OR: 1.21, 95% CI: 1.016-1.721, P = 0.032) and atrial fibrillation (OR: 6.17, 95% CI: 2.37-57.214, P = 0.003). Bivariate models using lactate levels and clearance at the same points revealed that blood lactate levels performed better than the clearance percentage.
Serial measurements of arterial blood lactate and LC help in obtaining early prognostic guidance in adult patients supported by VA-ECMO after cardiac surgery. Absolute lactate levels, compared with LC at the same time points, demonstrated better performance in differentiating mortality.
Laimoud M
,Machado P
,Lo MG
,Maghirang MJ
,Hakami E
,Qureshi R
... -
《ESC Heart Failure》
Acute kidney injury and cardiogenic shock severity for mortality risk stratification in patients supported with VA ECMO.
To assess the stage of acute kidney injury (AKI), as an index of organ perfusion, combined with shock severity, measured by the Society for Cardiovascular Angiography and Interventions (SCAI) shock stage classification, to stratify the risk of mortality in patients diagnosed with cardiogenic shock (CS) and supported with venoarterial extracorporeal membrane oxygenation (VA ECMO).
From January 2018 to December 2020, consecutive adult patients diagnosed with CS and received VA ECMO were retrospectively evaluated. The highest AKI stage within 48 h after ECMO initiation was assessed using the Kidney Disease: Improving Global Outcomes criteria. We included 216 patients with a mean age of 58.8 years and 31.0% were females. 88.4% of patients received ECMO for postcardiotomy, while 11.6% for medical CS. The total in-hospital mortality was 53.2%. AKI occurred in 182 (84.3%) patients receiving ECMO for CS. AKI stage 0, 1, 2, and 3 were present in 15.7%, 17.6%, 18.1%, and 48.6% of patients with in-hospital mortality of 26.5%, 26.3%, 61.5%, and 68.6%, respectively (P < 0.001). The AKI stage (P < 0.001), SCAI shock stage before ECMO (P = 0.008), and NYHA ≥ Class III on admission (P = 0.044) were independent predictors of in-hospital mortality. The area under the receiver operating characteristic curve of 0.754 (95% confidence interval: 0.690 to 0.811) for AKI stage combined with SCAI shock stage was better than those for AKI stage (0.676), SCAI shock stage (0.657), serum lactate level (0.682), SOFA score (0.644), SVAE score (0.582), and VIS score (0.530) prior to ECMO.
In this single-center CS population who received VA ECMO for circulatory support, predominantly postcardiotomy cases, AKI occurred in 84.3% of the patients. AKI stage, as an index of organ perfusion combined with shock severity measured by the SCAI shock classification, demonstrates a good correlation with in-hospital mortality.
Li C
,Wang Y
,Wang X
,Shao C
,Xin M
,Xu B
,Hao X
,Jia M
,Yang F
,Wang L
,Wang H
,Hou X
,Du Z
... -
《ESC Heart Failure》
Comparison of Two Modern Survival Prediction Tools, SORG-MLA and METSSS, in Patients With Symptomatic Long-bone Metastases Who Underwent Local Treatment With Surgery Followed by Radiotherapy and With Radiotherapy Alone.
Survival estimation for patients with symptomatic skeletal metastases ideally should be made before a type of local treatment has already been determined. Currently available survival prediction tools, however, were generated using data from patients treated either operatively or with local radiation alone, raising concerns about whether they would generalize well to all patients presenting for assessment. The Skeletal Oncology Research Group machine-learning algorithm (SORG-MLA), trained with institution-based data of surgically treated patients, and the Metastases location, Elderly, Tumor primary, Sex, Sickness/comorbidity, and Site of radiotherapy model (METSSS), trained with registry-based data of patients treated with radiotherapy alone, are two of the most recently developed survival prediction models, but they have not been tested on patients whose local treatment strategy is not yet decided.
(1) Which of these two survival prediction models performed better in a mixed cohort made up both of patients who received local treatment with surgery followed by radiotherapy and who had radiation alone for symptomatic bone metastases? (2) Which model performed better among patients whose local treatment consisted of only palliative radiotherapy? (3) Are laboratory values used by SORG-MLA, which are not included in METSSS, independently associated with survival after controlling for predictions made by METSSS?
Between 2010 and 2018, we provided local treatment for 2113 adult patients with skeletal metastases in the extremities at an urban tertiary referral academic medical center using one of two strategies: (1) surgery followed by postoperative radiotherapy or (2) palliative radiotherapy alone. Every patient's survivorship status was ascertained either by their medical records or the national death registry from the Taiwanese National Health Insurance Administration. After applying a priori designated exclusion criteria, 91% (1920) were analyzed here. Among them, 48% (920) of the patients were female, and the median (IQR) age was 62 years (53 to 70 years). Lung was the most common primary tumor site (41% [782]), and 59% (1128) of patients had other skeletal metastases in addition to the treated lesion(s). In general, the indications for surgery were the presence of a complete pathologic fracture or an impending pathologic fracture, defined as having a Mirels score of ≥ 9, in patients with an American Society of Anesthesiologists (ASA) classification of less than or equal to IV and who were considered fit for surgery. The indications for radiotherapy were relief of pain, local tumor control, prevention of skeletal-related events, and any combination of the above. In all, 84% (1610) of the patients received palliative radiotherapy alone as local treatment for the target lesion(s), and 16% (310) underwent surgery followed by postoperative radiotherapy. Neither METSSS nor SORG-MLA was used at the point of care to aid clinical decision-making during the treatment period. Survival was retrospectively estimated by these two models to test their potential for providing survival probabilities. We first compared SORG to METSSS in the entire population. Then, we repeated the comparison in patients who received local treatment with palliative radiation alone. We assessed model performance by area under the receiver operating characteristic curve (AUROC), calibration analysis, Brier score, and decision curve analysis (DCA). The AUROC measures discrimination, which is the ability to distinguish patients with the event of interest (such as death at a particular time point) from those without. AUROC typically ranges from 0.5 to 1.0, with 0.5 indicating random guessing and 1.0 a perfect prediction, and in general, an AUROC of ≥ 0.7 indicates adequate discrimination for clinical use. Calibration refers to the agreement between the predicted outcomes (in this case, survival probabilities) and the actual outcomes, with a perfect calibration curve having an intercept of 0 and a slope of 1. A positive intercept indicates that the actual survival is generally underestimated by the prediction model, and a negative intercept suggests the opposite (overestimation). When comparing models, an intercept closer to 0 typically indicates better calibration. Calibration can also be summarized as log(O:E), the logarithm scale of the ratio of observed (O) to expected (E) survivors. A log(O:E) > 0 signals an underestimation (the observed survival is greater than the predicted survival); and a log(O:E) < 0 indicates the opposite (the observed survival is lower than the predicted survival). A model with a log(O:E) closer to 0 is generally considered better calibrated. The Brier score is the mean squared difference between the model predictions and the observed outcomes, and it ranges from 0 (best prediction) to 1 (worst prediction). The Brier score captures both discrimination and calibration, and it is considered a measure of overall model performance. In Brier score analysis, the "null model" assigns a predicted probability equal to the prevalence of the outcome and represents a model that adds no new information. A prediction model should achieve a Brier score at least lower than the null-model Brier score to be considered as useful. The DCA was developed as a method to determine whether using a model to inform treatment decisions would do more good than harm. It plots the net benefit of making decisions based on the model's predictions across all possible risk thresholds (or cost-to-benefit ratios) in relation to the two default strategies of treating all or no patients. The care provider can decide on an acceptable risk threshold for the proposed treatment in an individual and assess the corresponding net benefit to determine whether consulting with the model is superior to adopting the default strategies. Finally, we examined whether laboratory data, which were not included in the METSSS model, would have been independently associated with survival after controlling for the METSSS model's predictions by using the multivariable logistic and Cox proportional hazards regression analyses.
Between the two models, only SORG-MLA achieved adequate discrimination (an AUROC of > 0.7) in the entire cohort (of patients treated operatively or with radiation alone) and in the subgroup of patients treated with palliative radiotherapy alone. SORG-MLA outperformed METSSS by a wide margin on discrimination, calibration, and Brier score analyses in not only the entire cohort but also the subgroup of patients whose local treatment consisted of radiotherapy alone. In both the entire cohort and the subgroup, DCA demonstrated that SORG-MLA provided more net benefit compared with the two default strategies (of treating all or no patients) and compared with METSSS when risk thresholds ranged from 0.2 to 0.9 at both 90 days and 1 year, indicating that using SORG-MLA as a decision-making aid was beneficial when a patient's individualized risk threshold for opting for treatment was 0.2 to 0.9. Higher albumin, lower alkaline phosphatase, lower calcium, higher hemoglobin, lower international normalized ratio, higher lymphocytes, lower neutrophils, lower neutrophil-to-lymphocyte ratio, lower platelet-to-lymphocyte ratio, higher sodium, and lower white blood cells were independently associated with better 1-year and overall survival after adjusting for the predictions made by METSSS.
Based on these discoveries, clinicians might choose to consult SORG-MLA instead of METSSS for survival estimation in patients with long-bone metastases presenting for evaluation of local treatment. Basing a treatment decision on the predictions of SORG-MLA could be beneficial when a patient's individualized risk threshold for opting to undergo a particular treatment strategy ranged from 0.2 to 0.9. Future studies might investigate relevant laboratory items when constructing or refining a survival estimation model because these data demonstrated prognostic value independent of the predictions of the METSSS model, and future studies might also seek to keep these models up to date using data from diverse, contemporary patients undergoing both modern operative and nonoperative treatments.
Level III, diagnostic study.
Lee CC
,Chen CW
,Yen HK
,Lin YP
,Lai CY
,Wang JL
,Groot OQ
,Janssen SJ
,Schwab JH
,Hsu FM
,Lin WH
... -
《-》