Medicine (Austin & Northern Health) - Research Publications

Permanent URI for this collection

Search Results

Now showing 1 - 10 of 107
  • Item
    No Preview Available
    Intensive exercise program after spinal cord injury (SCIPA full-on): A randomized controlled trial
    Galea, M ; Dunlop, S ; Geraghty, T ; Davis, G ; Nunn, A ; Olenko, L (Elsevier, 2018-01-01)
    Introduction/Background: While upper body training is effective for improving aerobic fitness and muscle strength after spinal cord injury (SCI), activity-based therapies (ABT) intended to activate the paralysed extremities have been reported to promote neurological improvement. We investigated the effectiveness of intensive ABT compared with upper body training for people with SCI. Material and method: One hundred and sixteen participants (C2-T12, AIS A-D) were recruited from six SCI units in Australia and New Zealand, and randomised to experimental or control groups. Experimental participants received a 12-week ABT program including locomotor training, functional electrical stimulation-assisted leg cycling, and trunk and lower extremity exercises, while control group participants undertook upper body strength and aerobic fitness training. The primary outcome measure was the ASIA motor scores for upper and lower extremities. Results: One hundred and three participants completed the interventions and were included in the primary analysis. Mean (SD) upper extremity motor scores for experimental (n = 49) and control (n = 54) groups were 41.45 (12.11) and 39.39 (11.94), respectively, with an adjusted mean between-group difference of −0.039 (95% CI: −1.12 to 1.04). Mean (SD) lower extremity motor scores were 12.51 (17) and 10.24 (17.19) for experimental and control groups, with an adjusted mean between-group difference of 0.895 (95% CI: −0.48 to 2.27). There were 15 serious adverse events within each group, but only one of these was related to the experimental intervention (bilateral femoral condyle and tibial plateau subchondral fractures). Conclusion: Activity-based therapy did not lead to greater improvements in ASIA motor scores compared to upper body training
  • Item
    No Preview Available
    Healthy Life-Year Costs of Treatment Speed From Arrival to Endovascular Thrombectomy in Patients With Ischemic Stroke A Meta-analysis of Individual Patient Data From 7 Randomized Clinical Trials
    Almekhlafi, MA ; Goyal, M ; Dippel, DWJ ; Majoie, CBLM ; Campbell, BCV ; Muir, KW ; Demchuk, AM ; Bracard, S ; Guillemin, F ; Jovin, TG ; Mitchell, P ; White, P ; Hill, MD ; Brown, S ; Saver, JL (AMER MEDICAL ASSOC, 2021-06)
    IMPORTANCE: The benefits of endovascular thrombectomy (EVT) are time dependent. Prior studies may have underestimated the time-benefit association because time of onset is imprecisely known. OBJECTIVE: To assess the lifetime outcomes associated with speed of endovascular thrombectomy in patients with acute ischemic stroke due to large-vessel occlusion (LVO). DATA SOURCES: PubMed was searched for randomized clinical trials of stent retriever thrombectomy devices vs medical therapy in patients with anterior circulation LVO within 12 hours of last known well time, and for which a peer-reviewed, complete primary results article was published by August 1, 2020. STUDY SELECTION: All randomized clinical trials of stent retriever thrombectomy devices vs medical therapy in patients with anterior circulation LVO within 12 hours of last known well time were included. DATA EXTRACTION/SYNTHESIS: Patient-level data regarding presenting clinical and imaging features and functional outcomes were pooled from the 7 retrieved randomized clinical trials of stent retriever thrombectomy devices (entirely or predominantly) vs medical therapy. All 7 identified trials published in a peer-reviewed journal (by August 1, 2020) contributed data. Detailed time metrics were collected including last known well-to-door (LKWTD) time; last known well/onset-to-puncture (LKWTP) time; last known well-to-reperfusion (LKWR) time; door-to-puncture (DTP) time; and door-to-reperfusion (DTR) time. MAIN OUTCOMES AND MEASURES: Change in healthy life-years measured as disability-adjusted life-years (DALYs). DALYs were calculated as the sum of years of life lost (YLL) owing to premature mortality and years of healthy life lost because of disability (YLD). Disability weights were assigned using the utility-weighted modified Rankin Scale. Age-specific life expectancies without stroke were calculated from 2017 US National Vital Statistics. RESULTS: Among the 781 EVT-treated patients, 406 (52.0%) were early-treated (LKWTP ≤4 hours) and 375 (48.0%) were late-treated (LKWTP >4-12 hours). In early-treated patients, LKWTD was 188 minutes (interquartile range, 151.3-214.8 minutes) and DTP 105 minutes (interquartile range, 76-135 minutes). Among the 298 of 380 (78.4%) patients with substantial reperfusion, median DTR time was 145.0 minutes (interquartile range, 111.5-185.5 minutes). Care process delays were associated with worse clinical outcomes in LKW-to-intervention intervals in early-treated patients and in door-to-intervention intervals in early-treated and late-treated patients, and not associated with LKWTD intervals, eg, in early-treated patients, for each 10-minute delay, healthy life-years lost were DTP 1.8 months vs LKWTD 0.0 months; P < .001. Considering granular time increments, the amount of healthy life-time lost associated with each 1 second of delay was DTP 2.2 hours and DTR 2.4 hours. CONCLUSIONS AND RELEVANCE: In this study, care delays were associated with loss of healthy life-years in patients with acute ischemic stroke treated with EVT, particularly in the postarrival time period. The finding that every 1 second of delay was associated with loss of 2.2 hours of healthy life may encourage continuous quality improvement in door-to-treatment times.
  • Item
    Thumbnail Image
    The Association Between Sarcopenia and Functional Improvement in Older and Younger Patients Who Completed Inpatient Rehabilitation: A Prospective Cohort Study
    Churilov, I ; Churilov, L ; Brock, K ; Murphy, D ; MacIsaac, RJJ ; Ekinci, EII (FRONTIERS MEDIA SA, 2021-10-21)
    Objective: To investigate the association between sarcopenia and functional improvement in patients older and younger than 65 years upon completion of an inpatient rehabilitation program. Design: Prospective cohort study. Participants: Adult consecutive patients who completed the inpatient rehabilitation program at a metropolitan tertiary referral hospital general inpatient rehabilitation unit. Methods: Sarcopenia status was determined using the European Working Group on Sarcopenia in Older People 2 algorithm, using muscle mass measured by BioImpedance Analysis and grip strength. Progress in rehabilitation was measured using change in the Functional Independence Measure and Goal Attainment Scaling score. To investigate the age group by sarcopenia status interaction we used quantile regression models with bootstrapped standard error estimation for functional improvement and linear regression model with robust standard error estimation for GAS score. Results: 257 participants [128 (50%) male, median age 63 years (IQR: 52-72)], 33(13%) with sarcopenia, completed inpatient rehabilitation [median length of stay 16 days (IQR: 11-27.5)]. Participants' median Functional Independence Measure change was 24 (IQR 15-33.5) and mean total Goal Attainment Scaling score was 57.6 (SD 10.2). Adjusting for admission Functional Independence Measure score, the median difference in Functional Independence Measure change between participants with and without sarcopenia was: -4.3 (95% CI: -10.6, 1.9); p = 0.17 in participants 65 years and younger, and 4.6 (95% CI: 1.0, 8.2); p = 0.01 in participants older than 65; age-by-sarcopenia interaction p = 0.02. Conclusions: Unlike younger people, older people with sarcopenia have greater functional improvement in inpatient rehabilitation than those without sarcopenia.
  • Item
    Thumbnail Image
    Feasibility of using self-reported ethnicity in pregnancy according to the gestation-related optimal weight classification: a cross-sectional study
    Lockie, E ; McCarthy, EA ; Hui, L ; Churilov, L ; Walker, SP (WILEY, 2018-05)
    OBJECTIVE: To evaluate the feasibility of self-reported ethnicity using the gestation-related optimal growth (GROW) classification in a contemporary multicultural antenatal population. DESIGN: Cross-sectional study. SETTING: Tertiary obstetric hospital in Melbourne, Australia. POPULATION: Pregnant women attending the antenatal clinic. METHODS: We surveyed pregnant women during April-June 2016 regarding their understanding of the term 'ethnicity', and how they would classify the ethnicity of themselves, their partner, and family members according to the Australian GROW classification. RESULTS: Two hundred and thirty-five women completed the survey. When describing 'ethnicity', most women (103, 44%) chose multiple descriptors, most frequently country of birth (54%) and region of ancestry (47%). Interpretation of 'ethnicity' varied significantly between ethnic groups: those choosing 'country of birth' were more likely to identify as Indian (odds ratio, OR 3.5, P = 0.03), whereas those choosing 'physical appearance' were more likely to identify as Chinese (OR 3.0, P = 0.047). Thirty participants (13%) were unable to describe their ethnicity from the available GROW options. Sixty-one (26%) respondents' ethnicity was inconsistent with that of their parents' heritage. A further 35% had a partner of different ethnicity. The agreement between country of birth and self-reported ethnicity was only fair (kappa 0.73, 95% confidence interval, 95% CI 0.64-0.82). CONCLUSION: This study confirms the complexity of defining ethnicity in contemporary multicultural settings. Self-reported ethnicity is often inaccurate, concepts of ethnicity vary by ethnic group, and country of birth is a poor descriptive surrogate. Adjustment for maternal ethnicity should be undertaken with caution in the customised assessment of fetal growth. TWEETABLE ABSTRACT: Is self-reported maternal ethnicity reliable? We think not.
  • Item
    Thumbnail Image
    A novel immune function biomarker identifies patients at risk of clinical events early following liver transplantation
    Sood, S ; Haifer, C ; Yu, L ; Pavlovic, J ; Churilov, L ; Gow, PJ ; Jones, RM ; Angus, PW ; Visvanathan, K ; Testro, AG (WILEY, 2017-04)
    Balancing immunosuppression after liver transplant is difficult, with clinical events common. We investigate whether a novel immune biomarker based on a laboratory platform with widespread availability that measures interferon γ (IFNγ) after stimulation with a lyophilized ball containing an adaptive and innate immune stimulant can predict events following transplantation. A total of 75 adult transplant recipients were prospectively monitored in a blinded, observational study; 55/75 (73.3%) patients experienced a total of 89 clinical events. Most events occurred within the first month. Low week 1 results were significantly associated with risk of early infection (area under the receiver operating characteristic curve [AUROC], 0.74; P = 0.008). IFNγ ≤ 1.30 IU/mL (likelihood ratio positive, 1.93; sensitivity, 71.4%; specificity, 63.0%) was associated with the highest risk for infection with minimal rejection risk. Nearly half the cohort (27/60, 45.0%) expressed IFNγ ≤ 1.30 IU/mL. Moreover, an elevated week 1 result was significantly associated with the risk of rejection within the first month after transplant (AUROC, 0.77; P = 0.002), but no episodes of infection. On multivariate logistic regression, IFNγ ≥ 4.49 IU/mL (odds ratio, 4.75) may be an independent predictor of rejection (P = 0.05). In conclusion, low IFNγ suggesting oversuppression is associated with infections, whereas high IFNγ indicating undersuppression is associated with rejection. This assay offers the potential to allow individualization and optimization of immunosuppression that could fundamentally alter the way patients are managed following transplantation. Liver Transplantation 23 487-497 2017 AASLD.
  • Item
    Thumbnail Image
    Epilepsy in families: Age at onset is a familial trait, independent of syndrome
    Ellis, CA ; Churilov, L ; Epstein, MP ; Xie, SX ; Bellows, ST ; Ottman, R ; Berkovic, SF (WILEY, 2019-07)
    OBJECTIVE: We tested 2 hypotheses regarding age at onset within familial epilepsies: (1) family members with epilepsy tend to have similar ages at onset, independent of epilepsy syndrome; and (2) age at onset is younger in successive generations after controlling for sampling bias. METHODS: We analyzed clinical data collected by the Epi4K Consortium (303 multiplex families, 1,120 individuals). To test hypothesis 1, we used both linear mixed models commonly used for heritability analysis and Cox regression models with frailty terms to assess clustering of onset within families after controlling for other predictors. To test hypothesis 2, we used mixed effects models, pairwise analyses, and survival analysis to address sampling-related bias that may mimic anticipation. RESULTS: Regarding hypothesis 1, age at seizure onset was significantly heritable (intraclass correlation coefficient = 0.17, p < 0.001) after adjusting for epilepsy type, sex, site, history of febrile seizure, and age at last observation. This finding remained significant after adjusting for epilepsy syndromes, and was robust across statistical methods in all families and in generalized families. Regarding hypothesis 2, the mean age at onset decreased in successive generations (p < 0.001). After adjusting for age at last observation, this effect was not significant in mixed effects models (p = 0.14), but remained significant in pairwise (p = 0.0003) and survival analyses (p = 0.02). INTERPRETATION: Age at seizure onset is an independent familial trait, and may have genetic determinants distinct from the determinants of particular epilepsy syndromes. Younger onsets in successive generations can be explained in part by sampling bias, but the presence of genetic anticipation cannot be excluded. ANN NEUROL 2019.
  • Item
    Thumbnail Image
    Health economic implications of postoperative complications following liver resection surgery: a systematic review
    Cosic, L ; Ma, R ; Churilov, L ; Nikfarjam, M ; Christophi, C ; Weinberg, L (WILEY, 2019-12)
    BACKGROUND: Limited data exists concerning the health economics of liver resection, with even less information on the costs emerging from complications, despite this remaining an important target from a health economic perspective. Our objective was to describe the financial burden of complications following liver resection. METHODS: We conducted a systematic search and included studies reporting resource use of in-hospital complications during the index liver resection admission. All indications for liver resection were considered. All techniques were considered. Data was collected using a data extraction table and a narrative synthesis was performed. RESULTS: We identified 12 eligible articles. There was considerable heterogeneity in study designs, patient populations and outcome definitions. We found weak evidence of increased costs associated with major liver resection compared to minor resections. We found robust evidence supporting the increasing economic burden arising from complications after liver resection. Acceptable evidence for increased cost due to the presence and grade of complication was found. Strong evidence concerning the association of length of stay with costs was demonstrated. CONCLUSIONS: The presence and grade of complications increase hospital cost across diverse settings. The costing methodology should be transparent and complication grading systems should be consistent in future studies.
  • Item
    Thumbnail Image
    Timing of diagnosis of gestational diabetes and pregnancy outcomes: A retrospective cohort
    Shub, A ; Chee, T ; Templeton, A ; Boyce, D ; McNamara, C ; Houlihan, C ; Churilov, L ; McCarthy, EA (WILEY, 2019-02)
    BACKGROUND: Recent guidelines suggest screening high-risk women in early pregnancy for gestational diabetes (GDM); however, there is little evidence to support this. AIMS: To compare pregnancy outcomes associated with diabetes for women with risk factors for GDM according to gestation of diagnosis. Early GDM was defined as a positive test before 20 weeks gestation, late GDM as a positive test at 20 or more weeks and no GDM when both tests were negative. MATERIALS AND METHODS: Retrospective analysis in an Australian tertiary hospital of women who underwent a glucose tolerance test in pregnancy prior to 20 weeks gestation, and a repeat test after 20 weeks gestation if the initial test was negative. Results were adjusted for maternal demographics. RESULTS: Women with early GDM (n = 170) were no more likely to experience the obstetric composite outcome than women with late GDM (n = 171) or no GDM (n = 547) (early odds ratio (OR) 1.16, 95%CI 0.79-1.71; late OR 0.78, 95%CI 0.53-1.12). Infants of women with early GDM, but not late GDM, were more likely (early OR 1.8, 95%CI 1.15-2.92; late OR 1.4, 95%CI 0.90-2.23) to have the neonatal composite outcome than infants of women without GDM, predominantly due to an increase in neonatal hypoglycaemia. CONCLUSIONS: This result may be due to careful management of GDM, or because, after adjustment for maternal demographics, the early diagnosis of GDM does not substantially increase rates of adverse outcomes compared to GDM diagnosed in later pregnancy or no GDM in women with risk factors for GDM.
  • Item
    Thumbnail Image
    Limited utility of routine chest X-ray in initial evaluation of neutropenic fever in patients with haematological diseases undergoing chemotherapy
    Estacio, O ; Loh, Z ; Baker, A ; Chong, G ; Grigg, A ; Churilov, L ; Hawkes, EA (WILEY, 2018-05)
    BACKGROUND: Routine chest X-ray (CXR) is recommended for neutropenic fever (NF) management however its role is relatively understudied in haematology patients. AIM: To investigate the utility of CXR in the diagnosis and management of patients with haematological conditions complicated by NF. METHODS: Retrospective, single-centre analysis of haematology patients admitted with NF between January 2011 and December 2015. Baseline demographics, treatment details and outcomes were collected from electronic patient records. CXR underwent independent radiology review. Primary endpoints were a proportion of NF episodes in which CXR detected a probable chest infection in the absence of respiratory symptoms/signs and/or resulted in a change in antibiotic management. RESULTS: Four hundred and thirty-five episodes were identified; CXR was performed in 75% of patients (65% within 2 days of NF). In 4 of 164 (2.4%) asymptomatic patients, CXR was consistent with infection, in contrast to 19 of 119 (16%) patients with clinical signs of respiratory infection. Only 3 of 283 (1.1%) CXR resulted in a change to antibiotics. CXR consistent with infection was not associated with increased mortality or increased admission length, although there was an association with intensive care unit admission (odds ratios: 7.61, 95% confidence interval: 2.04-28.31). CONCLUSION: In haematology patients with NF, CXR rarely detected chest infection or changed management in patients with no respiratory symptoms or signs. CXR in our institution is no longer part of routine assessment of NF in the absence of these features.
  • Item
    No Preview Available
    Aeromedical retrieval diagnostic trends during a period of Coronavirus 2019 lockdown
    Gardiner, FW ; Gillam, M ; Churilov, L ; Sharma, P ; Steere, M ; Hanna, M ; Hooper, A ; Quinlan, F (Cold Spring Harbor Laboratory, 2020-08-19)
    Abstract Background We aimed to compare the pre, lockdown, and post-lockdown aeromedical retrieval (AR) diagnostic reasons and patient demographics during a period of Coronavirus 2019 (COVID-19) social isolation. Methods An observational study with retrospective data collection, consisting of Australians who received an AR between the 26 January to the 23 June 2020. The main outcome measures were patient diagnostic category proportions and trends prior (28 January to 15 March), during (16 March to 4 May), and following (5 May to 23 June 2020) social isolation restrictions. Results There were 16981 ARs consisting of 1959 (11.5) primary evacuations (PE) and 12724 (88.5) inter-hospital transfer (IHT), with a population median age of 52 years old (interquartile range [IQR] 29.0–69.0), with 49.0% (n = 8283) of the cohort being male and 38.0% (n = 6399) being female. There were a total of 6 confirmed and 209 suspected cases of COVID-19, with the majority of cases (n = 114; 53.0%) in the social isolation period. As compared to pre-restriction, the odds of retrieval for the restriction and post-restriction period differed across time between the major diagnostic groups. This included, an increase in cardiovascular retrieval for both restriction and post-restriction periods (OR 1.12 95% CI 1.02-1.24 and OR 1.18 95% CI 1.08-1.30 respectively), increases in neoplasm in the post restriction period (OR 1.31 95% CI 1.04-1.64), and increases for congenital conditions in the restriction period (OR 2.56 95% CI 1.39-4.71). Cardiovascular and congenital conditions had increased rates of priority 1 patients in the restriction and post restriction periods. There was a decrease in endocrine and metabolic disease retrievals in the restriction period (OR 0.72 95% CI 0.53-0.98). There were lower odds during the post-restriction period for a retrievals of the respiratory system (OR 0.78 95% CI 0.67-0.93), and disease of the skin (OR 0.78 95% CI 0.6-1.0). Distribution between the 2019 and 2020 time periods differed (p< 0.05), with the lockdown period resulting in a significant reduction in activity. Conclusion The lockdown period resulted in increased AR rates of circulatory and congenital conditions. However, this period also resulted in a reduction of overall activity, possibly due to a reduction in other infectious disease rates, such as influenza, due to social distancing.