Type 2 diabetes leads to premature death and reduced quality of life for 8% of Americans. Nutrition management is critical to maintaining glycemic control, yet it is difficult to achieve due to the ...high individual differences in glycemic response to nutrition. Anticipating glycemic impact of different meals can be challenging not only for individuals with diabetes, but also for expert diabetes educators. Personalized computational models that can accurately forecast an impact of a given meal on an individual's blood glucose levels can serve as the engine for a new generation of decision support tools for individuals with diabetes. However, to be useful in practice, these computational engines need to generate accurate forecasts based on limited datasets consistent with typical self-monitoring practices of individuals with type 2 diabetes. This paper uses three forecasting machines: (i) data assimilation, a technique borrowed from atmospheric physics and engineering that uses Bayesian modeling to infuse data with human knowledge represented in a mechanistic model, to generate real-time, personalized, adaptable glucose forecasts; (ii) model averaging of data assimilation output; and (iii) dynamical Gaussian process model regression. The proposed data assimilation machine, the primary focus of the paper, uses a modified dual unscented Kalman filter to estimate states and parameters, personalizing the mechanistic models. Model selection is used to make a personalized model selection for the individual and their measurement characteristics. The data assimilation forecasts are empirically evaluated against actual postprandial glucose measurements captured by individuals with type 2 diabetes, and against predictions generated by experienced diabetes educators after reviewing a set of historical nutritional records and glucose measurements for the same individual. The evaluation suggests that the data assimilation forecasts compare well with specific glucose measurements and match or exceed in accuracy expert forecasts. We conclude by examining ways to present predictions as forecast-derived range quantities and evaluate the comparative advantages of these ranges.
The national adoption of electronic health records (EHR) promises to make an unprecedented amount of data available for clinical research, but the data are complex, inaccurate, and frequently ...missing, and the record reflects complex processes aside from the patient's physiological state. We believe that the path forward requires studying the EHR as an object of interest in itself, and that new models, learning from data, and collaboration will lead to efficient use of the valuable information currently locked in health records.
Abstract
Electronic health record phenotyping is the use of raw electronic health record data to assert characterizations about patients. Researchers have been doing it since the beginning of ...biomedical informatics, under different names. Phenotyping will benefit from an increasing focus on fidelity, both in the sense of increasing richness, such as measured levels, degree or severity, timing, probability, or conceptual relationships, and in the sense of reducing bias. Research agendas should shift from merely improving binary assignment to studying and improving richer representations. The field is actively researching new temporal directions and abstract representations, including deep learning. The field would benefit from research in nonlinear dynamics, in combining mechanistic models with empirical data, including data assimilation, and in topology. The health care process produces substantial bias, and studying that bias explicitly rather than treating it as merely another source of noise would facilitate addressing it.
Background Fields like nonlinear physics offer methods for analyzing time series, but many methods require that the time series be stationary—no change in properties over time.
Objective Medicine is ...far from stationary, but the challenge may be able to be ameliorated by reparameterizing time because clinicians tend to measure patients more frequently when they are ill and are more likely to vary.
Methods We compared time parameterizations, measuring variability of rate of change and magnitude of change, and looking for homogeneity of bins of temporal separation between pairs of time points. We studied four common laboratory tests drawn from 25 years of electronic health records on 4 million patients.
Results We found that sequence time—that is, simply counting the number of measurements from some start—produced more stationary time series, better explained the variation in values, and had more homogeneous bins than either traditional clock time or a recently proposed intermediate parameterization. Sequence time produced more accurate predictions in a single Gaussian process model experiment.
Conclusions Of the three parameterizations, sequence time appeared to produce the most stationary series, possibly because clinicians adjust their sampling to the acuity of the patient. Parameterizing by sequence time may be applicable to association and clustering experiments on electronic health record data. A limitation of this study is that laboratory data were derived from only one institution. Sequence time appears to be an important potential parameterization.
To study the relation between electronic health record (EHR) variables and healthcare process events.
Lagged linear correlation was calculated between five healthcare process events and 84 EHR ...variables (24 clinical laboratory values and 60 clinical concepts extracted from clinical notes) in a 24-year database. The EHR variables were clustered for each healthcare process event and interpreted.
Laboratory tests tended to cluster together and note concepts tended to cluster together. Within each of those two classes, the variables clustered into clinically sensible groupings. The exact groupings varied from healthcare process event to event, with the largest differences occurring between inpatient events and outpatient events.
Unlike previously reported pairwise associations between variables, which highlighted correlations across the laboratory-clinical note divide, incorporating healthcare process events appeared to be sensitive to the manner in which the variables were collected.
We believe that it may be possible to exploit this sensitivity to help knowledge engineers select variables and correct for biases.
We introduce data assimilation as a computational method that uses machine learning to combine data with human knowledge in the form of mechanistic models in order to forecast future states, to ...impute missing data from the past by smoothing, and to infer measurable and unmeasurable quantities that represent clinically and scientifically important phenotypes. We demonstrate the advantages it affords in the context of type 2 diabetes by showing how data assimilation can be used to forecast future glucose values, to impute previously missing glucose values, and to infer type 2 diabetes phenotypes. At the heart of data assimilation is the mechanistic model, here an endocrine model. Such models can vary in complexity, contain testable hypotheses about important mechanics that govern the system (eg, nutrition's effect on glucose), and, as such, constrain the model space, allowing for accurate estimation using very little data.
Ventilator dyssynchrony may be associated with increased delivered tidal volumes (V t s) and dynamic transpulmonary pressure (ΔP L,dyn ), surrogate markers of lung stress and strain, despite low V t ...ventilation. However, it is unknown which types of ventilator dyssynchrony are most likely to increase these metrics or if specific ventilation or sedation strategies can mitigate this potential.
A prospective cohort analysis to delineate the association between ten types of breaths and delivered V t , ΔP L,dyn , and transpulmonary mechanical energy.
Patients admitted to the medical ICU.
Over 580,000 breaths from 35 patients with acute respiratory distress syndrome (ARDS) or ARDS risk factors.
None.
Patients received continuous esophageal manometry. Ventilator dyssynchrony was identified using a machine learning algorithm. Mixed-effect models predicted V t , ΔP L,dyn , and transpulmonary mechanical energy for each type of ventilator dyssynchrony while controlling for repeated measures. Finally, we described how V t , positive end-expiratory pressure (PEEP), and sedation (Richmond Agitation-Sedation Scale) strategies modify ventilator dyssynchrony's association with these surrogate markers of lung stress and strain. Double-triggered breaths were associated with the most significant increase in V t , ΔP L,dyn , and transpulmonary mechanical energy. However, flow-limited, early reverse-triggered, and early ventilator-terminated breaths were also associated with significant increases in V t , ΔP L,dyn , and energy. The potential of a ventilator dyssynchrony type to increase V t , ΔP L,dyn , or energy clustered similarly. Increasing set V t may be associated with a disproportionate increase in high-volume and high-energy ventilation from double-triggered breaths, but PEEP and sedation do not clinically modify the interaction between ventilator dyssynchrony and surrogate markers of lung stress and strain.
Double-triggered, flow-limited, early reverse-triggered, and early ventilator-terminated breaths are associated with increases in V t , ΔP L,dyn , and energy. As flow-limited breaths are more than twice as common as double-triggered breaths, further work is needed to determine the interaction of ventilator dyssynchrony frequency to cause clinically meaningful changes in patient outcomes.
Current international guidelines recommend endoscopic resection for T1 colorectal cancer (CRC) with low-risk histology features and oncologic resection for those at high risk of lymphatic metastasis. ...Exact risk stratification is therefore crucial to avoid under-treatment as well as over-treatment. Endoscopic full-thickness resection (EFTR) has shown to be effective for treatment of non-lifting benign lesions. In this multicenter, retrospective study we aimed to evaluate efficacy, safety, and clinical value of EFTR for early CRC.
Records of 1234 patients undergoing EFTR for various indications at 96 centers were screened for eligibility. A total of 156 patients with histologic evidence of adenocarcinoma were identified. This cohort included 64 cases undergoing EFTR after incomplete resection of a malignant polyp (group 1) and 92 non-lifting lesions (group 2). Endpoints of the study were: technical success, R0-resection, adverse events, and successful discrimination of high-risk versus low-risk tumors.
Technical success was achieved in 144 out of 156 (92.3%). Mean procedural time was 42 minutes. R0 resection was achieved in 112 of 156 (71.8%). Subgroup analysis showed a R0 resection rate of 87.5% in Group 1 and 60.9% in Group 2 (P < .001). Severe procedure-related adverse events were recorded in 3.9% of patients. Discrimination between high-risk versus low-risk tumor was successful in 155 of 156 cases (99.3%). In Group 1, 84.1% were identified as low-risk lesions, whereas 16.3% in group 2 had low-risk features. In total, 53 patients (34%) underwent oncologic resection due to high-risk features whereas 98 patients (62%) were followed endoscopically.
In early colorectal cancer, EFTR is technically feasible and safe. It allows exact histological risk stratification and can avoid surgery for low-risk lesions. Prospective studies are required to further define indications for EFTR in malignant colorectal lesions and to evaluate long-term outcome.
Display omitted