This book constitutes a collection of research achievements mature enough to provide a firm and reliable basis on modular ontologies. It discusses the recent concepts, theories and techniques for ...knowledge modularization.
Substantial variability exists in the use of life-prolonging treatments for patients with stroke, especially near the end of life. This study explores patterns of palliative care utilization and ...death in hospitalized patients with stroke across the United States.
Using the 2010 to 2012 nationwide inpatient sample databases, we included all patients discharged with stroke identified by
codes. Strokes were subclassified as ischemic, intracerebral, and subarachnoid hemorrhage. We compared demographics, comorbidities, procedures, and outcomes between patients with and without a palliative care encounter (PCE) as defined by the
code V66.7. Pearson χ
test was used for categorical variables. Multivariate logistic regression was used to account for hospital, regional, payer, and medical severity factors to predict PCE use and death.
Among 395 411 patients with stroke, PCE was used in 6.2% with an increasing trend over time (
<0.05). We found a wide range in PCE use with higher rates in patients with older age, hemorrhagic stroke types, women, and white race (all
<0.001). Smaller and for-profit hospitals saw lower rates. Overall, 9.2% of hospitalized patients with stroke died, and PCE was significantly associated with death. Length of stay in decedents was shorter for patients who received PCE.
Palliative care use is increasing nationally for patients with stroke, especially in larger hospitals. Persistent disparities in PCE use and mortality exist in regards to age, sex, race, region, and hospital characteristics. Given the variations in PCE use, especially at the end of life, the use of mortality rates as a hospital quality measure is questioned.
With the accumulation of large amounts of health related data, predictive analytics could stimulate the transformation of reactive medicine towards Predictive, Preventive and Personalized (PPPM) ...Medicine, ultimately affecting both cost and quality of care. However, high-dimensionality and high-complexity of the data involved, prevents data-driven methods from easy translation into clinically relevant models. Additionally, the application of cutting edge predictive methods and data manipulation require substantial programming skills, limiting its direct exploitation by medical domain experts. This leaves a gap between potential and actual data usage. In this study, the authors address this problem by focusing on open, visual environments, suited to be applied by the medical community. Moreover, we review code free applications of big data technologies. As a showcase, a framework was developed for the meaningful use of data from critical care patients by integrating the MIMIC-II database in a data mining environment (RapidMiner) supporting scalable predictive analytics using visual tools (RapidMiner's Radoop extension). Guided by the CRoss-Industry Standard Process for Data Mining (CRISP-DM), the ETL process (Extract, Transform, Load) was initiated by retrieving data from the MIMIC-II tables of interest. As use case, correlation of platelet count and ICU survival was quantitatively assessed. Using visual tools for ETL on Hadoop and predictive modeling in RapidMiner, we developed robust processes for automatic building, parameter optimization and evaluation of various predictive models, under different feature selection schemes. Because these processes can be easily adopted in other projects, this environment is attractive for scalable predictive analytics in health research.
In this edition, we focused on big data analytics methods and tools that have been specifically developed for the domain of cultural heritage, as well as on experiences from the adaptation and/or ...application of general-purpose solutions in the domain of cultural heritage. The aim was to gather solutions, but also to summarise the lessons learnt, methodologies, and good practices that researchers and practitioners can use as a basis for their own work in the domain.
Time Series Management Systems: A Survey Jensen, Søren Kejser; Pedersen, Torben Bach; Thomsen, Christian
IEEE transactions on knowledge and data engineering,
11/2017, Letnik:
29, Številka:
11
Journal Article
Recenzirano
Odprti dostop
The collection of time series data increases as more monitoring and automation are being deployed. These deployments range in scale from an Internet of things (IoT) device located in a household to ...enormous distributed Cyber-Physical Systems (CPSs) producing large volumes of data at high velocity. To store and analyze these vast amounts of data, specialized Time Series Management Systems (TSMSs) have been developed to overcome the limitations of general purpose Database Management Systems (DBMSs) for times series management. In this paper, we present a thorough analysis and classification of TSMSs developed through academic or industrial research and documented through publications. Our classification is organized into categories based on the architectures observed during our analysis. In addition, we provide an overview of each system with a focus on the motivational use case that drove the development of the system, the functionality for storage and querying of time series a system implements, the components the system is composed of, and the capabilities of each system with regard to Stream Processing and Approximate Query Processing (AQP). Last, we provide a summary of research directions proposed by other researchers in the field and present our vision for a next generation TSMS.
We live in an increasingly interconnected world, with many organizations operating across countries or even continents. To serve their global user base, organizations are replacing their legacy DBMSs ...with cloud-based systems capable of scaling OLTP workloads to millions of users. CockroachDB is a scalable SQL DBMS that was built from the ground up to support these global OLTP workloads while maintaining high availability and strong consistency. Just like its namesake, CockroachDB is resilient to disasters through replication and automatic recovery mechanisms. This paper presents the design of CockroachDB and its novel transaction model that supports consistent geo-distributed transactions on commodity hardware. We describe how CockroachDB replicates and distributes data to achieve fault tolerance and high performance, as well as how its distributed SQL layer automatically scales with the size of the database cluster while providing the standard SQL interface that users expect. Finally, we present a comprehensive performance evaluation and share a couple of case studies of CockroachDB users. We conclude by describing lessons learned while building CockroachDB over the last five years.
As the primary approach to deriving decision-support insights, automated recurring routine analytic jobs account for a major part of cluster resource usages in modern enterprise data warehouses. ...These recurring routine jobs usually have stringent schedule and deadline determined by external business logic, and thus cause dreadful resource skew and severe resource over-provision in the cluster. In this paper, we present Grosbeak, a novel data warehouse that supports resource-aware incremental computing to process recurring routine jobs, smooths the resource skew, and optimizes the resource usage. Unlike batch processing in traditional data warehouses, Grosbeak leverages the fact that data is continuously ingested. It breaks an analysis job into small batches that incrementally process the progressively available data, and schedules these small-batch jobs intelligently when the cluster has free resources. In this demonstration, we showcase Grosbeak using real-world analysis pipelines. Users can interact with the data warehouse by registering recurring queries and observing the incremental scheduling behavior and smoothed resource usage pattern.
This open access book comprehensively covers the fundamentals of clinical data science, focusing on data collection, modelling and clinical applications. Topics covered in the first section on data ...collection include: data sources, data at scale (big data), data stewardship (FAIR data) and related privacy concerns. Aspects of predictive modelling using techniques such as classification, regression or clustering, and prediction model validation will be covered in the second section. The third section covers aspects of (mobile) clinical decision support systems, operational excellence and value-based healthcare. Fundamentals of Clinical Data Science is an essential resource for healthcare professionals and IT consultants intending to develop and refine their skills in personalized medicine, using solutions based on large datasets from electronic health records or telemonitoring programmes. The book’s promise is “no math, no code”and will explain the topics in a style that is optimized for a healthcare audience.