In this study, we propose an approach that provides a useful data summary related to a patient’s experience of pain. Because pain is a very important but subjective phenomenon that currently has no ...calibratable method for assessing it, we suggest an approach that uses calibratable biomarker sensors with the patient’s self-assessment of perceived pain. We surmise that such an approach may only be able to clearly distinguish between cases in which the available evidence is consistent. However, this information may provide clinicians with valuable insights, and as research progresses into how biomarkers are related to pain, more specific insights may emerge regarding how specific evidence inconsistencies may point to particular pain causes. We provide a brief overview of pain science, including the types of pain, contemporary pain theories, pain, and pain assessment techniques. Next, we present novel approaches to pain sensor development, including an overview of research on pain-related biomarker sensors and artificial intelligence methods for summarizing the evidence. We then provide some illustrations of the implementation of our approach. Some specifics are presented in the Methods section of this paper. For example, in a set of 379 patients, we observed 80% evidence of consistency and 5 types of inconsistencies. Information regarding the gender and individual differences in cyclooxygenase-2 and inducible nitric oxide synthase data on reported pain could contribute to the inconsistency. Different causes of inconsistencies are also attributed to cultural or temporal variability of cyclooxygenase-2 and inducible nitric oxide synthase (as well as their serum variation and half-life), visual analog scale, and other tools. We emphasize that this presentation is illustrative. Much work remains to be done before implementing and testing this approach in a clinically meaningful context.
When investigating covariate interactions and group associations with standard regression analyses, the relationship between the response variable and exposure may be difficult to characterize. When ...the relationship is nonlinear, linear modeling techniques do not capture the nonlinear information content. Statistical learning (SL) techniques with kernels are capable of addressing nonlinear problems without making parametric assumptions. However, these techniques do not produce findings relevant for epidemiologic interpretations. A simulated case-control study was used to contrast the information embedding characteristics and separation boundaries produced by a specific SL technique with logistic regression (LR) modeling representing a parametric approach. The SL technique was comprised of a kernel mapping in combination with a perceptron neural network. Because the LR model has an important epidemiologic interpretation, the SL method was modified to produce the analogous interpretation and generate odds ratios for comparison.
The SL approach is capable of generating odds ratios for main effects and risk factor interactions that better capture nonlinear relationships between exposure variables and outcome in comparison with LR.
The integration of SL methods in epidemiology may improve both the understanding and interpretation of complex exposure/disease relationships.
After the September 11 tragedies of 2001, scientists and law‐enforcement agencies have shown increasing concern that terrorist organizations and their “rogue” foreign government‐backers may resort to ...the use of chemical and/or biological agents against U.S. military or civilian targets. In addition to the right mix of policies, including security measures, intelligence gathering and training for medical personnel on how to recognize symptoms of biochemical warfare agents, the major success in combating terrorism lies in how best to respond to an attack using reliable analytical sensors. The public and regulatory agencies expect sensing methodologies and devices for homeland security to be very reliable. Quality data can only be generated by using analytical sensors that are validated and proven to be under strict design criteria, development and manufacturing controls. Electrochemical devices are ideally suited for obtaining the desired analytical information in a faster, simpler, and cheaper manner compared to traditional (lab‐based) assays and hence for meeting the requirements of decentralized biodefense applications. This articler presents a review of the major trends in monitoring technologies for chemical and biological warfare (CBW) agents. It focuses on research and development of sensors (particularly electrochemical ones), discusses how advances in molecular recognition might be used to design new multimission networked sensors (MULNETS) for homeland security. Decision flow‐charts for choosing particular analytical techniques for CBW agents are presented. Finally, the paths to designing sensors to meet the needs of today's measurement criteria are analyzed.
Significant interest exists in establishing radiologic imaging as a valid biomarker for assessing the response of cancer to a variety of treatments. To address this problem, we have chosen to study ...patients with metastatic colorectal carcinoma to learn whether statistical learning theory can improve the performance of radiologists using CT in predicting patient treatment response to therapy compared with the more traditional RECIST (Response Evaluation Criteria in Solid Tumors) standard.
Predictions of survival after 8 months in 38 patients with metastatic colorectal carcinoma using the Support Vector Machine (SVM) technique improved 30% when using additional information compared to WHO (World Health Organization) or RECIST measurements alone. With both Logistic Regression (LR) and SVM, there was no significant difference in performance between WHO and RECIST. The SVM and LR techniques also demonstrated that one radiologist consistently outperformed another.
This preliminary research study has demonstrated that SLT algorithms, properly used in a clinical setting, have the potential to address questions and criticisms associated with both RECIST and WHO scoring methods. We also propose that tumor heterogeneity, shape, etc. obtained from CT and/or MRI scans be added to the SLT feature vector for processing.
Statistical learning (SL) techniques can address non-linear relationships and small datasets but do not provide an output that has an epidemiologic interpretation.
A small set of clinical variables ...(CVs) for stage-1 non-small cell lung cancer patients was used to evaluate an approach for using SL methods as a preprocessing step for survival analysis. A stochastic method of training a probabilistic neural network (PNN) was used with differential evolution (DE) optimization. Survival scores were derived stochastically by combining CVs with the PNN. Patients (n = 151) were dichotomized into favorable (n = 92) and unfavorable (n = 59) survival outcome groups. These PNN derived scores were used with logistic regression (LR) modeling to predict favorable survival outcome and were integrated into the survival analysis (i.e. Kaplan-Meier analysis and Cox regression). The hybrid modeling was compared with the respective modeling using raw CVs. The area under the receiver operating characteristic curve (Az) was used to compare model predictive capability. Odds ratios (ORs) and hazard ratios (HRs) were used to compare disease associations with 95% confidence intervals (CIs).
The LR model with the best predictive capability gave Az = 0.703. While controlling for gender and tumor grade, the OR = 0.63 (CI: 0.43, 0.91) per standard deviation (SD) increase in age indicates increasing age confers unfavorable outcome. The hybrid LR model gave Az = 0.778 by combining age and tumor grade with the PNN and controlling for gender. The PNN score and age translate inversely with respect to risk. The OR = 0.27 (CI: 0.14, 0.53) per SD increase in PNN score indicates those patients with decreased score confer unfavorable outcome. The tumor grade adjusted hazard for patients above the median age compared with those below the median was HR = 1.78 (CI: 1.06, 3.02), whereas the hazard for those patients below the median PNN score compared to those above the median was HR = 4.0 (CI: 2.13, 7.14).
We have provided preliminary evidence showing that the SL preprocessing may provide benefits in comparison with accepted approaches. The work will require further evaluation with varying datasets to confirm these findings.
The primary objectives of this paper are: 1.) to apply Statistical Learning Theory (SLT), specifically Partial Least Squares (PLS) and Kernelized PLS (K-PLS), to the universal ..."feature-rich/case-poor" (also known as "large p small n", or "high-dimension, low-sample size") microarray problem by eliminating those features (or probes) that do not contribute to the "best" chromosome bio-markers for lung cancer, and 2.) quantitatively measure and verify (by an independent means) the efficacy of this PLS process. A secondary objective is to integrate these significant improvements in diagnostic and prognostic biomedical applications into the clinical research arena. That is, to devise a framework for converting SLT results into direct, useful clinical information for patient care or pharmaceutical research. We, therefore, propose and preliminarily evaluate, a process whereby PLS, K-PLS, and Support Vector Machines (SVM) may be integrated with the accepted and well understood traditional biostatistical "gold standard", Cox Proportional Hazard model and Kaplan-Meier survival analysis methods. Specifically, this new combination will be illustrated with both PLS and Kaplan-Meier followed by PLS and Cox Hazard Ratios (CHR) and can be easily extended for both the K-PLS and SVM paradigms. Finally, these previously described processes are contained in the Fine Feature Selection (FFS) component of our overall feature reduction/evaluation process, which consists of the following components: 1.) coarse feature reduction, 2.) fine feature selection and 3.) classification (as described in this paper) and prediction.
Our results for PLS and K-PLS showed that these techniques, as part of our overall feature reduction process, performed well on noisy microarray data. The best performance was a good 0.794 Area Under a Receiver Operating Characteristic (ROC) Curve (AUC) for classification of recurrence prior to or after 36 months and a strong 0.869 AUC for classification of recurrence prior to or after 60 months. Kaplan-Meier curves for the classification groups were clearly separated, with p-values below 4.5e-12 for both 36 and 60 months. CHRs were also good, with ratios of 2.846341 (36 months) and 3.996732 (60 months).
SLT techniques such as PLS and K-PLS can effectively address difficult problems with analyzing biomedical data such as microarrays. The combinations with established biostatistical techniques demonstrated in this paper allow these methods to move from academic research and into clinical practice.
Analysis of gene expression microarray datasets presents the high risk of over-fitting (spurious patterns) because of their feature-rich but case-poor nature. This paper describes our ongoing efforts ...to develop a method to combat over-fitting and determine the strongest signal in the dataset. A GA-SVM hybrid along with Gaussian noise (manual noise gain) is used to discover feature sets of minimal size that accurately classifies the cases under cross-validation. Initial results on a colorectal cancer dataset shows that the strongest signal (modest number of candidates) can be found by a binary search.
Parenchymal patterns defining the density of breast tissue are detected by advanced correlation pattern recognition in an integrated Computer-Aided Detection (CAD) and diagnosis system. Fractal ...signatures of density are modelled according to four clinical categories. A Support Vector Machine (SVM) in the primal formulation solves the multiclass problem using 'One-Versus-All' (OVA) and 'All-Versus-All' (AVA) decompositions, achieving 85% and 94% accuracy, respectively. Fully automated classification of breast density via a texture model derived from fractal dimension, dispersion, and lacunarity moves current qualitative methods forward to objective quantitative measures, amenable with the overarching vision of substantiating the role of density in epidemiological risk models of breast cancer.
The need for rapid and accurate detection systems is expanding and the utilization of cross-reactive sensor arrays to detect chemical warfare agents in conjunction with novel computational techniques ...may prove to be a potential solution to this challenge. We have investigated the detection, prediction, and classification of various organophosphate (OP) nerve agent simulants using sensor arrays with a novel learning scheme known as support vector machines (SVMs). The OPs tested include parathion, malathion, dichlorvos, trichlorfon, paraoxon, and diazinon. A new data reduction software program was written in MATLAB V. 6.1 to extract steady-state and kinetic data from the sensor arrays. The program also creates training sets by mixing and randomly sorting any combination of data categories into both positive and negative cases. The resulting signals were fed into SVM software for “pairwise” and “one” vs all classification. Experimental results for this new paradigm show a significant increase in classification accuracy when compared to artificial neural networks (ANNs). Three kernels, the S2000, the polynomial, and the Gaussian radial basis function (RBF), were tested and compared to the ANN. The following measures of performance were considered in the pairwise classification: receiver operating curve (ROC) A z indices, specificities, and positive predictive values (PPVs). The ROC A z values, specifities, and PPVs increases ranged from 5% to 25%, 108% to 204%, and 13% to 54%, respectively, in all OP pairs studied when compared to the ANN baseline. Dichlorvos, trichlorfon, and paraoxon were perfectly predicted. Positive prediction for malathion was 95%.
To establish radiologic imaging as a valid biomarker for assessing the response of cancer to different treatments. We study patients with metastatic colorectal carcinoma to learn whether Statistical ...Learning Theory (SLT) improves the performance of radiologists using Computer Tomography (CT) in predicting patient treatment response to therapy compared with traditional Response Evaluation Criteria in Solid Tumours (RECIST) standard. Preliminary research demonstrated that SLT algorithms can address questions and criticisms associated with both RECIST and World Health Organization (WHO) scoring methods. We add tumour heterogeneity, shape, etc., obtained from CT or MRI scans the feature vector for processing.