Hyperspectral image (HSI) usually holds information of land cover classes as a set of many contiguous narrow spectral wavelength bands. For its efficient thematic mapping or classification, band ...(feature) reduction strategies through Feature Extraction (FE) and/or Feature Selection (FS) methods for finding the intrinsic bands' information are typically applied. Principal Component Analysis (PCA) is a frequently employed unsupervised linear FE method whereas cumulative-variance accumulation is used for selecting top features from PCA data. However, PCA can fail to extract intrinsic structure of HSI due to global variance consideration and domination by visible and near infrared bands while cumulative-variance accumulation has no capability to exploit non-linear relationships among the transformed features produced by PCA-based FE methods. Consequently, we propose an information theoretic normalized Mutual Information (nMI)-based minimum Redundancy Maximum Relevance (mRMR) non-linear measure to select the intrinsic features from the transformed space of our previously proposed Segmented-Folded-PCA (Seg-Fol-PCA) and Spectrally Segmented-Folded-PCA (SSeg-Fol-PCA) FE methods. We extensively analyse the effectiveness of the proposed unsupervised FE and supervised FS combinations Seg-Fol-PCA-mRMR and SSeg-Fol-PCA-mRMR with that of PCA-based existing linear and non-linear state-of-the-art methods. In addition, cumulative variance-based top features pick-up strategy is considered with all FE methods and Renyi quadratic entropy-based FS is used with Kernel Entropy Component Analysis (Ker-ECA). The experimental results illustrate that SSeg-Fol-PCA-mRMR and Seg-Fol-PCA-mRMR obtain highest classification result e.g. 95.39% and 95.03% respectively for agricultural Indian Pines HSI, and 96.58% and 95.30% respectively for urban Washington DC Mall HSI while the classification accuracies using all original features of the HSIs are 70.28% and 91.90% respectively. Moreover, the proposed SSeg-Fol-PCA-mRMR and Seg-Fol-PCA-mRMR outperform all investigated combinations of FE and FS using the real HSI datasets.
•This study examines elderly's intention to use wearable healthcare technologies.•The paper has extended UTAUT2 using resistance to change, technology anxiety, and self-actualization.•A two stage ...SEM-Neural Network approach were applied to analyze the data.•Policymakers should focus on functional congruence, technology anxiety, and hedonic motivation.
Wearable healthcare technology (WHT) has the potential to improve access to healthcare information especially to the older population and empower them to play an active role in self-management of their health. Despite their potential benefits, the acceptance and usage of WHT among the elderly are considerably low. However, little research has been conducted to describe any systematic study of the elderly's intention to adopt WHT. The objective of this study was to develop a theoretical model on the basis of extended Unified Theory of Acceptance and Use of Technology (UTAUT2) with additional constructs- resistance to change, technology anxiety, and self-actualization, to investigate the key predictors of WHT adoption by elderly. The model used in the current study was analyzed in two steps. In the first step, a Structural Equation Modeling (SEM) was used to determine significant determinants that affect the adoption of WHT. In the second step, a neural network model was applied to validate the findings in step 1 and establish the relative importance of each determinant to the adoption of WHT. The findings revealed that social influence, performance expectancy, functional congruence, self-actualization, and hedonic motivation had a positive relationship with the adoption of WHT. In addition, technology anxiety and resistance to change posed important but negative influences on WHT acceptance. Surprisingly, the study did not find any significant relationship between effort expectancy and facilitating conditions with behavioral intention to use WHT by the elderly. The results of this research have strong theoretical contributions to the existing literature of WHT. It also provides valuable information for WHT developers and social planners in the design and execution of WHT for the elderly.
The remote sensing hyperspectral images (HSIs) usually comprise many important information of the land covers capturing through a set of hundreds of narrow and contiguous spectral wavelength bands. ...Appropriate classification performance can only offer the required knowledge from these immense bands of HSI since the classification result is not reasonable using all the original features (bands) of the HSI. Although it is not easy to calculate the intrinsic features from the bands, band (dimensionality) reduction techniques through feature extraction and feature selection are usually applied to increase the classification result and to fix the curse of dimensionality problem. Though the Principal Component Analysis (PCA) has been commonly adopted for the feature reduction of HSI, it can often fail to extract the local useful characteristics of the HSI for effective classification as it considers the global statistics of the HSI. Consequently, Segmented-PCA (SPCA), Spectrally-Segmented-PCA (SSPCA), Folded-PCA (FPCA) and Superpixelwise PCA (SuperPCA) have been introduced for better feature extraction of HSI in diverse ways. In this paper, feature extraction through SPCA & FPCA and SSPCA & FPCA, termed as Segmented-FPCA (SFPCA) and Spectrally-Segmented-FPCA (SSFPCA) respectively, has further been improved through applying FPCA on the highly correlated or spectrally separated bands' segments of the HSI rather than not applying the FPCA on the entire dataset directly. The proposed methods are compared and analysed for a real mixed agricultural and an urban HSI classification using per-pixel SVM classifier. The experimental result shows that the classification performance using SSFPCA and SFPCA outperforms that of using conventional PCA, SPCA, SSPCA, FPCA, SuperPCA and using the entire original dataset without employing any feature reduction. Moreover, the proposed feature extraction methods provide the least memory and computation cost complexity.
The hyperspectral remote sensing images (HSIs) are acquired to encompass the essential information of land objects through contiguous narrow spectral wavelength bands. The classification accuracy is ...not often satisfactory in a cost-effective way using the entire original HSI for practical applications. To enhance the classification result of HSIs the band reduction strategies are applied which can be divided into feature extraction and feature selection methods. PCA (Principal Component Analysis), a linear unsupervised statistical transformation, is frequently adopted for the extraction of features from HSIs. In this paper, PCA and SPCA (Segmented-PCA), SSPCA (Spectrally Segmented-PCA), FPCA (Folded-PCA) and MNF (Minimum Noise Fraction) as linear variants of PCA together with KPCA (Kernel-PCA) and KECA (kernel Entropy Component Analysis) as nonlinear variants of PCA have been investigated. The top transformed features were picked out using accumulation of variance for all other feature extraction methods except for MNF and KECA. MNF uses SNR (Signal-to-Noise Ratio) values and KECA employs Renyi quadratic entropy measurement for this purpose. The studied approaches are equated and analyzed for Indian Pine agricultural and urban Washington DC Mall HSI classification using SVM (Support Vector Machine) classifier. The experiment illustrates that the costly effective and improved classification performance of the feature extraction approaches over the performance using the entire original dataset. MNF offers the highest classification accuracy and FPCA offers the least space and time complexity with satisfactory classification result.
The rise of social media has changed how people view connections. Machine Learning (ML)-based sentiment analysis and news categorization help understand emotions and access news. However, most ...studies focus on complex models requiring heavy resources and slowing inference times, making deployment difficult in resource-limited environments. In this paper, we process both structured and unstructured data, determining the polarity of text using the TextBlob scheme to determine the sentiment of news headlines. We propose a Stochastic Gradient Descent (SGD)-based Ridge classifier (RC) for blending SGDR with an advanced string processing technique to effectively classify news articles. Additionally, we explore existing supervised and unsupervised ML algorithms to gauge the effectiveness of our SGDR classifier. The scalability and generalization capability of SGD and L2 regularization techniques in RCs to handle overfitting and balance bias and variance provide the proposed SGDR with better classification capability. Experimental results highlight that our string processing pipeline significantly boosts the performance of all ML models. Notably, our ensemble SGDR classifier surpasses all state-of-the-art ML algorithms, achieving an impressive 98.12% accuracy. McNemar’s significance tests reveal that our SGDR classifier achieves a 1% significance level improvement over K-Nearest Neighbor, Decision Tree, and AdaBoost and a 5% significance level improvement over other algorithms. These findings underscore the superior proficiency of linear models in news categorization compared to tree-based and nonlinear counterparts. This study contributes valuable insights into the efficacy of the proposed methodology, elucidating its potential for news categorization and sentiment analysis.
Subspace detection from high dimensional hyperspectral image (HSI) data cube has become an important area of research for efficient identification of ground objects. Standard feature extraction ...method such as Principal Component Analysis (PCA) has some drawbacks as it depends solely on global variance of the dataset generated. Folded-PCA (FPCA), an improvement of PCA, offers more benefits over PCA as it envisages both local and global structures of image contents and requires less computation and memory. These superior properties make FPCA more effective for feature extraction in high dimensional remote sensing images e.g. HSIs. Therefore, the proposed feature reduction method combines FPCA feature extraction with Normalized Cross Cumulative Residual Entropy (NCCRE) feature selection, termed as FPCA-NCCRE, for efficient features' subspace detection. NCCRE is utilised as a means of feature selection over the new features generated from FPCA to obtain a more informative subspace. It is experimented on a real mixed agricultural and an urban hyperspectral dataset. Finally, Kernel Support Vector Machine (KSVM) is implemented to calculate the classification accuracy using the detected subspace. From the experiments, it is observed that the proposed method outperforms the baseline approaches and obtains the highest accuracy of 97.67 and 98.57% on the two real hyperspectral images.
Diabetes is a persistent metabolic disorder linked to elevated levels of blood glucose, commonly referred to as blood sugar. This condition can have detrimental effects on the heart, blood vessels, ...eyes, kidneys, and nerves as time passes. It is a chronic ailment that arises when the body fails to produce enough insulin or is unable to effectively use the insulin it produces. When diabetes is not properly managed, it often leads to hyperglycemia, a condition characterized by elevated blood sugar levels or impaired glucose tolerance. This can result in significant harm to various body systems, including the nerves and blood vessels. In this paper, we propose a multiclass diabetes mellitus detection and classification approach using an extremely imbalanced Laboratory of Medical City Hospital data dynamics. We also formulate a new dataset that is moderately imbalanced based on the Laboratory of Medical City Hospital data dynamics. To correctly identify the multiclass diabetes mellitus, we employ three machine learning classifiers namely support vector machine, logistic regression, and k-nearest neighbor. We also focus on dimensionality reduction (feature selection-filter, wrapper, and embedded method) to prune the unnecessary features and to scale up the classification performance. To optimize the classification performance of classifiers, we tune the model by hyperparameter optimization with 10-fold grid search cross-validation. In the case of the original extremely imbalanced dataset with 70:30 partition and support vector machine classifier, we achieved maximum accuracy of 0.964, precision of 0.968, recall of 0.964, F1-score of 0.962, Cohen kappa of 0.835, and AUC of 0.99 by using top 4 feature according to filter method. By using the top 9 features according to wrapper-based sequential feature selection, the k-nearest neighbor provides an accuracy of 0.935 and 1.0 for the other performance metrics. For our created moderately imbalanced dataset with an 80:20 partition, the SVM classifier achieves a maximum accuracy of 0.938, and 1.0 for other performance metrics. For the multiclass diabetes mellitus detection and classification, our experiments outperformed conducted research based on the Laboratory of Medical City Hospital data dynamics.
Federated Averaging (FedAvg) is the baseline Federated Learning (FL) algorithm that applies the stochastic gradient descent for local model training and the arithmetic averaging of the local models' ...parameters for global model aggregation. Succeeding FL works commonly utilize the arithmetic averaging scheme of FedAvg for the aggregation. However, such arithmetic averaging is prone to the outlier model-updates, especially when the clients' data are non-Independent and Identically Distributed (non-IID). As such, the classical aggregation approach suffers from the dominance of the outlier updates and, consequently, causes high communication costs towards producing a decent global model. In this letter, we propose a robust aggregation strategy to alleviate the above issues. In particular, we propose first pruning the node-wise outlier updates (weights) from the local trained models and then performing the aggregation on the selected effective weights-set at each node. We provide the theoretical result of our method and conduct extensive experiments on the MNIST, CIFAR-10, and Shakespeare datasets with IID and non-IID settings, which demonstrate that our aggregation approach outperforms the state-of-the-art methods in terms of communication speedup, test-set performance and training convergence.
A hyperspectral image (HSI), which contains a number of contiguous and narrow spectral wavelength bands, is a valuable source of data for ground cover examinations. Classification using the entire ...original HSI suffers from the "curse of dimensionality" problem because (i) the image bands are highly correlated both spectrally and spatially, (ii) not every band can carry equal information, (iii) there is a lack of enough training samples for some classes, and (iv) the overall computational cost is high. Therefore, effective feature (band) reduction is necessary through feature extraction (FE) and/or feature selection (FS) for improving the classification in a cost-effective manner. Principal component analysis (PCA) is a frequently adopted unsupervised FE method in HSI classification. Nevertheless, its performance worsens when the dataset is noisy, and the computational cost becomes high. Consequently, this study first proposed an efficient FE approach using a normalized mutual information (NMI)-based band grouping strategy, where the classical PCA was applied to each band subgroup for intrinsic FE. Finally, the subspace of the most effective features was generated by the NMI-based minimum redundancy and maximum relevance (mRMR) FS criteria. The subspace of features was then classified using the kernel support vector machine. Two real HSIs collected by the AVIRIS and HYDICE sensors were used in an experiment. The experimental results demonstrated that the proposed feature reduction approach significantly improved the classification performance. It achieved the highest overall classification accuracy of 94.93% for the AVIRIS dataset and 99.026% for the HYDICE dataset. Moreover, the proposed approach reduced the computational cost compared with the studied methods.
Abstract Historically, the analysis of stimulus-dependent time–frequency patterns has been the cornerstone of most electroencephalography (EEG) studies. The abnormal oscillations in high-frequency ...waves associated with psychotic disorders during sensory and cognitive tasks have been studied many times. However, any significant dissimilarity in the resting-state low-frequency bands is yet to be established. Spectral analysis of the alpha and delta band waves shows the effectiveness of stimulus-independent EEG in identifying the abnormal activity patterns of pathological brains. A generalized model incorporating multiple frequency bands should be more efficient in associating potential EEG biomarkers with first-episode psychosis (FEP), leading to an accurate diagnosis. We explore multiple machine-learning methods, including random-forest, support vector machine, and Gaussian process classifier (GPC), to demonstrate the practicality of resting-state power spectral density (PSD) to distinguish patients of FEP from healthy controls. A comprehensive discussion of our preprocessing methods for PSD analysis and a detailed comparison of different models are included in this paper. The GPC model outperforms the other models with a specificity of 95.78% to show that PSD can be used as an effective feature extraction technique for analyzing and classifying resting-state EEG signals of psychiatric disorders.