Abstract Deep learning using convolutional neural networks is an actively emerging field in histological image analysis. This study explores deep learning methods for computer-aided classification in ...H&E stained histopathological whole slide images of gastric carcinoma. An introductory convolutional neural network architecture is proposed for two computerized applications, namely, cancer classification based on immunohistochemical response and necrosis detection based on the existence of tumor necrosis in the tissue. Classification performance of the developed deep learning approach is quantitatively compared with traditional image analysis methods in digital histopathology requiring prior computation of handcrafted features, such as statistical measures using gray level co-occurrence matrix, Gabor filter-bank responses, LBP histograms, gray histograms, HSV histograms and RGB histograms, followed by random forest machine learning. Additionally, the widely known AlexNet deep convolutional framework is comparatively analyzed for the corresponding classification problems. The proposed convolutional neural network architecture reports favorable results, with an overall classification accuracy of 0.6990 for cancer classification and 0.8144 for necrosis detection.
Recurrent neural networks (RNNs) are widely utilized in neural network research to capture spatiotemporal features in video data. However, their effectiveness heavily relies on the spatial features ...upon which they trained. This paper introduces innovative ensembles of features for constructing frame-wise structures by employing impactful neural network models with innovative training pipelines. These features are designed to enhance the recognition of hand gesture videos using RNN by leveraging temporal information. Recognizing hand gestures from videos is a complex task that presents considerable challenges. One notable challenge is the overlap in gesture motion, where different gesture categories exhibit similar hand poses within a single video clip. To overcome this issue, we were motivated to develop extensive and diverse features that offer a more comprehensive description of the gesture video clips, thereby mitigating recognition problems caused by images overlapping. Overall, our efforts to generate diverse features have yielded promising results in enhancing the recognition of hand gestures from videos, particularly in scenarios where overlap poses a significant challenge. We have combined the extracted features from a deep neural network trained from scratch with features obtained from various standard neural networks (Self-Organizing Map, Radial Base Function) that are trained to enhance the deep-trained features. The mutual arrangement for combining the shared features has configured new frame-wise image features. Furthermore, we have provided a performance comparison of the newly constructed frame-wise features through time-sharing to train RNN for recognition. The proposed models have been evaluated on two-hand gesture video datasets, where a preserving gesture sequence is crucial due to overlapping motions. Our work demonstrates a significant improvement in performance for both datasets.
The complexity of natural scenes makes it challenging to experimentally study the mechanisms behind human gaze behavior when viewing dynamic environments. Historically, eye movements were believed to ...be driven primarily by space-based attention towards locations with salient features. Increasing evidence suggests, however, that visual attention does not select locations with high saliency but operates on attentional units given by the objects in the scene. We present a new computational framework to investigate the importance of objects for attentional guidance. This framework is designed to simulate realistic scanpaths for dynamic real-world scenes, including saccade timing and smooth pursuit behavior. Individual model components are based on psychophysically uncovered mechanisms of visual attention and saccadic decision-making. All mechanisms are implemented in a modular fashion with a small number of well-interpretable parameters. To systematically analyze the importance of objects in guiding gaze behavior, we implemented five different models within this framework: two purely spatial models, where one is based on low-level saliency and one on high-level saliency, two object-based models, with one incorporating low-level saliency for each object and the other one not using any saliency information, and a mixed model with object-based attention and selection but space-based inhibition of return. We optimized each model’s parameters to reproduce the saccade amplitude and fixation duration distributions of human scanpaths using evolutionary algorithms. We compared model performance with respect to spatial and temporal fixation behavior, including the proportion of fixations exploring the background, as well as detecting, inspecting, and returning to objects. A model with object-based attention and inhibition, which uses saliency information to prioritize between objects for saccadic selection, leads to scanpath statistics with the highest similarity to the human data. This demonstrates that scanpath models benefit from object-based attention and selection, suggesting that object-level attentional units play an important role in guiding attentional processing.
Celotno besedilo
Dostopno za:
DOBA, IZUM, KILJ, NUK, PILJ, PNG, SAZU, SIK, UILJ, UKNU, UL, UM, UPUK
Most of the world’s 1500 active volcanoes are not instrumentally monitored, resulting in deadly eruptions which can occur without observation of precursory activity. The new Sentinel missions are now ...providing freely available imagery with unprecedented spatial and temporal resolutions, with payloads allowing for a comprehensive monitoring of volcanic hazards. We here present the volcano monitoring platform MOUNTS (Monitoring Unrest from Space), which aims for global monitoring, using multisensor satellite-based imagery (Sentinel-1 Synthetic Aperture Radar SAR, Sentinel-2 Short-Wave InfraRed SWIR, Sentinel-5P TROPOMI), ground-based seismic data (GEOFON and USGS global earthquake catalogues), and artificial intelligence (AI) to assist monitoring tasks. It provides near-real-time access to surface deformation, heat anomalies, SO2 gas emissions, and local seismicity at a number of volcanoes around the globe, providing support to both scientific and operational communities for volcanic risk assessment. Results are visualized on an open-access website where both geocoded images and time series of relevant parameters are provided, allowing for a comprehensive understanding of the temporal evolution of volcanic activity and eruptive products. We further demonstrate that AI can play a key role in such monitoring frameworks. Here we design and train a Convolutional Neural Network (CNN) on synthetically generated interferograms, to operationally detect strong deformation (e.g., related to dyke intrusions), in the real interferograms produced by MOUNTS. The utility of this interdisciplinary approach is illustrated through a number of recent eruptions (Erta Ale 2017, Fuego 2018, Kilauea 2018, Anak Krakatau 2018, Ambrym 2018, and Piton de la Fournaise 2018–2019). We show how exploiting multiple sensors allows for assessment of a variety of volcanic processes in various climatic settings, ranging from subsurface magma intrusion, to surface eruptive deposit emplacement, pre/syn-eruptive morphological changes, and gas propagation into the atmosphere. The data processed by MOUNTS is providing insights into eruptive precursors and eruptive dynamics of these volcanoes, and is sharpening our understanding of how the integration of multiparametric datasets can help better monitor volcanic hazards.
This paper presents a comprehensive review of the principle and application of deep learning in retinal image analysis. Many eye diseases often lead to blindness in the absence of proper clinical ...diagnosis and medical treatment. For example, diabetic retinopathy (DR) is one such disease in which the retinal blood vessels of human eyes are damaged. The ophthalmologists diagnose DR based on their professional knowledge, that is labor intensive. With the advances in image processing and artificial intelligence, computer vision-based techniques have been applied rapidly and widely in the field of medical images analysis and are becoming a better way to advance ophthalmology in practice. Such approaches utilize accurate visual analysis to identify the abnormality of blood vessels with improved performance over manual procedures. More recently, machine learning, in particular, deep learning, has been successfully implemented in this area. In this paper, we focus on recent advances in deep learning methods for retinal image analysis. We review the related publications since 1982, which include more than 80 papers for retinal vessels detections in the research scope spanning from segmentation to classification. Although deep learning has been successfully implemented in other areas, we found only 17 papers so far focus on retinal blood vessel segmentation. This paper characterizes each deep learning based segmentation method as described in the literature. Analyzing along with the limitations and advantages of each method. In the end, we offer some recommendations for future improvement for retinal image analysis.
Display omitted
•Disaggregated different human activities related to LULCC over a complex region.•Quantified the contribution of each driving factor to biophysical parameters using RF.•Strong spatial ...heterogeneity of biophysical parameters in the NTM.•Amplified signals of biophysical parameters over climate change.•Human-induced trends in biophysical parameters produced significant signatures.
Quantifying the variation of biophysical parameters and their driving mechanisms is essential for monitoring land surface environmental changes and for understanding the land–atmosphere interaction in the arid region. Due to the complexity of human activities, most researches are limited to climate change, whereas the response analysis of human activities to changes in biophysical parameters are still lacking or not comprehensively considered. Therefore, large biases and uncertainties still exist in the estimates of regional responses to global change. Firstly, we specifically quantified the main human activities related to land use/land cover change (LULCC) in the northern Tianshan Mountains (NTM), and identified the spatiotemporal changes of primary biophysical parameters, including Albedo, leaf area index (LAI), land surface temperature (LST), and Normalized Difference Vegetation Index (NDVI). Then, we tested the performance of the five models used, including multiple linear regression (MLR), random forest (RF), support vector regression (SVR), multi-layer perceptron (MLP), and K-nearest neighbor (KNN). RF outperformed others and was used to quantify and disaggregate the contribution of climate change and human activities to land surface parameters in the NTM. We found a strong spatial heterogeneity in the spatial variation of all biophysical parameters. Except for LST, the annual maximum Albedo, LAI, and NDVI showed a significant increasing trend in the NTM from 2000 to 2019 (p < 0.05). Generally, climate change contributed more to the biophysical parameters than human activities. However, the contribution of human activities to NDVI was 0.51, which was greater than that of climate change during 2000–2015. This study provides new insight on the impact of climate change and human activities on biophysical parameters and a scientific basis for model parameterization in the arid region.
In this work, we propose a novel denoising technique, the icosahedral mesh denoising network (IMD-Net) for closed genus-0 meshes. IMD-Net is a deep neural network that produces a denoised mesh in a ...single end-to-end pass, preserving and emphasizing natural object features in the process. A preprocessing step, exploiting the homeomorphism between genus-0 mesh and sphere, remeshes an irregular mesh using the regular mesh structure of a frequency subdivided icosahedron. Enabled by gauge equivariant convolutional layers arranged in a residual U-net, IMD-Net denoises the remeshing invariant to global mesh transformations as well as local feature constellations and orientations, doing so with a computational complexity of traditional conv2D kernel. The network is equipped with carefully crafted loss function that leverages differences between positional, normal and curvature fields of target and noisy mesh in a numerically stable fashion. In a first, two large shape datasets commonly used in related fields, ABC and ShapeNetCore , are introduced to evaluate mesh denoising. IMD-Net's competitiveness with existing state-of-the-art techniques is established using both metric evaluations and visual inspection of denoised models. Our code is publicly available at https://github.com/jjabo/IMD-Net .
Blindness is a main threat that affects the daily life activities of any human. Visual prostheses have been introduced to provide artificial vision to the blind with the aim of allowing them to ...restore confidence and independence. In this article, we propose an approach that involves four image enhancement techniques to facilitate object recognition and localization for visual prostheses users. These techniques are clip art representation of the objects, edge sharpening, corner enhancement and electrode dropout handling. The proposed techniques are tested in a real-time mixed reality simulation environment that mimics vision perceived by visual prostheses users. Twelve experiments were conducted to measure the performance of the participants in object recognition and localization. The experiments involved single objects, multiple objects and navigation. To evaluate the performance of the participants in objects recognition, we measure their recognition time, recognition accuracy and confidence level. For object localization, two metrics were used to measure the performance of the participants which are the grasping attempt time and the grasping accuracy. The results demonstrate that using all enhancement techniques simultaneously gives higher accuracy, higher confidence level and less time for recognizing and grasping objects in comparison to not applying the enhancement techniques or applying pair-wise combinations of them. Visual prostheses could benefit from the proposed approach to provide users with an enhanced perception.
Celotno besedilo
Dostopno za:
DOBA, IZUM, KILJ, NUK, PILJ, PNG, SAZU, SIK, UILJ, UKNU, UL, UM, UPUK
The Mountain-Oasis-Desert System (MODS) is the fundamental landscape component within the vast arid region of Central Asia. Human activities and natural processes cause surface displacement in the ...MODS, resulting in geoenvironmental issues and hazards. However, surface displacement and its driving mechanisms in the MODS are not understood. In this study, we used the time-series interferometric synthetic aperture radar (InSAR) method to detect deformation in the MODS in north Tianshan Mountain, Xinjiang and investigated its attribution from the geological conditions, groundwater level changes, and climate variability. The results along the vertical gradient indicate that decorrelation is severe in the high mountain areas with ice-marginal environments (over 3600 m above sea level (a.s.l.)) and the forest-covered mid-mountainous belt (1700-2800 m asl), rendering effective detection unfeasible. Dynamic characteristics are identified in subalpine areas with an elevation of 2800-3600 m asl, with maximum horizontal and vertical displacements of −80.2 mm/year and −58.6 mm/year, respectively. The seasonal acceleration is influenced by the combined effects of temperature and precipitation changes. We observed dense subsidence funnels and ground fissures associated with coal mining in the foreland hills (700-1700 m asl). Ground displacements here exceeded −50 mm/year two years after mining activities had ceased. Subsidence has expanded in the oases with elevations of less than 700 m asl, due to extensive groundwater extraction for agricultural irrigation. The correlation coefficient between groundwater level and displacement was 0.89 and 0.45 at wells in the agricultural and desert areas, respectively. The deformation exhibited seasonal variations associated with groundwater level. The deformation remains relatively stable in the surrounding oasis-affected desert areas, with weak fluctuations influenced by seasonal variations in groundwater level.
This paper proposes the use of Stacked Random Forests (SRF) for the classification of Polarimetric Synthetic Aperture Radar images. SRF apply several Random Forest instances in a sequence where each ...individual uses the class estimate of its predecessor as an additional feature. To this aim, the internal node tests are designed to work not only directly on the complex-valued image data, but also on spatially varying probability distributions and thus allow a seamless integration of RFs within the stacking framework. Experimental results show that the classification performance is consistently improved by the proposed approach, i.e., the achieved accuracy is increased by 4 % and 7 % for one fully- and one dual-polarimetric dataset. This increase only comes at the cost of a linear increased training and prediction time, which is rather limited as the method converges quickly.