The growth in the use of sensor technology has led to the demand for image fusion: signal processing techniques that can combine information received from different sensors into a single composite ...image in an efficient and reliable manner. This book brings together classical and modern algorithms and design architectures, demonstrating through applications how these can be implemented. Image Fusion: Algorithms and Applications provides a representative collection of the recent advances in research and development in the field of image fusion, demonstrating both spatial domain and transform domain fusion methods including Bayesian methods, statistical approaches, ICA and wavelet domain techniques. It also includes valuable material on image mosaics, remote sensing applications and performance evaluation. This book will be an invaluable resource to R engineers, academic researchers and system developers requiring the most up-to-date and complete information on image fusion algorithms, design architectures and applications. * Combines theory and practice to create a unique point of reference * Contains contributions from leading experts in this rapidly-developing field * Demonstrates potential uses in military, medical and civilian areas
The environmental challenges the world faces have never been greater or more complex. Global areas that are covered by forests and urban woodlands are threatened by large-scale forest fires that have ...increased dramatically during the last decades in Europe and worldwide, in terms of both frequency and magnitude. To this end, rapid advances in remote sensing systems including ground-based, unmanned aerial vehicle-based and satellite-based systems have been adopted for effective forest fire surveillance. In this paper, the recently introduced 360-degree sensor cameras are proposed for early fire detection, making it possible to obtain unlimited field of view captures which reduce the number of required sensors and the computational cost and make the systems more efficient. More specifically, once optical 360-degree raw data are obtained using an RGB 360-degree camera mounted on an unmanned aerial vehicle, we convert the equirectangular projection format images to stereographic images. Then, two DeepLab V3+ networks are applied to perform flame and smoke segmentation, respectively. Subsequently, a novel post-validation adaptive method is proposed exploiting the environmental appearance of each test image and reducing the false-positive rates. For evaluating the performance of the proposed system, a dataset, namely the “Fire detection 360-degree dataset”, consisting of 150 unlimited field of view images that contain both synthetic and real fire, was created. Experimental results demonstrate the great potential of the proposed system, which has achieved an F-score fire detection rate equal to 94.6%, hence reducing the number of required sensors. This indicates that the proposed method could significantly contribute to early fire detection.
Convolutional neural networks (CNNs) have resurged lately due to their state-of-the-art performance in various disciplines, such as computer vision, audio and text processing. However, CNNs have not ...been widely employed for remote sensing applications. In this paper, we propose a CNN architecture, named Modular-CNN, to improve the performance of building detectors that employ Histogram of Oriented Gradients (HOG) and Local Binary Patterns (LBP) in a remote sensing dataset. Additionally, we propose two improvements to increase the classification accuracy of Modular-CNN. The first improvement combines the power of raw and normalised features, while the second one concerns the Euler transformation of feature vectors. We demonstrate the effectiveness of our proposed Modular-CNN and the novel improvements in remote sensing and other datasets in a comparative study with other state-of-the-art methods.
Convolutional neural networks (CNN) have enabled significant improvements in pedestrian detection owing to the strong representation ability of the CNN features. However, it is generally difficult to ...reduce false positives on hard negative samples such as tree leaves, traffic lights, poles, etc. Some of these hard negatives can be removed by making use of high level semantic vision cues. In this paper, we propose a region-based CNN method which makes use of semantic cues for better pedestrian detection. Our method extends the Faster R-CNN detection framework by adding a branch of network for semantic image segmentation. The semantic network aims to compute complementary higher level semantic features to be integrated with the convolutional features. We make use of multi-resolution feature maps extracted from different network layers in order to ensure good detection accuracy for pedestrians at different scales. Boosted forest is used for training the integrated features in a cascaded manner for hard negatives mining. Experiments on the Caltech pedestrian dataset show improvements on detection accuracy with the semantic network. With the deep VGG16 model, our pedestrian detection method achieves robust detection performance on the Caltech dataset.
RGB-D salient object detection (SOD) demonstrates its superiority in detecting in complex environments due to the additional depth information introduced in the data. Inevitably, an independent ...stream is introduced to extract features from depth images, leading to extra computation and parameters. This methodology sacrifices the model size to improve the detection accuracy which may impede the practical application of SOD problems. To tackle this dilemma, we propose a dynamic knowledge distillation (DKD) method, along with a lightweight structure, which significantly reduces the computational burden while maintaining validity. This method considers the factors of both teacher and student performance within the training stage and dynamically assigns the distillation weight instead of applying a fixed weight on the student model. We also investigate the issue of RGB-D early fusion strategy in distillation and propose a simple noise elimination method to mitigate the impact of distorted training data caused by low quality depth maps. Extensive experiments are conducted on five public datasets to demonstrate that our method can achieve competitive performance with a fast inference speed (136FPS) compared to 12 prior methods.
With the ever-increasing demand in the analysis and understanding of aerial images in order to remotely recognize targets, this paper introduces a robust system for the detection and localization of ...cars in images captured by air vehicles and satellites. The system adopts a sliding-window approach. It compromises a window-evaluation and a window-classification subsystems. The performance of the proposed framework was evaluated on the Vaihingen dataset. Results demonstrate its superiority to the state of the art.
The task of enhancing the perception of a scene by combining information captured by different sensors is usually known as
image fusion. The
pyramid decomposition and the
Dual-Tree Wavelet Transform ...have been thoroughly applied in image fusion as analysis and synthesis tools. Using a number of
pixel-based and
region-based fusion rules, one can combine the important features of the input images in the transform domain to compose an enhanced image. In this paper, the authors test the efficiency of a transform constructed using
Independent Component Analysis (ICA) and
Topographic Independent Component Analysis bases in image fusion. The bases are obtained by offline training with images of similar context to the observed scene. The images are fused in the transform domain using novel pixel-based or region-based rules. The proposed schemes feature improved performance compared to traditional wavelet approaches with slightly increased computational complexity.
Pedestrian detection is a challenging task, mainly owing to the numerous appearances of human bodies. Modern detectors extract representative features via the deep neural network; however, they ...usually require a large training set and high-performance GPUs. For these cases, we propose a novel human detection approach that integrates a pretrained face detector based on multitask cascaded convolutional neural networks and a traditional pedestrian detector based on aggregate channel features via a score combination module. The proposed detector is a promising approach that can be used to handle pedestrian detection with limited datasets and computational resources. The proposed detector is investigated comprehensively in terms of parameter choices to optimize its performance. The robustness of the proposed detector in terms of the training set, test set, and threshold is observed via tests and cross dataset validations on various pedestrian datasets, including the INRIA, part of the ETHZ, and the Caltech and Citypersons datasets. Experiments have proved that this integrated detector yields a significant increase in recall and a decrease in the log average miss rate compared with sole use of the traditional pedestrian detector. At the same time, the proposed method achieves a comparable performance to FRCNN on the INRIA test set compared with sole use of the Aggregated Channel Features detector.
Recognizing the imperative need for biodiversity protection, the convention on biological diversity (CBD) has recently established new targets towards 2020, the so-called Aichi targets, and updated ...proposed sets of indicators to quantitatively monitor the progress towards these targets. Remote sensing has been increasingly contributing to timely, accurate, and cost-effective assessment of biodiversity-related characteristics and functions during the last years. However, most relevant studies constitute individual research efforts, rarely related with the extraction of widely adopted CBD biodiversity indicators. Furthermore, systematic operational use of remote sensing data by managing authorities has still been limited. In this study, the Aichi targets and the related CBD indicators whose monitoring can be facilitated by remote sensing are identified. For each headline indicator a number of recent remote sensing approaches able for the extraction of related properties are reviewed. Methods cover a wide range of fields, including: habitat extent and condition monitoring; species distribution; pressures from unsustainable management, pollution and climate change; ecosystem service monitoring; and conservation status assessment of protected areas. The advantages and limitations of different remote sensing data and algorithms are discussed. Sorting of the methods based on their reported accuracies is attempted, when possible. The extensive literature survey aims at reviewing highly performing methods that can be used for large-area, effective, and timely biodiversity assessment, to encourage the more systematic use of remote sensing solutions in monitoring progress towards the Aichi targets, and to decrease the gaps between the remote sensing and management communities.
Gastric cancer is one of the most frequent causes of cancer-related deaths worldwide. Gastric atrophy (GA) and gastric intestinal metaplasia (IM) of the mucosa of the stomach have been found to ...increase the risk of gastric cancer and are considered precancerous lesions. Therefore, the early detection of GA and IM may have a valuable role in histopathological risk assessment. However, GA and IM are difficult to confirm endoscopically and, following the Sydney protocol, their diagnosis depends on the analysis of glandular morphology and on the identification of at least one well-defined goblet cell in a set of hematoxylin and eosin (H&E) -stained biopsy samples. To this end, the precise segmentation and classification of glands from the histological images plays an important role in the diagnostic confirmation of GA and IM. In this paper, we propose a digital pathology end-to-end workflow for gastric gland segmentation and classification for the analysis of gastric tissues. The proposed GAGL-VTNet, initially, extracts both global and local features combining multi-scale feature maps for the segmentation of glands and, subsequently, it adopts a vision transformer that exploits the visual dependences of the segmented glands towards their classification. For the analysis of gastric tissues, segmentation of mucosa is performed through an unsupervised model combining energy minimization and a U-Net model. Then, features of the segmented glands and mucosa are extracted and analyzed. To evaluate the efficiency of the proposed methodology we created the GAGL dataset consisting of 85 WSI, collected from 20 patients. The results demonstrate the existence of significant differences of the extracted features between normal, GA and IM cases. The proposed approach for gland and mucosa segmentation achieves an object dice score equal to 0.908 and 0.967 respectively, while for the classification of glands it achieves an F1 score equal to 0.94 showing great potential for the automated quantification and analysis of gastric biopsies.