38 research outputs found
A data relocation approach for terrain surface analysis on multi-GPU systems: a case study on the total viewshed problem
Digital Elevation Models (DEMs) are important datasets for modelling the line
of sight, such as radio signals, sound waves and human vision. These are
commonly analyzed using rotational sweep algorithms. However, such algorithms
require large numbers of memory accesses to 2D arrays which, despite being
regular, result in poor data locality in memory. Here, we propose a new
methodology called skewed Digital Elevation Model (sDEM), which substantially
improves the locality of memory accesses and increases the inherent parallelism
involved in the computation of rotational sweep-based algorithms. In
particular, sDEM applies a data restructuring technique before accessing the
memory and performing the computation. To demonstrate the high efficiency of
sDEM, we use the problem of total viewshed computation as a case study
considering different implementations for single-core, multi-core, single-GPU
and multi-GPU platforms. We conducted two experiments to compare sDEM with (i)
the most commonly used geographic information systems (GIS) software and (ii)
the state-of-the-art algorithm. In the first experiment, sDEM is on average
8.8x faster than current GIS software despite being able to consider only few
points because of their limitations. In the second experiment, sDEM is 827.3x
faster than the state-of-the-art algorithm in the best case
Aerosol type classification with machine learning techniques applied to multiwavelength lidar data from EARLINET
Aerosol typing is essential for understanding atmospheric composition and its impact on the climate. Lidar-based aerosol typing has been often addressed with manual classification using optical property ranges. However, few works addressed it using automated classification with machine learning (ML) mainly due to the lack of annotated datasets. In this study, a high-vertical-resolution dataset is generated and annotated for the University of Granada (UGR) station in Southeastern Spain, which belongs to the European Aerosol Research Lidar Network (EARLINET), identifying five major aerosol types: Continental Polluted, Dust, Mixed, Smoke and Unknown. Six ML models – Decision Tree, Random Forest, Gradient Boosting, XGBoost, LightGBM and Neural Network- were applied to classify aerosol types using multiwavelength lidar data from EARLINET, for two system configurations: with and without depolarization data. LightGBM achieved the best performance, with precision, recall, and F1-Score above 90 % (with depolarization) and close to 87 % (without depolarization). The performance for each aerosol type was evaluated and dust classification improved by ∼ 30 % with depolarization, highlighting its critical role in distinguishing aerosol types. Validation against independent datasets, including a smoke case and a Saharan dust event, confirmed robust classification under real and extreme conditions. Compared to NATALI, a neural network-based EARLINET algorithm, the approach presented in this work shows improved aerosol classification accuracy, which emphasize the benefits of using high-resolution multiwavelength lidar data from real measurements. This highlights the potential of ML-based methods for robust and accurate aerosol typing, establishing a benchmark for future studies using multiwavelength lidar at high-resolution data from EARLINET.</p
TOMOBFLOW: feature-preserving noise filtering for electron tomography
<p>Abstract</p> <p>Background</p> <p>Noise filtering techniques are needed in electron tomography to allow proper interpretation of datasets. The standard linear filtering techniques are characterized by a tradeoff between the amount of reduced noise and the blurring of the features of interest. On the other hand, sophisticated anisotropic nonlinear filtering techniques allow noise reduction with good preservation of structures. However, these techniques are computationally intensive and are difficult to be tuned to the problem at hand.</p> <p>Results</p> <p>TOMOBFLOW is a program for noise filtering with capabilities of preservation of biologically relevant information. It is an efficient implementation of the Beltrami flow, a nonlinear filtering method that locally tunes the strength of the smoothing according to an edge indicator based on geometry properties. The fact that this method does not have free parameters hard to be tuned makes TOMOBFLOW a user-friendly filtering program equipped with the power of diffusion-based filtering methods. Furthermore, TOMOBFLOW is provided with abilities to deal with different types and formats of images in order to make it useful for electron tomography in particular and bioimaging in general.</p> <p>Conclusion</p> <p>TOMOBFLOW allows efficient noise filtering of bioimaging datasets with preservation of the features of interest, thereby yielding data better suited for post-processing, visualization and interpretation. It is available at the web site <url>http://www.ual.es/%7ejjfdez/SW/tomobflow.html</url>.</p
GHOST: Building Blocks for High Performance Sparse Linear Algebra on Heterogeneous Systems
Siting Multiple Observers for Maximum Coverage: An Accurate Approach
AbstractThe selection of the minimal number of observers that ensures the maximum visual coverage over an area represented by a digital elevation model (DEM) have great interest in many fields, e.g., telecommunications, environment planning, among others. However, this problem is complex and intractable when the number of points of the DEM is relatively high. This complexity is due to three issues: 1) the difficulty in determining the visibility of the terrain from one point, 2) the need to know the visibility at all points of the terrain and 3) the combinatorial complexity of the selection of observers.The recent progress in total-viewshed maps computation not only provides an efficient solution to the first two problems, but also opens other ways to new solutions that were unthinkable previously. This paper presents a new type of cartography, called the masked total viewshed map, and provides optimal solutions for both sequential and simultaneous observers location
A tuning approach for iterative multiple 3d stencil pipeline on GPUs: Anisotropic Nonlinear Diffusion algorithm as case study
This paper focuses on challenging applications that can be expressed as an iterative pipeline of multiple 3d stencil stages and explores their optimization space on GPUs. For this study, we selected a representative example from the field of digital signal processing, the Anisotropic Nonlinear Diffusion algorithm. An open issue to these applications is to determine the optimal fission/fusion level of the involved stages and whether that combination benefits from data tiling. This implies exploring a large space of all the possible fission/fusion combinations with and without tiling, thus making the process non-trivial. This study provides insights to reduce the optimization tuning space and programming effort of iterative multiple 3d stencils. Our results demonstrate that all combinations that fuse the bottleneck stencil with high halos update cost (> 25 % , this percentage can be measured or estimated experimentally for each single stencil) and high registers and shared memory accesses must not be considered in the exploration process. The optimal fission/fusion combination is up to 1.65× faster than the case in which we fully decompose our stencil without tiling and 5.3× faster with respect to the fully fused version on the NVIDIA GPUs
