Author(s) ID,Title,Year,DOI,Link,Abstract "56454431400;6603053853;57193876668;56031375100;57193879429;57164651300;55902859900;7401889926;56695201100;57193877725;","Cloud detection algorithm comparison and validation for operational Landsat data products",2017,"10.1016/j.rse.2017.03.026","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85017309987&doi=10.1016%2fj.rse.2017.03.026&partnerID=40&md5=e8f0b58d05400803537d7eb63c1f4ae7","Clouds are a pervasive and unavoidable issue in satellite-borne optical imagery. Accurate, well-documented, and automated cloud detection algorithms are necessary to effectively leverage large collections of remotely sensed data. The Landsat project is uniquely suited for comparative validation of cloud assessment algorithms because the modular architecture of the Landsat ground system allows for quick evaluation of new code, and because Landsat has the most comprehensive manual truth masks of any current satellite data archive. Currently, the Landsat Level-1 Product Generation System (LPGS) uses separate algorithms for determining clouds, cirrus clouds, and snow and/or ice probability on a per-pixel basis. With more bands onboard the Landsat 8 Operational Land Imager (OLI)/Thermal Infrared Sensor (TIRS) satellite, and a greater number of cloud masking algorithms, the U.S. Geological Survey (USGS) is replacing the current cloud masking workflow with a more robust algorithm that is capable of working across multiple Landsat sensors with minimal modification. Because of the inherent error from stray light and intermittent data availability of TIRS, these algorithms need to operate both with and without thermal data. In this study, we created a workflow to evaluate cloud and cloud shadow masking algorithms using cloud validation masks manually derived from both Landsat 7 Enhanced Thematic Mapper Plus (ETM +) and Landsat 8 OLI/TIRS data. We created a new validation dataset consisting of 96 Landsat 8 scenes, representing different biomes and proportions of cloud cover. We evaluated algorithm performance by overall accuracy, omission error, and commission error for both cloud and cloud shadow. We found that CFMask, C code based on the Function of Mask (Fmask) algorithm, and its confidence bands have the best overall accuracy among the many algorithms tested using our validation data. The Artificial Thermal-Automated Cloud Cover Algorithm (AT-ACCA) is the most accurate nonthermal-based algorithm. We give preference to CFMask for operational cloud and cloud shadow detection, as it is derived from a priori knowledge of physical phenomena and is operable without geographic restriction, making it useful for current and future land imaging missions without having to be retrained in a machine-learning environment. © 2017 Elsevier Inc." "56769708700;56422351400;24829272400;35099690000;8557497200;7004040532;","SEMANTIC3D.NET: A NEW LARGE-SCALE POINT CLOUD CLASSIFICATION BENCHMARK",2017,"10.5194/isprs-annals-IV-1-W1-91-2017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85040692236&doi=10.5194%2fisprs-annals-IV-1-W1-91-2017&partnerID=40&md5=fd39b83943b97e48d15923780fea366d","This paper presents a new 3D point cloud classification benchmark data set with over four billion manually labelled points, meant as input for data-hungry (deep) learning methods. We also discuss first submissions to the benchmark that use deep convolutional neural networks (CNNs) as a work horse, which already show remarkable performance improvements over state-of-the-art. CNNs have become the de-facto standard for many tasks in computer vision and machine learning like semantic segmentation or object detection in images, but have no yet led to a true breakthrough for 3D point cloud labelling tasks due to lack of training data. With the massive data set presented in this paper, we aim at closing this data gap to help unleash the full potential of deep learning methods for 3D labelling tasks. Our semantic3D.net data set consists of dense point clouds acquired with static terrestrial laser scanners. It contains 8 semantic classes and covers a wide range of urban outdoor scenes: churches, streets, railroad tracks, squares, villages, soccer fields and castles. We describe our labelling interface and show that our data set provides more dense and complete point clouds with much higher overall number of labelled points compared to those already available to the research community. We further provide baseline method descriptions and comparison between methods submitted to our online system. We hope semantic3D.net will pave the way for deep learning methods in 3D point cloud labelling to learn richer, more general 3D representations, and first submissions after only a few months indicate that this might indeed be the case. © 2017 Copernicus GmbH. All rights reserved." "7202071824;7005457386;25221840700;48662599000;40360989600;","Characterizing forest canopy structure with lidar composite metrics and machine learning",2011,"10.1016/j.rse.2011.04.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-79957608914&doi=10.1016%2fj.rse.2011.04.001&partnerID=40&md5=df382de967e42e1dcc4135350377bab0","A lack of reliable observations for canopy science research is being partly overcome by the gradual use of lidar remote sensing. This study aims to improve lidar-based canopy characterization with airborne laser scanners through the combined use of lidar composite metrics and machine learning models. Our so-called composite metrics comprise a relatively large number of lidar predictors that tend to retain as much information as possible when reducing raw lidar point clouds into a format suitable as inputs to predictive models of canopy structural variables. The information-rich property of such composite metrics is further complemented by machine learning, which offers an array of supervised learning models capable of relating canopy characteristics to high-dimensional lidar metrics via complex, potentially nonlinear functional relationships. Using coincident lidar and field data over an Eastern Texas forest in USA, we conducted a case study to demonstrate the ubiquitous power of the lidar composite metrics in predicting multiple forest attributes and also illustrated the use of two kernel machines, namely, support vector machine and Gaussian processes (GP). Results show that the two machine learning models in conjunction with the lidar composite metrics outperformed traditional approaches such as the maximum likelihood classifier and linear regression models. For example, the five-fold cross validation for GP regression models (vs. linear/log-linear models) yielded a root mean squared error of 1.06 (2.36) m for Lorey's height, 0.95 (3.43) m for dominant height, 5.34 (8.51) m2/ha for basal area, 21.4 (40.5) Mg/ha for aboveground biomass, 6.54 (9.88) Mg/ha for belowground biomass, 0.75 (2.76) m for canopy base height, 2.2 (2.76) m for canopy ceiling height, 0.015 (0.02) kg/m3 for canopy bulk density, 0.068 (0.133) kg/m2 for available canopy fuel, and 0.33 (0.39) m2/m2 for leaf area index. Moreover, uncertainty estimates from the GP regression were more indicative of the true errors in the predicted canopy variables than those from their linear counterparts. With the ever-increasing accessibility of multisource remote sensing data, we envision a concomitant expansion in the use of advanced statistical methods, such as machine learning, to explore the potentially complex relationships between canopy characteristics and remotely-sensed predictors, accompanied by a desideratum for improved error analysis. © 2011 Elsevier Inc." "55235444500;56382070400;27267770800;35263190200;54584369000;6505786772;36193003800;","A high-performance and in-season classification system of field-level crop types using time-series Landsat data and a machine learning approach",2018,"10.1016/j.rse.2018.02.045","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85043598532&doi=10.1016%2fj.rse.2018.02.045&partnerID=40&md5=25f09c47adb113f0e182698a3737a7e8","Accurate and timely spatial classification of crop types based on remote sensing data is important for both scientific and practical purposes. Spatially explicit crop-type information can be used to estimate crop areas for a variety of monitoring and decision-making applications such as crop insurance, land rental, supply-chain logistics, and financial market forecasting. However, there is no publically available spatially explicit in-season crop-type classification information for the U.S. Corn Belt (a landscape predominated by corn and soybean). Instead, researchers and decision-makers have to wait until four to six months after harvest to have such information from the previous year. The state-of-the-art research on crop-type classification has been shifted from relying on only spectral features of single static images to combining together spectral and time-series information. While Landsat data have a desirable spatial resolution for field-level crop-type classification, the ability to extract temporal phenology information based on Landsat data remains a challenge due to low temporal revisiting frequency and inevitable cloud contamination. To address this challenge and generate accurate, cost-effective, and in-season crop-type classification, this research uses the USDA's Common Land Units (CLUs) to aggregate spectral information for each field based on a time-series Landsat image data stack to largely overcome the cloud contamination issue while exploiting a machine learning model based on Deep Neural Network (DNN) and high-performance computing for intelligent and scalable computation of classification processes. Experiments were designed to evaluate what information is most useful for training the machine learning model for crop-type classification, and how various spatial and temporal factors affect the crop-type classification performance in order to derive timely crop type information. All experiments were conducted over Champaign County located in central Illinois, and a total of 1322 Landsat multi-temporal scenes including all the six optical spectral bands spanning from 2000 to 2015 were used. Computational experiments show the inclusion of temporal phenology information and evenly distributed spatial training samples in the study domain improves classification performance. The shortwave infrared bands show notably better performance than the widely used visible and near-infrared bands for classifying corn and soybean. In comparison with USDA's Crop Data Layer (CDL), this study found a relatively high Overall Accuracy (i.e. the number of the corrected classified fields divided by the number of the total fields) of 96% for classifying corn and soybean across all CLU fields in the Champaign County from 2000 to 2015. Furthermore, our approach achieved 95% Overall Accuracy by late July of the concurrent year for classifying corn and soybean. The findings suggest the methodology presented in this paper is promising for accurate, cost-effective, and in-season classification of field-level crop types, which may be scaled up to large geographic extents such as the U.S. Corn Belt. © 2018" "7202420840;57192694511;23398841900;22235943500;24082070800;","Multilevel cloud detection in remote sensing images based on deep learning",2017,"10.1109/JSTARS.2017.2686488","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85018512559&doi=10.1109%2fJSTARS.2017.2686488&partnerID=40&md5=0005957cc7563c8b3fd1a335c008035d","Cloud detection is one of the important tasks for remote sensing image processing. In this paper, a novel multilevel cloud detection method based on deep learning is proposed for remote sensing images. First, the simple linear iterative clustering (SLIC) method is improved to segment the image into good quality superpixels. Then, a deep convolutional neural network (CNN) with two branches is designed to extract the multiscale features from each superpixel and predict the superpixel as one of three classes including thick cloud, thin cloud, and noncloud. Finally, the predictions of all the superpixels in the image yield the cloud detection result. In the proposed cloud detection framework, the improved SLIC method can obtain accurate cloud boundaries by optimizing initial cluster centers, designing dynamic distance measure, and expanding search space. Moreover, different from traditional cloud detection methods that cannot achieve multilevel detection of cloud, the designed deep CNN model can not only detect cloud but also distinguish thin cloud from thick cloud. Experimental results indicate that the proposed method can detect cloud with higher accuracy and robustness than compared methods. © 2008-2012 IEEE." "54930662100;56055664200;55542665400;55893847400;","Model driven reconstruction of roofs from sparse LIDAR point clouds",2013,"10.1016/j.isprsjprs.2012.11.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84871791540&doi=10.1016%2fj.isprsjprs.2012.11.004&partnerID=40&md5=aba2ca626499742cb58f30183bf6a3f2","This article presents a novel, fully automatic method for the reconstruction of three-dimensional building models with prototypical roofs (CityGML LoD2) from LIDAR data and building footprints. The proposed method derives accurate results from sparse point data sets and is suitable for large area reconstruction. Sparse LIDAR data are widely available nowadays. Robust estimation methods such as RANSAC/MSAC, are applied to derive best fitting roof models in a model-driven way. For the identification of the most probable roof model, supervised machine learning methods (Support Vector Machines) are used. In contrast to standard approaches (where the best model is selected via MDL or AIC), supervised classification is able to incorporate additional features enabling a significant improvement in model selection accuracy. © 2012 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)." "37046713400;6603095001;","Landsat-based classification in the cloud: An opportunity for a paradigm shift in land cover monitoring",2017,"10.1016/j.rse.2017.05.025","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85020096180&doi=10.1016%2fj.rse.2017.05.025&partnerID=40&md5=ec8213125fc17b1acf9ad814b9e714bf","Satellite-derived land cover maps play an important role in many applications, including monitoring of smallholder-dominated agricultural landscapes. New cloud-based computing platforms and satellite sensors offer opportunities for generating land cover maps designed to meet the spatial and temporal requirements of specific applications. Such maps can be a significant improvement compared to existing products, which tend to be coarser than 300 m, are often not representative of areas with fast-paced land use change, and have a fixed set of cover classes. Here, we present two approaches for land cover classification using the Landsat archive within Google Earth Engine. Random forest classification was performed with (1) season-based composites, where median values of individual bands and vegetation indices were generated from four years for each of four seasons, and (2) metric-based composites, where different quantiles were computed for the entire four-year period. These approaches were tested for six land cover types spanning over 18,000 locations in Zambia, with ground “truth” determined by visual inspection of high-resolution imagery from Google Earth. The methods were trained on 30% of these points and tested on the remaining 70%, and results were also compared with existing land cover products. Overall accuracies of about 89% were achieved for the season- and metric-based approaches for individual classes, with 93% and 94% accuracy for distinguishing cropland from non-cropland. For the latter task, the existing Globeland30 dataset based on Landsat had much lower accuracies (around 77% on average), as did existing cover maps at coarser resolutions. Overall, the results support the use of either season or metric-based classification approaches. Both produce better results than those obtained from previous classifiers, which supports a general paradigm shift away from dependence on standard static products and towards custom generation of on-demand cover maps designed to fulfill the needs of each specific application. © 2017 Elsevier Inc." "36604588400;35758381900;23006934800;55918993800;","Improving the accuracy of rainfall rates from optical satellite sensors with machine learning - A random forests-based approach applied to MSG SEVIRI",2014,"10.1016/j.rse.2013.10.026","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84888031568&doi=10.1016%2fj.rse.2013.10.026&partnerID=40&md5=eb790432a7a8bc904f91d615c5867470","The present study aims to investigate the potential of the random forests ensemble classification and regression technique to improve rainfall rate assignment during day, night and twilight (resulting in 24-hour precipitation estimates) based on cloud physical properties retrieved from Meteosat Second Generation (MSG) Spinning Enhanced Visible and InfraRed Imager (SEVIRI) data.Random forests (RF) models contain a combination of characteristics that make them well suited for its application in precipitation remote sensing. One of the key advantages is the ability to capture non-linear association of patterns between predictors and response which becomes important when dealing with complex non-linear events like precipitation. Due to the deficiencies of existing optical rainfall retrievals, the focus of this study is on assigning rainfall rates to precipitating cloud areas in connection with extra-tropical cyclones in mid-latitudes including both convective and advective-stratiform precipitating cloud areas. Hence, the rainfall rates are assigned to rain areas previously identified and classified according to the precipitation formation processes. As predictor variables water vapor-IR differences and IR cloud top temperature are used to incorporate information on cloud top height. δT8.7-10.8 and δT10.8-12.1 are considered to supply information about the cloud phase. Furthermore, spectral SEVIRI channels (VIS0.6, VIS0.8, NIR1.6) and cloud properties (cloud effective radius, cloud optical thickness) are used to include information about the cloud water path during daytime, while suitable combinations of temperature differences (δT3.9-10.8, δT3.9-7.3) are considered during night-time.The development of the rainfall rate retrieval technique is realised in three steps. First, an extensive tuning study is carried out to customise each of the RF models. The daytime, night-time and twilight precipitation events have to be treated separately due to differing information content about the cloud properties between the different times of day. Secondly, the RF models are trained using the optimum values for the number of trees and number of randomly chosen predictor variables found in the tuning study. Finally, the final RF models are used to predict rainfall rates using an independent validation data set and the results are validated against co-located rainfall rates observed by a ground radar network. To train and validate the model, the radar-based RADOLAN RW product from the German Weather Service (DWD) is used which provides area-wide gauge-adjusted hourly precipitation information.Regarding the overall performance, as indicated by the coefficient of determination (Rsq), hourly rainfall rates show already a good correlation with Rsq. = 0.5 (day and night) and Rsq. = 0.48 (twilight) between the satellite and radar based observations. Higher temporal aggregation leads to better agreement. Rsq rises to 0.78 (day), 0.77 (night) and 0.75 (twilight) for 8-h interval. By comparing day, night and twilight performance it becomes evident that daytime precipitation is generally predicted best by the model. Twilight and night-time predictions are generally less accurate but only by a small margin. This may due to the smaller number of predictor variables during twilight and night-time conditions as well as less favourable radiative transfer conditions to obtain the cloud parameters during these periods.However, the results show that with the newly developed method it is possible to assign rainfall rates with good accuracy even on an hourly basis. Furthermore, the rainfall rates can be assigned during day, night and twilight conditions which enables the estimation of rainfall rates 24. h. day. © 2013." "55135808700;57203279842;56539588700;57192838041;15020631200;7003946094;56464782100;7006114701;","A 30-m landsat-derived cropland extent product of Australia and China using random forest machine learning algorithm on Google Earth Engine cloud computing platform",2018,"10.1016/j.isprsjprs.2018.07.017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051136400&doi=10.1016%2fj.isprsjprs.2018.07.017&partnerID=40&md5=31444a63e4b8602ebececc9c0cb54278","Mapping high resolution (30-m or better) cropland extent over very large areas such as continents or large countries or regions accurately, precisely, repeatedly, and rapidly is of great importance for addressing the global food and water security challenges. Such cropland extent products capture individual farm fields, small or large, and are crucial for developing accurate higher-level cropland products such as cropping intensities, crop types, crop watering methods (irrigated or rainfed), crop productivity, and crop water productivity. It also brings many challenges that include handling massively large data volumes, computing power, and collecting resource intensive reference training and validation data over complex geographic and political boundaries. Thereby, this study developed a precise and accurate Landsat 30-m derived cropland extent product for two very important, distinct, diverse, and large countries: Australia and China. The study used of eight bands (blue, green, red, NIR, SWIR1, SWIR2, TIR1, and NDVI) of Landsat-8 every 16-day Operational Land Imager (OLI) data for the years 2013–2015. The classification was performed by using a pixel-based supervised random forest (RF) machine learning algorithm (MLA) executed on the Google Earth Engine (GEE) cloud computing platform. Each band was time-composited over 4–6 time-periods over a year using median value for various agro-ecological zones (AEZs) of Australia and China. This resulted in a 32–48-layer mega-file data-cube (MFDC) for each of the AEZs. Reference training and validation data were gathered from: (a) field visits, (b) sub-meter to 5-m very high spatial resolution imagery (VHRI) data, and (c) ancillary sources such as from the National agriculture bureaus. Croplands versus non-croplands knowledge base for training the RF algorithm were derived from MFDC using 958 reference-training samples for Australia and 2130 reference-training samples for China. The resulting 30-m cropland extent product was assessed for accuracies using independent validation samples: 900 for Australia and 1972 for China. The 30-m cropland extent product of Australia showed an overall accuracy of 97.6% with a producer's accuracy of 98.8% (errors of omissions = 1.2%), and user's accuracy of 79% (errors of commissions = 21%) for the cropland class. For China, overall accuracies were 94% with a producer's accuracy of 80% (errors of omissions = 20%), and user's accuracy of 84.2% (errors of commissions = 15.8%) for cropland class. Total cropland areas of Australia were estimated as 35.1 million hectares and 165.2 million hectares for China. These estimates were higher by 8.6% for Australia and 3.9% for China when compared with the traditionally derived national statistics. The cropland extent product further demonstrated the ability to estimate sub-national cropland areas accurately by providing an R2 value of 0.85 when compared with province-wise cropland areas of China. The study provides a paradigm-shift on how cropland maps are produced using multi-date remote sensing. These products can be browsed at www.croplands.org and made available for download at NASA's Land Processes Distributed Active Archive Center (LP DAAC) https://www.lpdaac.usgs.gov/node/1282. © 2018 The Author(s)" "24075297800;6603900587;","Tree-centric mapping of forest carbon density from airborne laser scanning and hyperspectral data",2016,"10.1111/2041-210X.12575","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84970021971&doi=10.1111%2f2041-210X.12575&partnerID=40&md5=720917790c30aff635e319dd11ca0e0c","Forests are a major component of the global carbon cycle, and accurate estimation of forest carbon stocks and fluxes is important in the context of anthropogenic global change. Airborne laser scanning (ALS) data sets are increasingly recognized as outstanding data sources for high-fidelity mapping of carbon stocks at regional scales. We develop a tree-centric approach to carbon mapping, based on identifying individual tree crowns (ITCs) and species from airborne remote sensing data, from which individual tree carbon stocks are calculated. We identify ITCs from the laser scanning point cloud using a region-growing algorithm and identifying species from airborne hyperspectral data by machine learning. For each detected tree, we predict stem diameter from its height and crown-width estimate. From that point on, we use well-established approaches developed for field-based inventories: above-ground biomasses of trees are estimated using published allometries and summed within plots to estimate carbon density. We show this approach is highly reliable: tests in the Italian Alps demonstrated a close relationship between field- and ALS-based estimates of carbon stocks (r2 = 0·98). Small trees are invisible from the air, and a correction factor is required to accommodate this effect. An advantage of the tree-centric approach over existing area-based methods is that it can produce maps at any scale and is fundamentally based on field-based inventory methods, making it intuitive and transparent. Airborne laser scanning, hyperspectral sensing and computational power are all advancing rapidly, making it increasingly feasible to use ITC approaches for effective mapping of forest carbon density also inside wider carbon mapping programs like REDD++. © 2016 The Authors. Methods in Ecology and Evolution © 2016 British Ecological Society" "52364778400;8255132900;6506700491;57191510259;","Neural networks and support vector machine algorithms for automatic cloud classification of whole-sky ground-based images",2015,"10.1109/LGRS.2014.2356616","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84907992182&doi=10.1109%2fLGRS.2014.2356616&partnerID=40&md5=f11b6a978e842c7cd619d6b9b36ffdb6","Clouds are one of the most important meteorological phenomena affecting the Earth radiation balance. The increasing development of whole-sky images enables temporal and spatial high-resolution sky observations and provides the possibility to understand and quantify cloud effects more accurately. In this letter, an attempt has been made to examine the machine learning [multilayer perceptron (MLP) neural networks and support vector machine (SVM)] capabilities for automatic cloud detection in whole-sky images. The approaches have been tested on a significant number of whole-sky images (containing a variety of cloud overages in different seasons and at different daytimes) from Vigna di Valle and Tor Vergata test sites, located near Rome. The pixel values of red, green, and blue bands of the images have been used as inputs of the mentionedmodels, while the outputs provided classified pixels in terms of cloud coverage or others (cloud-free pixels and sun). For the test data set, the overall accuracies of 95.07%, with a standard deviation of 3.37, and 93.66%, with a standard deviation of 4.45, have been obtained from MLP neural networks and SVM models, respectively. Although the two approaches generally generate similar accuracies, the MLP neural networks gave a better performance in some specific cases where the SVM generates poor accuracy. © 2014 IEEE." "14822202800;7202175203;7005477332;7201463831;15844487400;7405857939;","A data-mining approach to associating MISR smoke plume heights with MODIS fire measurements",2007,"10.1016/j.rse.2006.08.014","https://www.scopus.com/inward/record.uri?eid=2-s2.0-33846898494&doi=10.1016%2fj.rse.2006.08.014&partnerID=40&md5=1a520423a70198fe1796df97a0630c01","Satellites provide unique perspectives on aerosol global and regional spatial and temporal distributions, and offer compelling evidence that visibility and air quality are affected by particulate matter transported over long distances. The heights at which emissions are injected into the atmosphere are major factors governing downwind dispersal. In order to better understand the environmental factors determining injection heights of smoke plumes from wildfires, we have developed a prototype system for automatically searching through several years of MISR and MODIS data to locate fires and the associated smoke plumes and to retrieve injection heights and other relevant measurements from them. We are refining this system and assembling a statistical database, aimed at understanding how injection height relates to the fire severity and local weather conditions. In this paper we focus on our working proof-of-concept system that demonstrates how machine-learning and data mining methods aid in processing of massive volumes of satellite data. Automated algorithms for distinguishing smoke from clouds and other aerosols, identifying plumes, and extracting height data are described. Preliminary results are presented from application to MISR and MODIS data collected over North America during the summer of 2004. © 2006 Elsevier Inc. All rights reserved." "35993628500;56224472600;57211415629;24376360400;","Assessment of multi-temporal, multi-sensor radar and ancillary spatial data for grasslands monitoring in Ireland using machine learning approaches",2014,"10.1016/j.rse.2014.05.018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84902972855&doi=10.1016%2fj.rse.2014.05.018&partnerID=40&md5=bbce338ece09bb7fee575ff41c19c901","Accurate inventories of grasslands are important for studies of carbon dynamics, biodiversity conservation and agricultural management. For regions with persistent cloud cover the use of multi-temporal synthetic aperture radar (SAR) data provides an attractive solution for generating up-to-date inventories of grasslands. This is even more appealing considering the data that will be available from upcoming missions such as Sentinel-1 and ALOS-2. In this study, the performance of three machine learning algorithms; Random Forests (RF), Support Vector Machines (SVM) and the relatively underused Extremely Randomised Trees (ERT) is evaluated for discriminating between grassland types over two large heterogeneous areas of Ireland using multi-temporal, multi-sensor radar and ancillary spatial datasets. A detailed accuracy assessment shows the efficacy of the three algorithms to classify different types of grasslands. Overall accuracies. © 2014 Elsevier Inc." "55948735600;26665743300;","Bayesclumpy: Bayesian inference with clumpy dusty torus models",2009,"10.1088/0004-637X/696/2/2075","https://www.scopus.com/inward/record.uri?eid=2-s2.0-70449770857&doi=10.1088%2f0004-637X%2f696%2f2%2f2075&partnerID=40&md5=462b52712104d33c4b51a24331af289c","Our aim is to present a fast and general Bayesian inference framework based on the synergy between machine learning techniques and standard sampling methods and apply it to infer the physical properties of clumpy dusty torus using infrared photometric high spatial resolution observations of active galactic nuclei. We make use of the Metropolis-Hastings Markov Chain Monte Carlo algorithm for sampling the posterior distribution function. Such distribution results from combining all a priori knowledge about the parameters of the model and the information introduced by the observations. The main difficulty resides in the fact that the model used to explain the observations is computationally demanding and the sampling is very time consuming. For this reason, we apply a set of artificial neural networks that are used to approximate and interpolate a database of models. As a consequence, models not present in the original database can be computed ensuring continuity. We focus on the application of this solution scheme to the recently developed public database of clumpy dusty torus models. The machine learning scheme used in this paper allows us to generate any model from the database using only a factor of 10-4 of the original size of the database and a factor of 10-3 in computing time. The posterior distribution obtained for each model parameter allows us to investigate how the observations constrain the parameters and which ones remain partially or completely undetermined, providing statistically relevant confidence intervals. As an example, the application to the nuclear region of Centaurus A shows that the optical depth of the clouds, the total number of clouds, and the radial extent of the cloud distribution zone are well constrained using only six filters. The code is freely available from the authors. © 2009. The American Astronomical Society. All rights reserved.." "56157868600;55893936000;57205005098;24468572800;","Synergistic use of radar sentinel-1 and optical sentinel-2 imagery for crop mapping: A case study for Belgium",2018,"10.3390/rs10101642","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85055414575&doi=10.3390%2frs10101642&partnerID=40&md5=bf942253e130cba2d35d96ae28fa75ae","A timely inventory of agricultural areas and crop types is an essential requirement for ensuring global food security and allowing early crop monitoring practices. Satellite remote sensing has proven to be an increasingly more reliable tool to identify crop types. With the Copernicus program and its Sentinel satellites, a growing source of satellite remote sensing data is publicly available at no charge. Here, we used joint Sentinel-1 radar and Sentinel-2 optical imagery to create a crop map for Belgium. To ensure homogenous radar and optical inputs across the country, Sentinel-1 12-day backscatter mosaics were created after incidence angle normalization, and Sentinel-2 normalized difference vegetation index (NDVI) images were smoothed to yield 10-daily cloud-free mosaics. An optimized random forest classifier predicted the eight crop types with a maximum accuracy of 82% and a kappa coefficient of 0.77. We found that a combination of radar and optical imagery always outperformed a classification based on single-sensor inputs, and that classification performance increased throughout the season until July, when differences between crop types were largest. Furthermore, we showed that the concept of classification confidence derived from the random forest classifier provided insight into the reliability of the predicted class for each pixel, clearly showing that parcel borders have a lower classification confidence. We concluded that the synergistic use of radar and optical data for crop classification led to richer information increasing classification accuracies compared to optical-only classification. Further work should focus on object-level classification and crop monitoring to exploit the rich potential of combined radar and optical observations. © 2018 by the authors." "19639722300;23991212200;57191851405;23768374200;57202650482;","Could Machine Learning Break the Convection Parameterization Deadlock?",2018,"10.1029/2018GL078202","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048971617&doi=10.1029%2f2018GL078202&partnerID=40&md5=59a3938bd30ceecc2da8fd645b58b368","Representing unresolved moist convection in coarse-scale climate models remains one of the main bottlenecks of current climate simulations. Many of the biases present with parameterized convection are strongly reduced when convection is explicitly resolved (i.e., in cloud resolving models at high spatial resolution approximately a kilometer or so). We here present a novel approach to convective parameterization based on machine learning, using an aquaplanet with prescribed sea surface temperatures as a proof of concept. A deep neural network is trained with a superparameterized version of a climate model in which convection is resolved by thousands of embedded 2-D cloud resolving models. The machine learning representation of convection, which we call the Cloud Brain (CBRAIN), can skillfully predict many of the convective heating, moistening, and radiative features of superparameterization that are most important to climate simulation, although an unintended side effect is to reduce some of the superparameterization's inherent variance. Since as few as three months' high-frequency global training data prove sufficient to provide this skill, the approach presented here opens up a new possibility for a future class of convection parameterizations in climate models that are built “top-down,” that is, by learning salient features of convection from unusually explicit simulations. ©2018. The Authors." "6603053853;7102386628;7401889926;","Development of the landsat data continuity mission cloud-cover assessment algorithms",2012,"10.1109/TGRS.2011.2164087","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84859753252&doi=10.1109%2fTGRS.2011.2164087&partnerID=40&md5=5bfe88188b076104c9fbad4e6ad975a6","The upcoming launch of the Operational Land Imager (OLI) will start the next era of the Landsat program. However, the Automated Cloud-Cover Assessment (CCA) (ACCA) algorithm used on Landsat 7 requires a thermal band and is thus not suited for OLI. There will be a thermal instrument on the Landsat Data Continuity Mission (LDCM)the Thermal Infrared Sensorwhich may not be available during all OLI collections. This illustrates a need for CCA for LDCM in the absence of thermal data. To research possibilities for full-resolution OLI cloud assessment, a global data set of 207 Landsat 7 scenes with manually generated cloud masks was created. It was used to evaluate the ACCA algorithm, showing that the algorithm correctly classified 79.9% of a standard test subset of 3.95 10 9. The data set was also used to develop and validate two successor algorithms for use with OLI dataone derived from an off-the-shelf machine learning package and one based on ACCA but enhanced by a simple neural network. These comprehensive CCA algorithms were shown to correctly classify pixels as cloudy or clear 88.5% and 89.7% of the time, respectively. © 2012 IEEE." "6603669676;24281070000;57189711110;57188730414;7003485805;57188719941;6602243908;55996140200;57204881871;6506116372;26032353800;57204874588;34882466400;7007128655;7801472628;6506912173;32667623300;6602485938;35795531300;16031319400;56199424800;8261749500;56076107300;57204879199;6507365226;24468840300;54416399800;16744797500;6507288684;","Near real-time agriculture monitoring at national scale at parcel resolution: Performance assessment of the Sen2-Agri automated system in various cropping systems around the world",2019,"10.1016/j.rse.2018.11.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057630103&doi=10.1016%2fj.rse.2018.11.007&partnerID=40&md5=fbeaa1da49edcf4de561bbf311387920","The convergence of new EO data flows, new methodological developments and cloud computing infrastructure calls for a paradigm shift in operational agriculture monitoring. The Copernicus Sentinel-2 mission providing a systematic 5-day revisit cycle and free data access opens a completely new avenue for near real-time crop specific monitoring at parcel level over large countries. This research investigated the feasibility to propose methods and to develop an open source system able to generate, at national scale, cloud-free composites, dynamic cropland masks, crop type maps and vegetation status indicators suitable for most cropping systems. The so-called Sen2-Agri system automatically ingests and processes Sentinel-2 and Landsat 8 time series in a seamless way to derive these four products, thanks to streamlined processes based on machine learning algorithms and quality controlled in situ data. It embeds a set of key principles proposed to address the new challenges arising from countrywide 10 m resolution agriculture monitoring. The full-scale demonstration of this system for three entire countries (Ukraine, Mali, South Africa) and five local sites distributed across the world was a major challenge met successfully despite the availability of only one Sentinel-2 satellite in orbit. In situ data were collected for calibration and validation in a timely manner allowing the production of the four Sen2-Agri products over all the demonstration sites. The independent validation of the monthly cropland masks provided for most sites overall accuracy values higher than 90%, and already higher than 80% as early as the mid-season. The crop type maps depicting the 5 main crops for the considered study sites were also successfully validated: overall accuracy values higher than 80% and F1 Scores of the different crop type classes were most often higher than 0.65. These respective results pave the way for countrywide crop specific monitoring system at parcel level bridging the gap between parcel visits and national scale assessment. These full-scale demonstration results clearly highlight the operational agriculture monitoring capacity of the Sen2-Agri system to exploit in near real-time the observation acquired by the Sentinel-2 mission over very large areas. Scaling this open source system on cloud computing infrastructure becomes instrumental to support market transparency while building national monitoring capacity as requested by the AMIS and GEOGLAM G-20 initiatives. © 2018" "56803886700;35096575300;55070340000;56217320100;","Traffic sign detection in MLS acquired point clouds for geometric and image-based semantic inventory",2016,"10.1016/j.isprsjprs.2016.01.019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84958268091&doi=10.1016%2fj.isprsjprs.2016.01.019&partnerID=40&md5=9ba016a453d58e4dedec974e800e0096","Nowadays, mobile laser scanning has become a valid technology for infrastructure inspection. This technology permits collecting accurate 3D point clouds of urban and road environments and the geometric and semantic analysis of data became an active research topic in the last years. This paper focuses on the detection of vertical traffic signs in 3D point clouds acquired by a LYNX Mobile Mapper system, comprised of laser scanning and RGB cameras. Each traffic sign is automatically detected in the LiDAR point cloud, and its main geometric parameters can be automatically extracted, therefore aiding the inventory process. Furthermore, the 3D position of traffic signs are reprojected on the 2D images, which are spatially and temporally synced with the point cloud. Image analysis allows for recognizing the traffic sign semantics using machine learning approaches. The presented method was tested in road and urban scenarios in Galicia (Spain). The recall results for traffic sign detection are close to 98%, and existing false positives can be easily filtered after point cloud projection. Finally, the lack of a large, publicly available Spanish traffic sign database is pointed out. © 2016 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)." "14043813400;6603672796;7101677832;","Cloud classification of satellite radiance data by multicategory support vector machines",2004,"10.1175/1520-0426(2004)021<0159:CCOSRD>2.0.CO;2","https://www.scopus.com/inward/record.uri?eid=2-s2.0-1842731703&doi=10.1175%2f1520-0426%282004%29021%3c0159%3aCCOSRD%3e2.0.CO%3b2&partnerID=40&md5=b0fc20d5ab0b0f27e38cb4e8ae83aa7f","Two-category support vector machines (SVMs) have become very popular in the machine learning community for classification problems and have recently been shown to have good optimality properties for classification purposes. Treating multicategory problems as a series of binary problems is common in the SVM paradigm. However, this approach may fail under a variety of circumstances. The multicategory support vector machine (MSVM), which extends the binary SVM to the multicategory case in a symmetric way, and has good theoretical properties, has recently been proposed. The proposed MSVM in addition provides a unifying framework when there are either equal or unequal misclassification costs, and when there is a possibly nonrepresentative training set. Illustrated herein is the potential of the MSVM as an efficient cloud detection and classification algorithm for use in Earth Observing System models, which require knowledge of whether or not a radiance profile is cloud free. If the profile is not cloud free, it is valuable to have information concerning the type of cloud, for example, ice or water. The MSVM has been applied to simulated MODIS channel data to classify the radiance profiles as coming from clear sky, water clouds, or ice clouds, and the results are promising. It can be seen in simple examples, and application to Moderate Resolution Imaging Spectroradiometer (MODIS) observations, that the method is an improvement over channel-by-channel partitioning. It is believed that the MSVM will be a very useful tool for classification problems in atmospheric sciences. © 2004 American Meteorological Society." "57197839233;57203833129;6603474675;","Using image recognition to automate assessment of cultural ecosystem services from social media photographs",2018,"10.1016/j.ecoser.2017.09.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85029485678&doi=10.1016%2fj.ecoser.2017.09.004&partnerID=40&md5=b7b7787ea8b048657cbc47786df24782","Quantifying and mapping cultural ecosystem services is complex because of their intangibility. Data from social media, such as geo-tagged photographs, have been proposed for mapping cultural use or appreciation of ecosystems. However, manual content analysis and classification of large numbers of photographs is time consuming. This study develops a novel method for automating content analysis of social media photographs for ecosystem services assessment. The approach applies an online machine learning algorithm – Google Cloud Vision – to analyse over 20,000 photographs from Singapore, and uses hierarchical clustering to group these photographs. The accuracy of the classification was assessed by comparison with manual classification. Over 20% of photographs were taken of nature, being of animals or plants. The distribution of nature photographs was concentrated around particular natural attractions, and nature photographs were more likely to occur in parks and areas of high vegetation cover. The approach developed for clustering photographs was accurate and saved approximately 170 h of manual work. The method provides an indicator of cultural ecosystem services that can be applied rapidly over large areas. Automated assessment and mapping of cultural ecosystem services could be used to inform urban planning. © 2017 Elsevier B.V." "46061054400;56984291600;57203119863;","Surface water mapping by deep learning",2017,"10.1109/JSTARS.2017.2735443","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85028462983&doi=10.1109%2fJSTARS.2017.2735443&partnerID=40&md5=9283603714e218ae3e8c4ca89938aa0a","Mapping of surface water is useful in a variety of remote sensing applications, such as estimating the availability of water, measuring its change in time, and predicting droughts and floods. Using the imagery acquired by currently active Landsat missions, a surface water map can be generated from any selected region as often as every 8 days. Traditional Landsat water indices require carefully selected threshold values that vary depending on the region being imaged and on the atmospheric conditions. They also suffer from many false positives, arising mainly from snow and ice, and from terrain and cloud shadows being mistaken for water. Systems that produce high-quality water maps usually rely on ancillary data and complex rule-based expert systems to overcome these problems. Here, we instead adopt a data-driven, deep-learning-based approach to surface water mapping. We propose a fully convolutional neural network that is trained to segment water on Landsat imagery. Our proposed model, named DeepWaterMap, learns the characteristics of water bodies from data drawn from across the globe. The trained model separates water from land, snow, ice, clouds, and shadows using only Landsat bands as input. Our code and trained models are publicly available at http://live.ece.utexas.edu/research/deepwatermap/. © 2008-2012 IEEE." "27168701100;34875408300;56243280300;55870809600;35770445300;","A hybrid conditional random field for estimating the underlying ground surface from airborne LiDAR data",2009,"10.1109/TGRS.2009.2017738","https://www.scopus.com/inward/record.uri?eid=2-s2.0-67949104959&doi=10.1109%2fTGRS.2009.2017738&partnerID=40&md5=b33c58ceaad36b64c759ac9562332c2e","Recent advances in airborne light detection and ranging (LiDAR) technology allow rapid and inexpensive generation of digital surface models (DSMs), 3-D point clouds of buildings, vegetations, cars, and natural terrain features over large regions. However, in many applications, such as flood modeling and landslide prediction, digital terrain models (DTMs), the topography of the bare-Earth surface, are needed. This paper introduces a novel machine learning approach to automatically extract DTMs from their corresponding DSMs. We first classify each point as being either ground or nonground, using supervised learning techniques applied to a variety of features. For the points which are classified as ground, we use the LiDAR measurements as an estimate of the surface height, but, for the nonground points, we have to interpolate between nearby values, which we do using a Gaussian random field. Since our model contains both discrete and continuous latent variables, and is a discriminative (rather than generative) probabilistic model, we call it a hybrid conditional random field. We show that a Maximum a Posteriori estimate of the surface height can be efficiently estimated by using a variant of the Expectation Maximization algorithm. Experiments demonstrate that the accuracy of this learning-based approach outperforms the previous best systems, based on manually tuned heuristics. © 2006 IEEE." "35115334500;25223276500;7004260140;35498942700;8545284000;9737845400;23970956600;6602238915;","Object Classification and Recognition From Mobile Laser Scanning Point Clouds in a Road Environment",2016,"10.1109/TGRS.2015.2476502","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84943179157&doi=10.1109%2fTGRS.2015.2476502&partnerID=40&md5=b386cf7fc5fb22f6300ac265d4db3054","Automatic methods are needed to efficiently process the large point clouds collected using a mobile laser scanning (MLS) system for surveying applications. Machine-learning-based object recognition from MLS point clouds in a road and street environment was studied in order to create maps from the road environment infrastructure. The developed automatic processing workflow included the following phases: the removal of the ground and buildings, segmentation, segment classification, and object location estimation. Several novel geometry-based features, which were previously applied in autonomous driving and general point cloud processing, were applied for the segment classification of MLS point clouds. The features were divided into three sets, i.e., local descriptor histograms (LDHs), spin images, and general shape and point distribution features, respectively. These were used in the classification of the following roadside objects: trees, lamp posts, traffic signs, cars, pedestrians, and hoardings. The accuracy of the object recognition workflow was evaluated using a data set that contained more than 400 objects. LDHs and spin images were applied for the first time for machine-learning-based object classification in MLS point clouds in the surveying applications of the road and street environment. The use of these features improved the classification accuracy by 9.6% (resulting in 87.9% accuracy) compared with the accuracy obtained using 17 general shape and point distribution features that represent the current state of the art in the field of MLS; therefore, significant improvement in the classification accuracy was achieved. Connected component segmentation and ground extraction were the cause of most of the errors and should be thus improved in the future. © 2015 IEEE." "34167797600;6602866234;8531925900;56429171400;57190881419;","Ready-to-use methods for the detection of clouds, cirrus, snow, shadow, water and clear sky pixels in Sentinel-2 MSI images",2016,"10.3390/rs8080666","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84983755378&doi=10.3390%2frs8080666&partnerID=40&md5=ccf4a71122104e506c5208907fc240a3","Classification of clouds, cirrus, snow, shadows and clear sky areas is a crucial step in the pre-processing of optical remote sensing images and is a valuable input for their atmospheric correction. The Multi-Spectral Imager on board the Sentinel-2's of the Copernicus program offers optimized bands for this task and delivers unprecedented amounts of data regarding spatial sampling, global coverage, spectral coverage, and repetition rate. Efficient algorithms are needed to process, or possibly reprocess, those big amounts of data. Techniques based on top-of-atmosphere reflectance spectra for single-pixels without exploitation of external data or spatial context offer the largest potential for parallel data processing and highly optimized processing throughput. Such algorithms can be seen as a baseline for possible trade-offs in processing performance when the application of more sophisticated methods is discussed. We present several ready-to-use classification algorithms which are all based on a publicly available database of manually classified Sentinel-2A images. These algorithms are based on commonly used and newly developed machine learning techniques which drastically reduce the amount of time needed to update the algorithms when new images are added to the database. Several ready-to-use decision trees are presented which allow to correctly label about 91% of the spectra within a validation dataset. While decision trees are simple to implement and easy to understand, they offer only limited classification skill. It improves to 98% when the presented algorithm based on the classical Bayesian method is applied. This method has only recently been used for this task and shows excellent performance concerning classification skill and processing performance. A comparison of the presented algorithms with other commonly used techniques such as random forests, stochastic gradient descent, or support vector machines is also given. Especially random forests and support vector machines show similar classification skill as the classical Bayesian method. © 2016 by the authors." "57209423225;55680966400;24401175200;7004082496;6506388437;7801347466;","Contextual classification of point cloud data by exploiting individual 3D neigbourhoods",2015,"10.5194/isprsannals-II-3-W4-271-2015","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84969193116&doi=10.5194%2fisprsannals-II-3-W4-271-2015&partnerID=40&md5=8430427b18dd4f6348d5377620be257a","The fully automated analysis of 3D point clouds is of great importance in photogrammetry, remote sensing and computer vision. For reliably extracting objects such as buildings, road inventory or vegetation, many approaches rely on the results of a point cloud classification, where each 3D point is assigned a respective semantic class label. Such an assignment, in turn, typically involves statistical methods for feature extraction and machine learning. Whereas the different components in the processing workflow have extensively, but separately been investigated in recent years, the respective connection by sharing the results of crucial tasks across all components has not yet been addressed. This connection not only encapsulates the interrelated issues of neighborhood selection and feature extraction, but also the issue of how to involve spatial context in the classification step. In this paper, we present a novel and generic approach for 3D scene analysis which relies on (i) individually optimized 3D neighborhoods for (ii) the extraction of distinctive geometric features and (iii) the contextual classification of point cloud data. For a labeled benchmark dataset, we demonstrate the beneficial impact of involving contextual information in the classification process and that using individual 3D neighborhoods of optimal size significantly increases the quality of the results for both pointwise and contextual classification. © 2015 Copernicus GmbH. All Rights Reserved." "25630461700;57190487295;6603360932;48161356100;","Google earth engine, open-access satellite data, and machine learning in support of large-area probabilisticwetland mapping",2017,"10.3390/rs9121315","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85038215275&doi=10.3390%2frs9121315&partnerID=40&md5=3b85aac08b4c961044e4ceef1124fbe2","Modern advances in cloud computing and machine-leaning algorithms are shifting the manner in which Earth-observation (EO) data are used for environmental monitoring, particularly as we settle into the era of free, open-access satellite data streams. Wetland delineation represents a particularly worthy application of this emerging research trend, since wetlands are an ecologically important yet chronically under-represented component of contemporary mapping and monitoring programs, particularly at the regional and national levels. Exploiting Google Earth Engine and R Statistical software, we developed a workflow for predicting the probability of wetland occurrence using a boosted regression tree machine-learning framework applied to digital topographic and EO data. Working in a 13,700 km2 study area in northern Alberta, our best models produced excellent results, with AUC (area under the receiver-operator characteristic curve) values of 0.898 and explained-deviance values of 0.708. Our results demonstrate the central role of high-quality topographic variables for modeling wetland distribution at regional scales. Including optical and/or radar variables into the workflow substantially improved model performance, though optical data performed slightly better. Converting our wetland probability-of-occurrence model into a binary Wet-Dry classification yielded an overall accuracy of 85%, which is virtually identical to that derived from the Alberta MergedWetland Inventory (AMWI): the contemporary inventory used by the Government of Alberta. However, our workflow contains several key advantages over that used to produce the AMWI, and provides a scalable foundation for province-wide monitoring initiatives. © 2017 by the author." "57205734453;6603104514;7102768607;7101806481;6603683681;24479181300;","Quasi-stellar object selection algorithm using time variability and machine learning: Selection of 1620 quasi-stellar object candidates from MACHO Large Magellanic Cloud database",2011,"10.1088/0004-637X/735/2/68","https://www.scopus.com/inward/record.uri?eid=2-s2.0-79959918266&doi=10.1088%2f0004-637X%2f735%2f2%2f68&partnerID=40&md5=b00538db799ee2602106f9f5fe9a3973","We present a new quasi-stellar object (QSO) selection algorithm using a Support Vector Machine, a supervised classification method, on a set of extracted time series features including period, amplitude, color, and autocorrelation value. We train a model that separates QSOs from variable stars, non-variable stars, and microlensing events using 58 known QSOs, 1629 variable stars, and 4288 non-variables in the MAssive Compact Halo Object (MACHO) database as a training set. To estimate the efficiency and the accuracy of the model, we perform a cross-validation test using the training set. The test shows that the model correctly identifies 80% of known QSOs with a 25% false-positive rate. The majority of the false positives are Be stars. We applied the trained model to the MACHO Large Magellanic Cloud (LMC) data set, which consists of 40 million light curves, and found 1620 QSO candidates. During the selection none of the 33,242 known MACHO variables were misclassified as QSO candidates. In order to estimate the true false-positive rate, we crossmatched the candidates with astronomical catalogs including the Spitzer Surveying the Agents of a Galaxy's Evolution LMC catalog and a few X-ray catalogs. The results further suggest that the majority of the candidates, more than 70%, are QSOs. © 2011. The American Astronomical Society. All rights reserved.." "57205734453;6603104514;7004471421;7102768607;55271541100;6701513404;24448555800;","The EPOCH Project: I. Periodic variable stars in the EROS-2 LMC database â",2014,"10.1051/0004-6361/201323252","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84902076715&doi=10.1051%2f0004-6361%2f201323252&partnerID=40&md5=f80045e3fae596f518c2f591f5d2d8b7","The EPOCH (EROS-2 periodic variable star classification using machine learning) project aims to detect periodic variable stars in the EROS-2 light curve database. In this paper, we present the first result of the classification of periodic variable stars in the EROS-2 LMC database. To classify these variables, we first built a training set by compiling known variables in the Large Magellanic Cloud area from the OGLE and MACHO surveys. We crossmatched these variables with the EROS-2 sources and extracted 22 variability features from 28âE 392 light curves of the corresponding EROS-2 sources. We then used the random forest method to classify the EROS-2 sources in the training set. We designed the model to separate not only δ Scuti stars, RR Lyraes, Cepheids, eclipsing binaries, and long-period variables, the superclasses, but also their subclasses, such as RRab, RRc, RRd, and RRe for RR Lyraes, and similarly for the other variable types. The model trained using only the superclasses shows 99% recall and precision, while the model trained on all subclasses shows 87% recall and precision. We applied the trained model to the entire EROS-2 LMC database, which contains about 29 million sources, and found 117234 periodic variable candidates. Out of these 117234 periodic variables, 55 285 have not been discovered by either OGLE or MACHO variability studies. This set comprises 1906 δ Scuti stars, 6607 RR Lyraes, 638 Cepheids, 178 Type II Cepheids, 34562 eclipsing binaries, and 11394 long-period variables. © ESO, 2014." "7401836526;55558939800;57204297370;36097134700;","Earth System Modeling 2.0: A Blueprint for Models That Learn From Observations and Targeted High-Resolution Simulations",2017,"10.1002/2017GL076101","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85039757485&doi=10.1002%2f2017GL076101&partnerID=40&md5=4e56b653854b8b320621ee6b843ca5a1","Climate projections continue to be marred by large uncertainties, which originate in processes that need to be parameterized, such as clouds, convection, and ecosystems. But rapid progress is now within reach. New computational tools and methods from data assimilation and machine learning make it possible to integrate global observations and local high-resolution simulations in an Earth system model (ESM) that systematically learns from both and quantifies uncertainties. Here we propose a blueprint for such an ESM. We outline how parameterization schemes can learn from global observations and targeted high-resolution simulations, for example, of clouds and convection, through matching low-order statistics between ESMs, observations, and high-resolution simulations. We illustrate learning algorithms for ESMs with a simple dynamical system that shares characteristics of the climate system; and we discuss the opportunities the proposed framework presents and the challenges that remain to realize it. ©2017. American Geophysical Union. All Rights Reserved." "55609188900;23988925500;6507159121;","Generating 3D city models without elevation data",2017,"10.1016/j.compenvurbsys.2017.01.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85009347977&doi=10.1016%2fj.compenvurbsys.2017.01.001&partnerID=40&md5=2999c2052780c3cb0d09e45ae14b633b","Elevation datasets (e.g. point clouds) are an essential but often unavailable ingredient for the construction of 3D city models. We investigate in this paper to what extent can 3D city models be generated solely from 2D data without elevation measurements. We show that it is possible to predict the height of buildings from 2D data (their footprints and attributes available in volunteered geoinformation and cadastre), and then extrude their footprints to obtain 3D models suitable for a multitude of applications. The predictions have been carried out with machine learning techniques (random forests) using 10 different attributes and their combinations, which mirror different scenarios of completeness of real-world data. Some of the scenarios resulted in surprisingly good performance (given the circumstances): we have achieved a mean absolute error of 0.8m in the inferred heights, which satisfies the accuracy recommendations of CityGML for LOD1 models and the needs of several GIS analyses. We show that our method can be used in practice to generate 3D city models where there are no elevation data, and to supplement existing datasets with 3D models of newly constructed buildings to facilitate rapid update and maintenance of data. © 2017 Elsevier Ltd" "14822202800;7004315232;23476370700;55383124200;","An operational MISR pixel classifier using support vector machines",2007,"10.1016/j.rse.2006.06.021","https://www.scopus.com/inward/record.uri?eid=2-s2.0-33846903804&doi=10.1016%2fj.rse.2006.06.021&partnerID=40&md5=50b0e626310abd9a0761451160c059f1","The Multi-angle Imaging SpectroRadiometer (MISR) data products now include a scene classification for each 1.1-km pixel that was developed using Support Vector Machines (SVMs), a cutting-edge machine learning technique for supervised classification. Using a combination of spectral, angular, and texture features, each pixel is classified as land, water, cloud, aerosol, or snow/ice, with the aerosol class further divided into smoke, dust, and other aerosols. The classifier was trained by MISR scientists who labeled hundreds of scenes using a custom interactive tool that showed them the results of the training in real time, making the process significantly faster. Preliminary validation shows that the accuracy of the classifier is approximately 81% globally at the 1.1-km pixel level. Applications of this classifier include global studies of cloud and aerosol distribution, as well as data mining applications such as searching for smoke plumes. This is one of the largest and most ambitious operational uses of machine learning techniques for a remote-sensing instrument, and the success of this system will hopefully lead to further use of this approach. © 2006 Elsevier Inc. All rights reserved." "56959874000;26027537400;23479679000;7403918616;","Prediction of diffuse solar irradiance using machine learning and multivariable regression",2016,"10.1016/j.apenergy.2016.08.093","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84983490932&doi=10.1016%2fj.apenergy.2016.08.093&partnerID=40&md5=974f9a58386065a9fd44159c532f8bad","The paper studies the horizontal global, direct-beam and sky-diffuse solar irradiance data measured in Hong Kong from 2008 to 2013. A machine learning algorithm was employed to predict the horizontal sky-diffuse irradiance and conduct sensitivity analysis for the meteorological variables. Apart from the clearness index (horizontal global/extra atmospheric solar irradiance), we found that predictors including solar altitude, air temperature, cloud cover and visibility are also important in predicting the diffuse component. The mean absolute error (MAE) of the logistic regression using the aforementioned predictors was less than 21.5 W/m2 and 30 W/m2 for Hong Kong and Denver, USA, respectively. With the systematic recording of the five variables for more than 35 years, the proposed model would be appropriate to estimate of long-term diffuse solar radiation, study climate change and develope typical meteorological year in Hong Kong and places with similar climates. © 2016 Elsevier Ltd" "57206827164;7201933692;7006461358;7006613644;","Cloud removal based on sparse representation via multitemporal dictionary learning",2016,"10.1109/TGRS.2015.2509860","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84954489763&doi=10.1109%2fTGRS.2015.2509860&partnerID=40&md5=e32206582ad425ca2e9635a293eeed49","Cloud covers, which generally appear in optical remote sensing images, limit the use of collected images in many applications. It is known that removing these cloud effects is a necessary preprocessing step in remote sensing image analysis. In general, auxiliary images need to be used as the reference images to determine the true ground cover underneath cloud-contaminated areas. In this paper, a new cloud removal approach, which is called multitemporal dictionary learning (MDL), is proposed. Dictionaries of the cloudy areas (target data) and the cloud-free areas (reference data) are learned separately in the spectral domain. The removal process is conducted by combining coefficients from the reference image and the dictionary learned from the target image. This method could well recover the data contaminated by thin and thick clouds or cloud shadows. Our experimental results show that the MDL method is effective in removing clouds from both quantitative and qualitative viewpoints. © 1980-2012 IEEE." "56135196400;7403872687;7401526171;7005052907;6507378331;","A deep neural network modeling framework to reduce bias in satellite precipitation products",2016,"10.1175/JHM-D-15-0075.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84961392473&doi=10.1175%2fJHM-D-15-0075.1&partnerID=40&md5=3fe822f55a1058a64c9d4d6448cc4ad6","Despite the advantage of global coverage at high spatiotemporal resolutions, satellite remotely sensed precipitation estimates still suffer from insufficient accuracy that needs to be improved for weather, climate, and hydrologic applications. This paper presents a framework of a deep neural network (DNN) that improves the accuracy of satellite precipitation products, focusing on reducing the bias and false alarms. The state-of-the-art deep learning techniques developed in the area of machine learning specialize in extracting structural information from a massive amount of image data, which fits nicely into the task of retrieving precipitation data from satellite cloud images. Stacked denoising autoencoder (SDAE), a widely used DNN, is applied to perform bias correction of satellite precipitation products. A case study is conducted on the Precipitation Estimation from Remotely Sensed Information Using Artificial Neural Networks Cloud Classification System (PERSIANN-CCS) with spatial resolution of 0.08° × 0.08° over the central United States, where SDAE is used to process satellite cloud imagery to extract information over a window of 15 × 15 pixels. In the study, the summer of 2012 (June-August) and the winter of 2012/13 (December-February) serve as the training periods, while the same seasons of the following year (summer of 2013 and winter of 2013/14) are used for validation purposes. To demonstrate the effectiveness of the methodology outside the study area, three more regions are selected for additional validation. Significant improvements are achieved in both rain/no-rain (R/NR) detection and precipitation rate quantification: the results make 33% and 43% corrections on false alarm pixels and 98% and 78% bias reductions in precipitation rates over the validation periods of the summer and winter seasons, respectively. © 2016 American Meteorological Society." "57194850453;6603210105;","Point cloud segmentation towards urban ground modeling",2009,"10.1109/URS.2009.5137562","https://www.scopus.com/inward/record.uri?eid=2-s2.0-70350147610&doi=10.1109%2fURS.2009.5137562&partnerID=40&md5=320443227cbaa25a6fd0b6b289631698","This paper presents a new method for segmentation and interpretation of 3D point clouds from mobile LIDAR data. The main contribution of this work is the automatic detection and classification of artifacts located at the ground level. The detection is based on Top-Hat of hole filling algorithm of range images. Then, several features are extracted from the detected connected components (CCs). Afterward, a stepwise forward variable selection by using Wilk's Lambda criterion is performed. Finally, CCs are classified in four categories (lampposts, pedestrians, cars, the others) by using a SVM machine learning method. © 2009 IEEE." "56402112700;9036557400;57201291990;57200517230;23491844400;15725936000;","Classification and mapping of paddy rice by combining Landsat and SAR time series data",2018,"10.3390/rs10030447","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85044229783&doi=10.3390%2frs10030447&partnerID=40&md5=cf14f0ed9e630967d13125e962905ec7","Rice is an important food resource, and the demand for rice has increased as population has expanded. Therefore, accurate paddy rice classification and monitoring are necessary to identify and forecast rice production. Satellite data have been often used to produce paddy rice maps with more frequent update cycle (e.g., every year) than field surveys. Many satellite data, including both optical and SAR sensor data (e.g., Landsat, MODIS, and ALOS PALSAR), have been employed to classify paddy rice. In the present study, time series data from Landsat, RADARSAT-1, and ALOS PALSAR satellite sensors were synergistically used to classify paddy rice through machine learning approaches over two different climate regions (sites A and B). Six schemes considering the composition of various combinations of input data by sensor and collection date were evaluated. Scheme 6 that fused optical and SAR sensor time series data at the decision level yielded the highest accuracy (98.67% for site A and 93.87% for site B). Performance of paddy rice classification was better in site A than site B, which consists of heterogeneous land cover and has low data availability due to a high cloud cover rate. This study also proposed Paddy Rice Mapping Index (PMI) considering spectral and phenological characteristics of paddy rice. PMI represented well the spatial distribution of paddy rice in both regions. Google Earth Engine was adopted to produce paddy rice maps over larger areas using the proposed PMI-based approach. © 2018 by the authors." "57191358778;34881852900;8518459900;57194334833;56089561900;","Cloud detection for high-resolution satellite imagery using machine learning and multi-feature fusion",2016,"10.3390/rs8090715","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85017572164&doi=10.3390%2frs8090715&partnerID=40&md5=1890508ca35f388d671db39becf113bf","The accurate location of clouds in images is prerequisite for many high-resolution satellite imagery applications such as atmospheric correction, land cover classifications, and target recognition. Thus, we propose a novel approach for cloud detection using machine learning and multi-feature fusion based on a comparative analysis of typical spectral, textural, and other feature differences between clouds and backgrounds. To validate this method, we tested it on 102 Gao Fen-1(GF-1) and Gao Fen-2(GF-2) satellite images. The overall accuracy of our multi-feature fusion method for cloud detection was more than 91.45%, and the Kappa coefficient for all the tested images was greater than 80%. The producer and user accuracy were also higher at 93.67% and 95.67%, respectively; both of these values were higher than the values for the other tested feature fusion methods. Our results show that this novel multi-feature approach yields better accuracy than other feature fusion methods. In post-processing, we applied an object-oriented method to remove the influence of highly reflective ground objects and further improved the accuracy. Compared to traditional methods, our new method for cloud detection is accurate, exhibits good scalability, and produces consistent results when mapping clouds of different types and sizes over various land surfaces that contain natural vegetation, agriculture land, built-up areas, and water bodies. © 2016 by the authors." "26323439300;55975488200;56372281600;35725245400;55336245800;8302273000;7003358874;23005893600;","Categorizing grassland vegetation with full-waveform airborne laser scanning: A feasibility study for detecting natura 2000 habitat types",2014,"10.3390/rs6098056","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84907462233&doi=10.3390%2frs6098056&partnerID=40&md5=eeea6fdbcab64f606ed462cfd97124e4","There is increasing demand for reliable, high-resolution vegetation maps covering large areas. Airborne laser scanning data is available for large areas with high resolution and supports automatic processing, therefore, it is well suited for habitat mapping. Lowland hay meadows are widespread habitat types in European grasslands, and also have one of the highest species richness. The objective of this study was to test the applicability of airborne laser scanning for vegetation mapping of different grasslands, including the Natura 2000 habitat type lowland hay meadows. Full waveform leaf-on and leaf-off point clouds were collected from a Natura 2000 site in Sopron, Hungary, covering several grasslands. The LIDAR data were processed to a set of rasters representing point attributes including reflectance, echo width, vegetation height, canopy openness, and surface roughness measures, and these were fused to a multi-band pseudo-image. Random forest machine learning was used for classifying this dataset. Habitat type, dominant plant species and other features of interest were noted in a set of 140 field plots. Two sets of categories were used: five classes focusing on meadow identification and the location of lowland hay meadows, and 10 classes, including eight different grassland vegetation categories. For five classes, an overall accuracy of 75% was reached, for 10 classes, this was 68%. The method delivers unprecedented fine resolution vegetation maps for management and ecological research. We conclude that high-resolution full-waveform LIDAR data can be used to detect grassland vegetation classes relevant for Natura 2000. © 2014 by the authors." "57200017708;57216328909;","Segmentation of LiDAR point clouds for building extraction",2009,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84868557607&partnerID=40&md5=33b207a4d0c433517dfc9277a8b78b13","The objective of segmentation on point clouds is to spatially group points with similar properties into homogeneous regions. Segmentation is a fundamental issue in processing point clouds data acquired by LiDAR and the quality of segmentation largely determines the success of information retrieval. Unlike the image or TIN model, the point clouds do not explicitly represent topology information. As a result, most existing segmentation methods for image and TIN have encountered two difficulties. First, converting data from irregular 3-D point clouds to other models usually leads to information loss; this is particularly a serious drawback for range image based algorithms. Second, the high computation cost of converting a large volume of point data is a considerable problem for any large scale LiDAR applications In this paper, we investigate the strategy to develop LiDAR segmentation methods directly based on point clouds data model. We first discuss several potential local similarity measures based on discrete computation geometry and machine learning. A prototype algorithm based on advanced similarity measures and supported by fast nearest neighborhood search is proposed and implemented. Our experiments show that the proposed method is efficient and robust comparing with the algorithms based on image and TIN. The paper will review popular segmentation methods in related disciplines and present the segmentation results of the proposed method for diverse buildings with different levels of difficulty." "25648879400;55242145000;57193806900;","iPathology: Robotic applications and management of plants and plant diseases",2017,"10.3390/su9061010","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85020627885&doi=10.3390%2fsu9061010&partnerID=40&md5=a030861024e60b8aab7f8717d8aa3013","The rapid development of new technologies and the changing landscape of the online world (e.g., Internet of Things (IoT), Internet of All, cloud-based solutions) provide a unique opportunity for developing automated and robotic systems for urban farming, agriculture, and forestry. Technological advances in machine vision, global positioning systems, laser technologies, actuators, and mechatronics have enabled the development and implementation of robotic systems and intelligent technologies for precision agriculture. Herein, we present and review robotic applications on plant pathology and management, and emerging agricultural technologies for intra-urban agriculture. Greenhouse advanced management systems and technologies have been greatly developed in the last years, integrating IoT and WSN (Wireless Sensor Network). Machine learning, machine vision, and AI (Artificial Intelligence) have been utilized and applied in agriculture for automated and robotic farming. Intelligence technologies, using machine vision/learning, have been developed not only for planting, irrigation, weeding (to some extent), pruning, and harvesting, but also for plant disease detection and identification. However, plant disease detection still represents an intriguing challenge, for both abiotic and biotic stress. Many recognition methods and technologies for identifying plant disease symptoms have been successfully developed; still, the majority of them require a controlled environment for data acquisition to avoid false positives. Machine learning methods (e.g., deep and transfer learning) present promising results for improving image processing and plant symptom identification. Nevertheless, diagnostic specificity is a challenge for microorganism control and should drive the development of mechatronics and robotic solutions for disease management. © 2017 by the authors." "23493093600;7201498455;7402760338;26428503000;56431832800;7007124673;55721771000;6602125464;","Achieving accuracy requirements for forest biomass mapping: A spaceborne data fusion method for estimating forest biomass and LiDAR sampling error",2013,"10.1016/j.rse.2012.11.016","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84871395709&doi=10.1016%2fj.rse.2012.11.016&partnerID=40&md5=8195523ce2a9d555feb18ad1f2ea76db","The synergistic use of active and passive remote sensing (i.e., data fusion) demonstrates the ability of spaceborne light detection and ranging (LiDAR), synthetic aperture radar (SAR) and multispectral imagery for achieving the accuracy requirements of a global forest biomass mapping mission (±20Mgha-1 or 20%, the greater of the two, for at least 80% of grid cells). A data fusion approach also provides a means to extend 3D information from discrete spaceborne LiDAR measurements of forest structure across scales much larger than that of the LiDAR footprint. For estimating biomass, these measurements mix a number of errors including those associated with LiDAR footprint sampling over regional-global extents. A general framework for mapping above ground live forest biomass density (AGB) with a data fusion approach is presented and verified using data from NASA field campaigns near Howland, ME, USA, to assess AGB and LiDAR sampling errors across a regionally representative landscape. We combined SAR and Landsat-derived optical (passive optical) image data to identify contiguous areas (>0.5ha) that are relatively homogenous in remote sensing metrics (forest patches). We used this image-derived data with simulated spaceborne LiDAR derived from orbit and cloud cover simulations and airborne data from NASA's Laser Vegetation Imaging Sensor (LVIS) to compute AGB and estimate LiDAR sampling error for forest patches and 100m, 250m, 500m, and 1km grid cells. At both the patch and grid scales, we evaluated differences in AGB estimation and sampling error from the combined use of LiDAR with both SAR and passive optical and with either SAR or passive optical alone. First, this data fusion approach demonstrates that incorporating forest patches into the AGB mapping framework can provide sub-grid forest information for coarser grid-level AGB reporting. Second, a data fusion approach for estimating AGB using simulated spaceborne LiDAR with SAR and passive optical image combinations reduced forest AGB sampling errors 12%-38% from those where LiDAR is used with SAR or passive optical alone. In absolute terms, sampling errors were reduced from 14-40Mgha-1 to 11-28Mgha-1 across all grid scales and prediction methods, where minimum sampling errors were 11, 15, 18, and 22Mgha-1 for 1km, 500m, 250m, and 100m grid scales, respectively. Third, spaceborne global scale accuracy requirements were achieved whereby at least 80% of the grid cells at 100m, 250m, 500m, and 1km grid levels met AGB accuracy requirements using a combination of passive optical and SAR along with machine learning methods to predict vegetation structure metrics for forested areas without LiDAR samples. Finally, using either passive optical or SAR, accuracy requirements were met at the 500m and 250m grid level, respectively. © 2012 Elsevier Inc." "55176818100;7004479957;","Prognostic Validation of a Neural Network Unified Physics Parameterization",2018,"10.1029/2018GL078510","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85049858028&doi=10.1029%2f2018GL078510&partnerID=40&md5=80359a63ae5b7d3ae4100ca764b206cc","Weather and climate models approximate diabatic and sub-grid-scale processes in terms of grid-scale variables using parameterizations. Current parameterizations are designed by humans based on physical understanding, observations, and process modeling. As a result, they are numerically efficient and interpretable, but potentially oversimplified. However, the advent of global high-resolution simulations and observations enables a more robust approach based on machine learning. In this letter, a neural network-based parameterization is trained using a near-global aqua-planet simulation with a 4-km resolution (NG-Aqua). The neural network predicts the apparent sources of heat and moisture averaged onto (160 km)2 grid boxes. A numerically stable scheme is obtained by minimizing the prediction error over multiple time steps rather than single one. In prognostic single-column model tests, this scheme matches both the fluctuations and equilibrium of NG-Aqua simulation better than the Community Atmosphere Model does. ©2018. American Geophysical Union. All Rights Reserved." "55918310000;36604588400;35758381900;55918993800;","Comparison of four machine learning algorithms for their applicability in satellite-based optical rainfall retrievals",2016,"10.1016/j.atmosres.2015.09.021","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84954105449&doi=10.1016%2fj.atmosres.2015.09.021&partnerID=40&md5=3737f991049618246525fde63262558a","Machine learning (ML) algorithms have successfully been demonstrated to be valuable tools in satellite-based rainfall retrievals which show the practicability of using ML algorithms when faced with high dimensional and complex data. Moreover, recent developments in parallel computing with ML present new possibilities for training and prediction speed and therefore make their usage in real-time systems feasible. This study compares four ML algorithms - random forests (RF), neural networks (NNET), averaged neural networks (AVNNET) and support vector machines (SVM) - for rainfall area detection and rainfall rate assignment using MSG SEVIRI data over Germany. Satellite-based proxies for cloud top height, cloud top temperature, cloud phase and cloud water path serve as predictor variables. The results indicate an overestimation of rainfall area delineation regardless of the ML algorithm (averaged bias=1.8) but a high probability of detection ranging from 81% (SVM) to 85% (NNET). On a 24-hour basis, the performance of the rainfall rate assignment yielded R2 values between 0.39 (SVM) and 0.44 (AVNNET). Though the differences in the algorithms' performance were rather small, NNET and AVNNET were identified as the most suitable algorithms. On average, they demonstrated the best performance in rainfall area delineation as well as in rainfall rate assignment. NNET's computational speed is an additional advantage in work with large datasets such as in remote sensing based rainfall retrievals. However, since no single algorithm performed considerably better than the others we conclude that further research in providing suitable predictors for rainfall is of greater necessity than an optimization through the choice of the ML algorithm. © 2015 Elsevier B.V." "36021472900;7102909972;26666957000;16028782500;57203381681;7005140378;6701736763;","Savannah woody structure modelling and mapping using multi-frequency (X-, C- and L-band) Synthetic Aperture Radar data",2015,"10.1016/j.isprsjprs.2015.04.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84928709049&doi=10.1016%2fj.isprsjprs.2015.04.007&partnerID=40&md5=9c60b368f6285dc94c059ccb7daebb8e","Structural parameters of the woody component in African savannahs provide estimates of carbon stocks that are vital to the understanding of fuelwood reserves, which is the primary source of energy for 90% of households in South Africa (80% in Sub-Saharan Africa) and are at risk of over utilisation. The woody component can be characterised by various quantifiable woody structural parameters, such as tree cover, tree height, above ground biomass (AGB) or canopy volume, each been useful for different purposes. In contrast to the limited spatial coverage of ground-based approaches, remote sensing has the ability to sense the high spatio-temporal variability of e.g. woody canopy height, cover and biomass, as well as species diversity and phenological status - a defining but challenging set of characteristics typical of African savannahs. Active remote sensing systems (e.g. Light Detection and Ranging - LiDAR; Synthetic Aperture Radar - SAR), on the other hand, may be more effective in quantifying the savannah woody component because of their ability to sense within-canopy properties of the vegetation and its insensitivity to atmosphere and clouds and shadows. Additionally, the various components of a particular target's structure can be sensed differently with SAR depending on the frequency or wavelength of the sensor being utilised. This study sought to test and compare the accuracy of modelling, in a Random Forest machine learning environment, woody above ground biomass (AGB), canopy cover (CC) and total canopy volume (TCV) in South African savannahs using a combination of X-band (TerraSAR-X), C-band (RADARSAT-2) and L-band (ALOS PALSAR) radar datasets. Training and validation data were derived from airborne LiDAR data to evaluate the SAR modelling accuracies. It was concluded that the L-band SAR frequency was more effective in the modelling of the CC (coefficient of determination or R2 of 0.77), TCV (R2 of 0.79) and AGB (R2 of 0.78) metrics in Southern African savannahs than the shorter wavelengths (X- and C-band) both as individual and combined (X+C-band) datasets. The addition of the shortest wavelengths also did not assist in the overall reduction of prediction error across different vegetation conditions (e.g. dense forested conditions, the dense shrubby layer and sparsely vegetated conditions). Although the integration of all three frequencies (X+C+L-band) yielded the best overall results for all three metrics (R2=0.83 for CC and AGB and R2=0.85 for TCV), the improvements were noticeable but marginal in comparison to the L-band alone. The results, thus, do not warrant the acquisition of all three SAR frequency datasets for tree structure monitoring in this environment. © 2015 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)." "6701922957;55246081800;14042154700;57203072527;16507612700;14061095500;55695950800;7006934028;7102081813;7004633875;55666021400;8891681500;6602539167;55816647500;","OGLE-IV real-time transient search",2014,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84908353300&partnerID=40&md5=bdb90e35ece57dd29c005ad27ebf92ab","We present the design and first results of a real-time search for transients within the 650 sq. deg. area around the Magellanic Clouds, conducted as part of the OGLE-IV project and aimed at detecting supernovae, novae and other events. The average sampling of about four days from September to May, yielded a detection of 238 transients in 2012/2013 and 2013/2014 seasons. The superb photometric and astrometric quality of the OGLE data allows for numerous applications of the discovered transients. We use this sample to prepare and train a Machine Learning-based automated classifier for early light curves, which distinguishes major classes of transients with more than 80% of correct answers. Spectroscopically classified 49 supernovae Type Ia are used to construct a Hubble Diagram with statistical scatter of about 0.3 mag and fill the least populated region of the redshifts range in the Union sample. We investigate the influence of host galaxy environments on supernovae statistics and find the mean host extinction of AI = 0. 19 ± 0 .10 mag and AV = 0.39 ± 0. 21 mag based on a subsample of supernovae Type Ia. We show that the positional accuracy of the survey is of the order of 0.5 pixels (0.″ 13) and that the OGLE-IV Transient Detection System is capable of detecting transients within the nuclei of galaxies. We present a few interesting cases of nuclear transients of unknown type. All data on the OGLE transients are made publicly available to the astronomical community via the OGLE website." "36617274700;35221569200;55706577200;55995838700;","Efficient resources provisioning based on load forecasting in cloud",2014,"10.1155/2014/321231","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84896357536&doi=10.1155%2f2014%2f321231&partnerID=40&md5=9f1942a10b56337cdcf0451a409eff62","Cloud providers should ensure QoS while maximizing resources utilization. One optimal strategy is to timely allocate resources in a fine-grained mode according to application's actual resources demand. The necessary precondition of this strategy is obtaining future load information in advance. We propose a multi-step-ahead load forecasting method, KSwSVR, based on statistical learning theory which is suitable for the complex and dynamic characteristics of the cloud computing environment. It integrates an improved support vector regression algorithm and Kalman smoother. Public trace data taken from multitypes of resources were used to verify its prediction accuracy, stability, and adaptability, comparing with AR, BPNN, and standard SVR. Subsequently, based on the predicted results, a simple and efficient strategy is proposed for resource provisioning. CPU allocation experiment indicated it can effectively reduce resources consumption while meeting service level agreements requirements. © 2014 Rongdong Hu et al." "54790503500;57203513586;34874017400;57207877450;56135550100;","Cloud and cloud shadow detection in Landsat imagery based on deep convolutional neural networks",2019,"10.1016/j.rse.2019.03.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063157898&doi=10.1016%2fj.rse.2019.03.007&partnerID=40&md5=d78614d214573bb85a7805b2ce62a370","This paper formulates cloud and cloud shadow detection as a semantic segmentation problem and proposes a deep convolutional neural network (CNN) based method to detect them in Landsat imagery. Different from traditional machine learning methods, deep CNN-based methods convolve the entire input image to extract multi-level spatial and spectral features, and then deconvolve these features to produce the detailed segmentation. In this way, multi-level features from the whole image and all the bands are utilized to label each pixel as cloud, thin cloud, cloud shadow or clear. An adaption of SegNet with 13 convolutional layers and 13 deconvolution layers is proposed in this study. The method is applied to 38 Landsat 7 images and 32 Landsat 8 images which are globally distributed and have pixel-wise cloud and cloud shadow reference masks provided by the U.S. Geological Survey (USGS). In order to process such large images using the adapted SegNet model on a desktop computer, the Landsat Collection 1 scenes are split into non-overlapping 512 * 512 30 m pixel image blocks. 60% of these blocks are used to train the model using the backpropagation algorithm, 10% of the blocks are used to validate the model and tune its parameters, and the remaining 30% of the blocks are used for performance evaluation. Compared with the cloud and cloud shadow masks produced by CFMask, which are provided with the Landsat Collection 1 data, the overall accuracies are significantly improved from 89.88% and 84.58% to 95.26% and 95.47% for the Landsat 7 and Landsat 8 images respectively. The proposed method benefits from the multi-level spatial and spectral features, and results in more than a 40% increase in user's accuracy and in more than a 20% increase in producer's accuracy for cloud shadow detection in Landsat 8 imagery. The issues for operational implementation are discussed. © 2019 Elsevier Inc." "54963866700;7005397699;57202645627;6603135119;7102775296;57189389987;","Mapping vegetation and land use types in Fanjingshan National Nature Reserve using google earth engine",2018,"10.3390/rs10060927","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048961694&doi=10.3390%2frs10060927&partnerID=40&md5=d2bf486501a7772164c71d2c8e4c6874","Fanjingshan National Nature Reserve (FNNR) is a biodiversity hotspot in China that is part of a larger, multi-use landscape where farming, grazing, tourism, and other human activities occur. The steep terrain and persistent cloud cover pose challenges to robust vegetation and land use mapping. Our objective is to develop satellite image classification techniques that can reliably map forest cover and land use while minimizing the cloud and terrain issues, and provide the basis for long-term monitoring. Multi-seasonal Landsat image composites and elevation ancillary layers effectively minimize the persistent cloud cover and terrain issues. Spectral vegetation index (SVI) products and shade/illumination normalization approaches yield significantly higher mapping accuracies, compared to non-normalized spectral bands. Advanced machine learning image classification routines are implemented through the cloud-based Google Earth Engine platform. Optimal classifier parameters (e.g., number of trees and number of features for random forest classifiers) were achieved by using tuning techniques. Accuracy assessment results indicate consistent and effective overall classification (i.e., above 70% mapping accuracies) can be achieved using multi-temporal SVI composites with simple illumination normalization and elevation ancillary data, despite the fact limited training and reference data are available. This efficient and open-access image analysis workflow provides a reliable methodology to remotely monitor forest cover and land use in FNNR and other mountainous forested, cloud prevalent areas. © 2018 by the authors." "57188768363;57213577362;57184968600;","Programming challenges of chatbot: Current and future prospective",2018,"10.1109/R10-HTC.2017.8288910","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047404762&doi=10.1109%2fR10-HTC.2017.8288910&partnerID=40&md5=c2db8b6d1dcc330e4edc2f3ecf6ef31f","In the modern Era of technology, Chatbots is the next big thing in the era of conversational services. Chatbots is a virtual person who can effectively talk to any human being using interactive textual skills. Currently, there are many cloud base Chatbots services which are available for the development and improvement of the chatbot sector such as IBM Watson, Microsoft bot, AWS Lambda, Heroku and many others. A virtual person is based on machine learning and Artificial Intelligence (AI) concepts and due to dynamic nature, there is a drawback in the design and development of these chatbots as they have built-in AI, NLP, programming and conversion services. This paper gives an overview of cloud-based chatbots technologies along with programming of chatbots and challenges of programming in current and future Era of chatbot. © 2017 IEEE." "56332041000;24478692500;7004390513;6508152778;57204171085;55611861900;36720718800;55941606700;6701639567;7202115571;16307483000;7005056887;","Innovation in rangeland monitoring: annual, 30 m, plant functional type percent cover maps for U.S. rangelands, 1984–2017",2018,"10.1002/ecs2.2430","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054833588&doi=10.1002%2fecs2.2430&partnerID=40&md5=de119c9c5d884c6ac53fbbda435ddbe0","Innovations in machine learning and cloud-based computing were merged with historical remote sensing and field data to provide the first moderate resolution, annual, percent cover maps of plant functional types across rangeland ecosystems to effectively and efficiently respond to pressing challenges facing conservation of biodiversity and ecosystem services. We utilized the historical Landsat satellite record, gridded meteorology, abiotic land surface data, and over 30,000 field plots within a Random Forests model to predict per-pixel percent cover of annual forbs and grasses, perennial forbs and grasses, shrubs, and bare ground over the western United States from 1984 to 2017. Results were validated using three independent collections of plot-level measurements, and resulting maps display land cover variation in response to changes in climate, disturbance, and management. The maps, which will be updated annually at the end of each year, provide exciting opportunities to expand and improve rangeland conservation, monitoring, and management. The data open new doors for scientific investigation at an unprecedented blend of temporal fidelity, spatial resolution, and geographic scale. © 2018 The Authors." "56315409000;56396097300;34868367300;48061685100;","Ground-based image analysis: A tutorial on machine-learning techniques and applications",2016,"10.1109/MGRS.2015.2510448","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84976421877&doi=10.1109%2fMGRS.2015.2510448&partnerID=40&md5=ef2d8cab592ce91b4212d9821d29b88a","Ground-based whole-sky cameras have opened up new opportunities for monitoring the earth's atmosphere. These cameras are an important complement to satellite images by providing geoscientists with cheaper, faster, and more localized data. The images captured by whole-sky imagers (WSI) can have high spatial and temporal resolution, which is an important prerequisite for applications such as solar energy modeling, cloud attenuation analysis, local weather prediction, and more. Extracting the valuable information from the huge amount of image data by detecting and analyzing the various entities in these images is challenging. However, powerful machine-learning techniques have become available to aid with the image analysis. This article provides a detailed explanation of recent developments in these techniques and their applications in ground-based imaging, aiming to bridge the gap between computer vision and remote sensing with the help of illustrative examples. We demonstrate the advantages of using machine-learning techniques in ground-based image analysis via three primary applications: segmentation, classification, and denoising. © 2013 IEEE." "14048914300;26030013700;55575497900;6602798575;","Mapping pan-European land cover using Landsat spectral-temporal metrics and the European LUCAS survey",2019,"10.1016/j.rse.2018.12.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057741708&doi=10.1016%2fj.rse.2018.12.001&partnerID=40&md5=d13a7f27bcf843cc3c7b83a79e9efe65","This study analyzed, for the first time, the potential of combining the large European-wide land survey LUCAS (Land Use/Cover Area frame Survey) and Landsat-8 data for mapping pan-European land cover and land use. We used annual and seasonal spectral-temporal metrics and environmental features to map 12 land cover and land use classes across Europe. The spectral-temporal metrics provided an efficient means to capture seasonal variations of land surface spectra and to reduce the impact of clouds and cloud-shadows by relaxing the otherwise strong cloud cover limitations imposed by image-based classification methods. The best classification model was based on Landsat-8 data from three years (2014–2016) and achieved an accuracy of 75.1%, nearly 2 percentage points higher than the classification model based on a single year of Landsat data (2015). Our results indicate that annual pan-European land cover maps are feasible, but that temporally dynamic classes like artificial land, cropland, and grassland still benefit from more frequent satellite observations. The produced pan-European land cover map compared favorably to the existing CORINE (Coordination of Information on the Environment) 2012 land cover dataset. The mapped country-wide area proportions strongly correlated with LUCAS-estimated area proportions (r = 0.98). Differences between mapped and LUCAS sample-based area estimates were highest for broadleaved forest (map area was 9% higher). Grassland and seasonal cropland areas were 7% higher than the LUCAS estimate, respectively. In comparison, the correlation between LUCAS and CORINE area proportions was weaker (r = 0.84) and varied strongly by country. CORINE substantially overestimated seasonal croplands by 63% and underestimated grassland proportions by 37%. Our study shows that combining current state-of-the-art remote sensing methods with the large LUCAS database improves pan-European land cover mapping. Although this study focuses on European land cover, the unique combination of large survey data and machine learning of spectral-temporal metrics, may also serve as a reference case for other regions. The pan-European land cover map for 2015 developed in this study is available under https://doi.pangaea.de/10.1594/PANGAEA.896282. © 2018 Elsevier Inc." "38663687700;55696622200;36445129100;7004040532;","Fusing Meter-Resolution 4-D InSAR Point Clouds and Optical Images for Semantic Urban Infrastructure Monitoring",2017,"10.1109/TGRS.2016.2554563","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84991101920&doi=10.1109%2fTGRS.2016.2554563&partnerID=40&md5=40cf460fa5cfaea1b1cff21e3a147f87","Using synthetic aperture radar (SAR) interferometry to monitor long-term millimeter-level deformation of urban infrastructures, such as individual buildings and bridges, is an emerging and important field in remote sensing. In the state-of-the-art methods, deformation parameters are retrieved and monitored on a pixel basis solely in the SAR image domain. However, the inevitable side-looking imaging geometry of SAR results in undesired occlusion and layover in urban area, rendering the current method less competent for a semantic-level monitoring of different urban infrastructures. This paper presents a framework of a semantic-level deformation monitoring by linking the precise deformation estimates of SAR interferometry and the semantic classification labels of optical images via a 3-D geometric fusion and semantic texturing. The proposed approach provides the first 'SARptical' point cloud of an urban area, which is the SAR tomography point cloud textured with attributes from optical images. This opens a new perspective of InSAR deformation monitoring. Interesting examples on bridge and railway monitoring are demonstrated. © 2016 IEEE." "23009358600;22234792700;56098531800;8369807100;7402674597;","Unmanned aircraft system advances health mapping of fragile polar vegetation",2017,"10.1111/2041-210X.12833","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85023629657&doi=10.1111%2f2041-210X.12833&partnerID=40&md5=c6a881ed2376891c15c947b5153eb8b8","Plants like mosses can be sensitive stress markers of subtle shifts in Arctic and Antarctic environmental conditions, including climate change. Traditional ground-based monitoring of fragile polar vegetation is, however, invasive, labour intensive and physically demanding. High-resolution multispectral satellite observations are an alternative, but even their recent highest achievable spatial resolution is still inadequate, resulting in a significant underestimation of plant health due to spectral mixing and associated reflectance impurities. To resolve these obstacles, we have developed a new method that uses low-altitude unmanned aircraft system (UAS) hyperspectral images of sub-decimeter spatial resolution. Machine-learning support vector regressions (SVR) were employed to infer Antarctic moss vigour from quantitative remote sensing maps of plant canopy chlorophyll content and leaf density. The same maps were derived for comparison purposes from the WorldView-2 high spatial resolution (2.2 m) multispectral satellite data. We found SVR algorithms to be highly efficient in estimating plant health indicators with acceptable root mean square errors (RMSE). The systematic RMSEs for chlorophyll content and leaf density were 3.5–6.0 and 1.3–2.0 times smaller, respectively, than the unsystematic errors. However, application of correctly trained SVR machines on space-borne multispectral images considerably underestimated moss chlorophyll content, while stress indicators retrieved from UAS data were found to be comparable with independent field measurements, providing statistically significant regression coefficients of determination (median r2 =.50, pt test =.0072). This study demonstrates the superior performance of a cost-efficient UAS mapping platform, which can be deployed even under the continuous cloud cover that often obscures optical high-altitude airborne and satellite observations. Antarctic moss vigour maps of appropriate resolution could provide timely and spatially explicit warnings of environmental stress events, including those triggered by climate change. Since our polar vegetation health assessment method is based on physical principles of quantitative spectroscopy, it could be adapted to other short-stature and fragmented plant communities (e.g. tundra grasslands), including alpine and desert regions. It therefore shows potential to become an operational component of any ecological monitoring sensor network. © 2017 The Authors. Methods in Ecology and Evolution © 2017 British Ecological Society" "57194590031;57202650452;36085250400;56460431800;6603353298;56216874200;","Towards global-scale seagrass mapping and monitoring using Sentinel-2 on Google Earth Engine: The case study of the Aegean and Ionian Seas",2018,"10.3390/rs10081227","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051668046&doi=10.3390%2frs10081227&partnerID=40&md5=41f40db24f358f89be0e60254602d1f5","Seagrasses are traversing the epoch of intense anthropogenic impacts that significantly decrease their coverage and invaluable ecosystem services, necessitating accurate and adaptable, global-scale mapping and monitoring solutions. Here, we combine the cloud computing power of Google Earth Engine with the freely available Copernicus Sentinel-2 multispectral image archive, image composition, and machine learning approaches to develop a methodological workflow for large-scale, high spatiotemporal mapping and monitoring of seagrass habitats. The present workflow can be easily tuned to space, time and data input; here, we show its potential, mapping 2510.1 km2 of P. oceanica seagrasses in an area of 40,951 km2 between 0 and 40 m of depth in the Aegean and Ionian Seas (Greek territorial waters) after applying support vector machines to a composite of 1045 Sentinel-2 tiles at 10-m resolution. The overall accuracy of P. oceanica seagrass habitats features an overall accuracy of 72% following validation by an independent field data set to reduce bias. We envision that the introduced flexible, time- and cost-efficient cloud-based chain will provide the crucial seasonal to interannual baseline mapping and monitoring of seagrass ecosystems in global scale, resolving gain and loss trends and assisting coastal conservation, management planning, and ultimately climate change mitigation. © 2018 by the authors." "56022064500;55666021400;57203072527;7102081813;6701922957;8891681500;16507612700;6602539167;14042154700;56398525000;14061095500;55816647500;56800976200;","The OGLE collection of variable stars. Eclipsing binaries in the Magellanic system",2016,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85011920540&partnerID=40&md5=96f6f8fb1d87b9f99db856f88dacf03b","We present the collection of eclipsing binaries in the Large and Small Magellanic Clouds, based on the OGLE survey. It contains 48 605 systems, 40204 belonging to the LMC and 8401 to the SMC. Out of the total number of presented here binaries, 16 374 are the new discoveries. We present the time-series photometry obtained for the selected objects during the fourth phase of the OGLE project. The catalog has been created using a two step machine learning procedure based on the Random Forest algorithm." "55183805600;9235059500;7201772493;35767091400;6603186048;","Comparison of NAIP orthophotography and rapideye satellite imagery for mapping of mining and mine reclamation",2014,"10.1080/15481603.2014.912874","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84901800835&doi=10.1080%2f15481603.2014.912874&partnerID=40&md5=7329d63e2b40bc369e3b48f10c814ab6","National Agriculture Imagery Program (NAIP) orthophotography is a potentially useful data source for land cover classification in the United States due to its nationwide and generally cloud-free coverage, low cost to the public, frequent update interval, and high spatial resolution. Nevertheless, there are challenges when working with NAIP imagery, especially regarding varying viewing geometry, radiometric normalization, and calibration. In this article, we compare NAIP orthophotography and RapidEye satellite imagery for high-resolution mapping of mining and mine reclamation within a mountaintop coal surface mine in the southern coalfields of West Virginia, USA. Two classification algorithms, support vector machines and random forests, were used to classify both data sets. Compared to the RapidEye classification, the NAIP classification resulted in lower overall accuracy and kappa and higher allocation disagreement and quantity disagreement. However, the accuracy of the NAIP classification was improved by reducing the number of classes mapped, using the near-infrared band, using textural measures and feature selection, and reducing the spatial resolution slightly by pixel aggregation or by applying a Gaussian low-pass filter. With such strategies, NAIP data can be a potential alternative to RapidEye satellite data for classification of surface mine land cover. © 2014 Taylor and Francis." "57202441820;57147072600;","A Review on deep learning techniques for 3D sensed data classification",2019,"10.3390/rs11121499","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068178990&doi=10.3390%2frs11121499&partnerID=40&md5=a855733383ef6c44a94bdecae23cb9d9","Over the past decade deep learning has driven progress in 2D image understanding. Despite these advancements, techniques for automatic 3D sensed data understanding, such as point clouds, is comparatively immature. However, with a range of important applications from indoor robotics navigation to national scale remote sensing there is a high demand for algorithms that can learn to automatically understand and classify 3D sensed data. In this paper we review the current state-of-the-art deep learning architectures for processing unstructured Euclidean data. We begin by addressing the background concepts and traditional methodologies. We review the current main approaches, including RGB-D, multi-view, volumetric and fully end-to-end architecture designs. Datasets for each category are documented and explained. Finally, we give a detailed discussion about the future of deep learning for 3D sensed data, using literature to justify the areas where future research would be most valuable. © 2019 by the authors." "57216112748;7006518879;23492163200;7005792869;57191414865;57196192125;55709751800;","Foliar and woody materials discriminated using terrestrial LiDAR in a mixed natural forest",2018,"10.1016/j.jag.2017.09.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85032197039&doi=10.1016%2fj.jag.2017.09.004&partnerID=40&md5=5325722a631b8fd03586bdd4be971f7a","Separation of foliar and woody materials using remotely sensed data is crucial for the accurate estimation of leaf area index (LAI) and woody biomass across forest stands. In this paper, we present a new method to accurately separate foliar and woody materials using terrestrial LiDAR point clouds obtained from ten test sites in a mixed forest in Bavarian Forest National Park, Germany. Firstly, we applied and compared an adaptive radius near-neighbor search algorithm with a fixed radius near-neighbor search method in order to obtain both radiometric and geometric features derived from terrestrial LiDAR point clouds. Secondly, we used a random forest machine learning algorithm to classify foliar and woody materials and examined the impact of understory and slope on the classification accuracy. An average overall accuracy of 84.4% (Kappa = 0.75) was achieved across all experimental plots. The adaptive radius near-neighbor search method outperformed the fixed radius near-neighbor search method. The classification accuracy was significantly higher when the combination of both radiometric and geometric features was utilized. The analysis showed that increasing slope and understory coverage had a significant negative effect on the overall classification accuracy. Our results suggest that the utilization of the adaptive radius near-neighbor search method coupling both radiometric and geometric features has the potential to accurately discriminate foliar and woody materials from terrestrial LiDAR data in a mixed natural forest. © 2017 Elsevier B.V." "56566776600;35233206400;55768214000;6602453684;57205723652;27567900100;7006613644;","A New Cloud Computing Architecture for the Classification of Remote Sensing Data",2017,"10.1109/JSTARS.2016.2603120","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85027470222&doi=10.1109%2fJSTARS.2016.2603120&partnerID=40&md5=2189c4e161712bccb689248a085596db","This paper proposes a new distributed architecture for supervised classification of large volumes of earth observation data on a cloud computing environment. The architecture supports distributed execution, network communication, and fault tolerance in a transparent way to the user. The architecture is composed of three abstraction layers, which support the definition and implementation of applications by researchers from different scientific investigation fields. The implementation of architecture is also discussed. A software prototype (available online), which runs machine learning routines implemented on the cloud using the Waikato Environment for Knowledge Analysis (WEKA), a popular free software licensed under the GNU General Public License, is used for validation. Performance issues are addressed through an experimental analysis in which two supervised classifiers available in WEKA were used: random forest and support vector machines. This paper further describes how to include other classification methods in the available software prototype. © 2008-2012 IEEE." "36604588400;35758381900;23006934800;55918993800;","Precipitation estimates from MSG SEVIRI daytime, nighttime, and twilight data with random forests",2014,"10.1175/JAMC-D-14-0082.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84910652230&doi=10.1175%2fJAMC-D-14-0082.1&partnerID=40&md5=aa13437b339aa829d67d4936fc8eeb5c","A new rainfall retrieval technique for determining rainfall rates in a continuous manner (day, twilight, and night) resulting in a 24-h estimation applicable to midlatitudes is presented. The approach is based on satellite-derived information on cloud-top height, cloud-top temperature, cloud phase, and cloud water path retrieved from Meteosat Second Generation (MSG) Spinning Enhanced Visible and Infrared Imager (SEVIRI) data and uses the random forests (RF) machine-learning algorithm. The technique is realized in three steps: (i) precipitating cloud areas are identified, (ii) the areas are separated into convective and advective-stratiform precipitating areas, and (iii) rainfall rates are assigned separately to the convective and advective-stratiform precipitating areas. Validation studies were carried out for each individual step as well as for the overall procedure using collocated ground-based radar data. Regarding each individual step, the models for rain area and convective precipitation detection produce good results. Both retrieval steps show a general tendency toward elevated prediction skill during summer months and daytime. The RF models for rainfall-rate assignment exhibit similar performance patterns, yet it is noteworthy how well the model is able to predict rainfall rates during nighttime and twilight. The performance of the overall procedure shows a very promising potential to estimate rainfall rates at high temporal and spatial resolutions in an automated manner. The near-real-time continuous applicability of the technique with acceptable prediction performances at 3-8-hourly intervals is particularly remarkable. This provides a very promising basis for future investigations into precipitation estimation based on machine-learning approaches and MSG SEVIRI data. © 2014 American Meteorological Society." "57197786350;35280133000;","Detection and classification of pole-like objects from mobile mapping data",2015,"10.5194/isprsannals-II-3-W5-57-2015","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85055427388&doi=10.5194%2fisprsannals-II-3-W5-57-2015&partnerID=40&md5=baaf9ec0798008a81fae1bd16365c0f2","Laser scanners on a vehicle-based mobile mapping system can capture 3D point-clouds of roads and roadside objects. Since roadside objects have to be maintained periodically, their 3D models are useful for planning maintenance tasks. In our previous work, we proposed a method for detecting cylindrical poles and planar plates in a point-cloud. However, it is often required to further classify pole-like objects into utility poles, streetlights, traffic signals and signs, which are managed by different organizations. In addition, our previous method may fail to extract low pole-like objects, which are often observed in urban residential areas. In this paper, we propose new methods for extracting and classifying pole-like objects. In our method, we robustly extract a wide variety of poles by converting point-clouds into wireframe models and calculating cross-sections between wireframe models and horizontal cutting planes. For classifying pole-like objects, we subdivide a pole-like object into five subsets by extracting poles and planes, and calculate feature values of each subset. Then we apply a supervised machine learning method using feature variables of subsets. In our experiments, our method could achieve excellent results for detection and classification of pole-like objects. © 2015 Copernicus GmbH. All Rights Reserved." "57205266761;55340046900;13612339700;23010378300;","Combining Sentinel-1 and Sentinel-2 data for improved land use and land cover mapping of monsoon regions",2018,"10.1016/j.jag.2018.08.011","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059936380&doi=10.1016%2fj.jag.2018.08.011&partnerID=40&md5=a10f164500a200a9fb45074f46fa087b","Land use and land cover maps can support our understanding of coupled human-environment systems and provide important information for environmental modeling and water resource management. Satellite data are a valuable source for land use and land cover mapping. However, cloud-free or weather independent data are necessary to map cloud-prone regions. This particularly applies to monsoon regions such as the Chennai basin, located in the north of Tamil Nadu and the south of Andhra Pradesh, India, which is influenced by the South Asian Monsoon and has abundant cloud cover, throughout the monsoon season. The Basin is characterized by small-scale agriculture with multiple cropping seasons and the rapidly developing metropolitan area of Chennai. This study aims to generate a land use and land cover map of the Chennai Basin for the cropping season of Rabi 2015/16 and to assess the influence of combining the new ESA Copernicus satellites Sentinel-1 and -2 on classification accuracies. A Random Forest based wrapper approach was applied to select the most relevant radar (Sentinel-1) images for the combination with the optical (Sentinel-2) data. Area proportion weighted accuracy with 95% confidence interval were estimated for the Random Forest models, which differentiated 13 land cover classes. The highest overall accuracy of 91.53% ± 0.89 pp was achieved with a combination of 1 Sentinel-2 and 8 Sentinel-1 scenes. This is an improvement of 5.68 pp over a classification with Sentinel-2 data only. An addition of further Sentinel-1 scenes showed no improvement in overall accuracy. The strongest improvement in class-specific accuracy was achieved for paddy fields. This study shows for the first time how land use and land cover classifications in cloud-prone monsoon regions with small-scale agriculture and multiple cropping patterns can be improved by combining Sentinel-1 and Sentinel-2 data. © 2018 Elsevier B.V." "57195956298;7003352529;36844796700;6505985320;6505946652;35221145900;24476089100;6701751679;","Estimating vegetation biomass and cover across large plots in shrub and grass dominated drylands using terrestrial lidar and machine learning",2018,"10.1016/j.ecolind.2017.09.034","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85030456606&doi=10.1016%2fj.ecolind.2017.09.034&partnerID=40&md5=2a2219373f5e5114ae9d1235f48b53f4","Terrestrial laser scanning (TLS) has been shown to enable an efficient, precise, and non-destructive inventory of vegetation structure at ranges up to hundreds of meters. We developed a method that leverages TLS collections with machine learning techniques to model and map canopy cover and biomass of several classes of short-stature vegetation across large plots. We collected high-definition TLS scans of 26 1-ha plots in desert grasslands and big sagebrush shrublands in southwest Idaho, USA. We used the Random Forests machine learning algorithm to develop decision tree models predicting the biomass and canopy cover of several vegetation classes from statistical descriptors of the aboveground heights of TLS points. Manual measurements of vegetation characteristics collected within each plot served as training and validation data. Models based on five or fewer TLS descriptors of vegetation heights were developed to predict the canopy cover fraction of shrubs (R2 = 0.77, RMSE = 7%), annual grasses (R2 = 0.70, RMSE = 21%), perennial grasses (R2 = 0.36, RMSE = 12%), forbs (R2 = 0.52, RMSE = 6%), bare earth or litter (R2 = 0.49, RMSE = 19%), and the biomass of shrubs (R2 = 0.71, RMSE = 175 g) and herbaceous vegetation (R2 = 0.61, RMSE = 99 g) (all values reported are out-of-bag). Our models explained much of the variability between predictions and manual measurements, and yet we expect that future applications could produce even better results by reducing some of the methodological sources of error that we encountered. Our work demonstrates how TLS can be used efficiently to extend manual measurement of vegetation characteristics from small to large plots in grasslands and shrublands, with potential application to other similarly structured ecosystems. Our method shows that vegetation structural characteristics can be modeled without classifying and delineating individual plants, a challenging and time-consuming step common in previous methods applying TLS to vegetation inventory. Improving application of TLS to studies of shrub-steppe ecosystems will serve immediate management needs by enhancing vegetation inventories, environmental modeling studies, and the ability to train broader datasets collected from air and space. © 2017 Elsevier Ltd" "36816165900;57195632558;7003352529;36844796700;6505985320;6505946652;35221145900;24476089100;","Lidar aboveground vegetation biomass estimates in shrublands: Prediction, uncertainties and application to coarser scales",2017,"10.3390/rs9090903","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85029372450&doi=10.3390%2frs9090903&partnerID=40&md5=8b80ecbf005ce464051b76e6c09a4b30","Our study objectives were to model the aboveground biomass in a xeric shrub-steppe landscape with airborne light detection and ranging (Lidar) and explore the uncertainty associated with the models we created. We incorporated vegetation vertical structure information obtained from Lidar with ground-measured biomass data, allowing us to scale shrub biomass from small field sites (1 m subplots and 1 ha plots) to a larger landscape. A series of airborne Lidar-derived vegetation metrics were trained and linked with the field-measured biomass in Random Forests (RF) regression models. A Stepwise Multiple Regression (SMR) model was also explored as a comparison. Our results demonstrated that the important predictors from Lidar-derived metrics had a strong correlation with field-measured biomass in the RF regression models with a pseudo R2 of 0.76 and RMSE of 125 g/m2 for shrub biomass and a pseudo R2 of 0.74 and RMSE of 141 g/m2 for total biomass, and a weak correlation with field-measured herbaceous biomass. The SMR results were similar but slightly better than RF, explaining 77-79% of the variance, with RMSE ranging from 120 to 129 g/m2 for shrub and total biomass, respectively. We further explored the computational efficiency and relative accuracies of using point cloud and raster Lidar metrics at different resolutions (1 m to 1 ha). Metrics derived from the Lidar point cloud processing led to improved biomass estimates at nearly all resolutions in comparison to raster-derived Lidar metrics. Only at 1 m were the results from the point cloud and raster products nearly equivalent. The best Lidar prediction models of biomass at the plot-level (1 ha) were achieved when Lidar metrics were derived from an average of fine resolution (1 m) metrics to minimize boundary effects and to smooth variability. Overall, both RF and SMR methods explained more than 74% of the variance in biomass, with the most important Lidar variables being associated with vegetation structure and statistical measures of this structure (e.g., standard deviation of height was a strong predictor of biomass). Using our model results, we developed spatially-explicit Lidar estimates of total and shrub biomass across our study site in the Great Basin, U.S.A., for monitoring and planning in this imperiled ecosystem. © 2017 by the authors." "26323439300;35725245400;56372281600;55975488200;23005893600;","Mapping natura 2000 habitat conservation status in a pannonic salt steppe with airborne laser scanning",2015,"10.3390/rs70302991","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84926367025&doi=10.3390%2frs70302991&partnerID=40&md5=170ba23bf7f69c26487bd32c84ab181a","Natura 2000 Habitat Conservation Status is currently evaluated based on fieldwork. However, this is proving to be unfeasible over large areas. The use of remote sensing is increasingly encouraged but covering the full range of ecological variables by such datasets and ensuring compatibility with the traditional assessment methodology has not been achieved yet. We aimed to test Airborne Laser Scanning (ALS) as a source for mapping all variables required by the local official conservation status assessment scheme and to develop an automated method that calculates Natura 2000 conservation status at 0.5 m raster resolution for 24 km2 of Pannonic Salt Steppe habitat (code 1530). We used multi-temporal (summer and winter) ALS point clouds with full-waveform recording and a density of 10 pt/m2. Some required variables were derived from ALS product rasters; others involved vegetation classification layers calculated by machine learning and fuzzy categorization. Thresholds separating favorable and unfavorable values of each variable required by the national assessment scheme were manually calibrated from 10 plots where field-based assessment was carried out. Rasters representing positive and negative scores for each input variable were integrated in a ruleset that exactly follows the Hungarian Natura 2000 assessment scheme for grasslands. Accuracy of each parameter and the final conservation status score and category was evaluated by 10 independent assessment plots. We conclude that ALS is a suitable data source for Natura 2000 assessments in grasslands, and that the national grassland assessment scheme can successfully be used as a GIS processing model for conservation status, ensuring that the output is directly comparable with traditional field based assessments. © 2015 by the authors." "7102252875;53363838700;16315027400;54893731100;7403597367;","Development of an intelligent environmental knowledge system for sustainable agricultural decision support",2014,"10.1016/j.envsoft.2013.10.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84890435004&doi=10.1016%2fj.envsoft.2013.10.004&partnerID=40&md5=56b7eedb0b38e29b7d78b839a76b2596","The purpose of this research was to develop a knowledge recommendation architecture based on unsupervised machine learning and unified resource description framework (RDF) for integrated environmental sensory data sources. In developing this architecture, which is very useful for agricultural decision support systems, we considered web based large-scale dynamic data mining, contextual knowledge extraction, and integrated knowledge representation methods. Five different environmental data sources were considered to develop and test the proposed knowledge recommendation framework called Intelligent Environmental Knowledgebase (i-EKbase); including Bureau of Meteorology SILO, Australian Water Availability Project, Australian Soil Resource Information System, Australian National Cosmic Ray Soil Moisture Monitoring Facility, and NASA's Moderate Resolution Imaging Spectroradiometer. Unsupervised clustering techniques based on Principal Component Analysis (PCA), Fuzzy-C-Means (FCM) and Self-organizing map (SOM) were used to create a 2D colour knowledge map representing the dynamics of the i-EKbase to provide ""prior knowledge"" about the integrated knowledgebase. Prior availability of recommendations from the knowledge base could potentially optimize the accessibility and usability issues related to big data sets and minimize the overall application costs. RDF representation has made i-EKbase flexible enough to publish and integrate on the Linked Open Data cloud. This newly developed system was evaluated as an expert agricultural decision support for sustainable water resource management case study in Australia at Tasmania with promising results. © 2013." "56539588700;57203279842;55135808700;57192838041;15020631200;7003946094;56464782100;","Mapping cropland extent of Southeast and Northeast Asia using multi-year time-series Landsat 30-m data using a random forest classifier on the Google Earth Engine Cloud",2019,"10.1016/j.jag.2018.11.014","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071608895&doi=10.1016%2fj.jag.2018.11.014&partnerID=40&md5=5c59f43feb1e4bde7fd7069a95587ffc","Cropland extent maps are useful components for assessing food security. Ideally, such products are a useful addition to countrywide agricultural statistics since they are not politically biased and can be used to calculate cropland area for any spatial unit from an individual farm to various administrative unites (e.g., state, county, district) within and across nations, which in turn can be used to estimate agricultural productivity as well as degree of disturbance on food security from natural disasters and political conflict. However, existing cropland extent maps over large areas (e.g., Country, region, continent, world) are derived from coarse resolution imagery (250 m to 1 km pixels) and have many limitations such as missing fragmented and\or small farms with mixed signatures from different crop types and\or farming practices that can be, confused with other land cover. As a result, the coarse resolution maps have limited useflness in areas where fields are small (<1 ha), such as in Southeast Asia. Furthermore, coarse resolution cropland maps have known uncertainties in both geo-precision of cropland location as well as accuracies of the product. To overcome these limitations, this research was conducted using multi-date, multi-year 30-m Landsat time-series data for 3 years chosen from 2013 to 2016 for all Southeast and Northeast Asian Countries (SNACs), which included 7 refined agro-ecological zones (RAEZ) and 12 countries (Indonesia, Thailand, Myanmar, Vietnam, Malaysia, Philippines, Cambodia, Japan, North Korea, Laos, South Korea, and Brunei). The 30-m (1 pixel = 0.09 ha) data from Landsat 8 Operational Land Imager (OLI) and Landsat 7 Enhanced Thematic Mapper (ETM+) were used in the study. Ten Landsat bands were used in the analysis (blue, green, red, NIR, SWIR1, SWIR2, Thermal, NDVI, NDWI, LSWI) along with additional layers of standard deviation of these 10 bands across 1 year, and global digital elevation model (GDEM)-derived slope and elevation bands. To reduce the impact of clouds, the Landsat imagery was time-composited over four time-periods (Period 1: January- April, Period 2: May-August, and Period 3: September-December) over 3-years. Period 4 was the standard deviation of all 10 bands taken over all images acquired during the 2015 calendar year. These four period composites, totaling 42 band data-cube, were generated for each of the 7 RAEZs. The reference training data (N = 7849) generated for the 7 RAEZ using sub-meter to 5-m very high spatial resolution imagery (VHRI) helped generate the knowledge-base to separate croplands from non-croplands. This knowledge-base was used to code and run a pixel-based random forest (RF) supervised machine learning algorithm on the Google Earth Engine (GEE) cloud computing environment to separate croplands from non-croplands. The resulting cropland extent products were evaluated using an independent reference validation dataset (N = 1750) in each of the 7 RAEZs as well as for the entire SNAC area. For the entire SNAC area, the overall accuracy was 88.1% with a producer's accuracy of 81.6% (errors of omissions = 18.4%) and user's accuracy of 76.7% (errors of commissions = 23.3%). For each of the 7 RAEZs overall accuracies varied from 83.2 to 96.4%. Cropland areas calculated for the 12 countries were compared with country areas reported by the United Nations Food and Agriculture Organization and other national cropland statistics resulting in an R2 value of 0.93. The cropland areas of provinces were compared with the province statistics that showed an R2 = 0.95 for South Korea and R2 = 0.94 for Thailand. The cropland products are made available on an interactive viewer at www.croplands.org and for download at National Aeronautics and Space Administration's (NASA) Land Processes Distributed Active Archive Center (LP DAAC): https://lpdaac.usgs.gov/node/1281. © 2019" "26667199300;7005523706;55931997700;57204762108;9039420300;","Near-real-time non-obstructed flood inundation mapping using synthetic aperture radar",2019,"10.1016/j.rse.2018.11.008","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85056999557&doi=10.1016%2fj.rse.2018.11.008&partnerID=40&md5=fa94b37022efd6acc5f0ec2632925602","In the event of a flood disaster, first response agencies need inundation maps produced in near real time (NRT). Such maps can be generated using satellite-based information. In this study, we developed mapping techniques that rely on synthetic aperture radar (SAR) on-board earth-orbiting platforms. SAR provides valid ground surface measurements through cloud cover with high resolution and sampling frequency that has recently increased through multiple missions. Despite numerous efforts, automatic processing of SAR data to derive accurate inundation maps still poses challenges. To address them, we have developed an NRT system named RAdar-Produced Inundation Diary (RAPID). RAPID integrates four processing steps: classification based on statistics, morphological processing, multi-threshold-based compensation, and machine-learning correction. Besides SAR data, the system integrates multisource remote-sensing data products, including land cover classification, water occurrence, hydrographical, water type, and river width products. In comparison to expert handmade flood maps, the fully-automated RAPID system exhibited “overall,” “producer,” and “user” accuracies of 93%, 77%, and 75%, respectively. RAPID accommodates commonly encountered over- and under-detections caused by noise-like speckle, water-like radar response areas, strong scatterers, and isolated inundation areas—errors that are in common practice to ignore, mask out, or be filtered out by coarsening the effective resolution. RAPID can serve as the kernel algorithm to derive flood inundation products from satellites—both existing and to be launched—equipped with high-resolution SAR sensors, including Envisat, Radarsat, NISAR, Advanced Land Observation Satellite (ALOS)-1/2, Sentinel-1, and TerraSAR-X. © 2018 Elsevier Inc." "56036795100;36816335800;57199057325;57199051279;35330367300;","Development of a support vector machine based cloud detection method for MODIS with the adjustability to various conditions",2018,"10.1016/j.rse.2017.11.003","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85037536784&doi=10.1016%2fj.rse.2017.11.003&partnerID=40&md5=f465f293236b6e9e009f07ffaef8f153","Common requirements for cloud detection methods including the adjustability with respect to incorrect results are clarified, and a method is proposed that satisfies the requirements by applying the support vector machine (SVM). Because the conditions of clouds and Earth's surfaces vary widely, incorrect results in actual cloud detection operations are unavoidable. Cloud detection methods therefore should be adjustable to easily reduce the frequency of incorrect results under certain conditions, without causing new incorrect results under other conditions. Cloud detection methods are also required to resolve a characteristic issue: the boundary between clear-sky and cloudy-sky areas in nature is vague, because the density of the cloud particles continuously varies. This vagueness makes the cloud definition subjective. Furthermore, the training dataset preparation for machine learning should avoid circular arguments. The SVM learning is generally less likely to result in overfitting: this study suggests that only typical data are sufficient for the SVM training dataset. By incorporating the discriminant analysis (DA), it is possible to subjectively determine the definition of typical cloudy and clear sky and to obtain typical cloud data without direct cloud detection. In an approach to adjust the classifier, data typical of certain conditions that lead to incorrect results are added to the training dataset. In this study, an adjustment procedure is proposed, which quantitatively judges, whether an addition is actually effective for reduction of the frequency of incorrect results. Another approach for the adjustment is improving feature space used for cloud detection. Indices as quantitative guidance to estimate whether an addition or elimination of a feature actually reduces the frequency of incorrect results can be obtained from the analysis of the support vectors. The cloud detection method incorporating the SVM is therefore able to integrate practical adjustment procedures. Applications of this method to Moderate Resolution Imaging Spectroradiometer (MODIS) data demonstrate that the concept of the method satisfies the requirements and the adjustability to various conditions can be realized. © 2017 Elsevier Inc." "52364778400;35772803100;56607323800;8255132900;6508003494;","Multilayer perceptron neural networks model for meteosat second generation SEVIRI daytime cloud masking",2015,"10.3390/rs70201529","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84928785522&doi=10.3390%2frs70201529&partnerID=40&md5=ef34b3ac61adf606a31ef53c4cf6a9d5","A multilayer perceptron neural network cloud mask for Meteosat Second Generation SEVIRI (Spinning Enhanced Visible and Infrared Imager) images is introduced and evaluated. The model is trained for cloud detection on MSG SEVIRI daytime data. It consists of a multi-layer perceptron with one hidden sigmoid layer, trained with the error back-propagation algorithm. The model is fed by six bands of MSG data (0.6, 0.8, 1.6, 3.9, 6.2 and 10.8 μm) with 10 hidden nodes. The multiple-layer perceptrons lead to a cloud detection accuracy of 88.96%, when trained to map two predefined values that classify cloud and clear sky. The network was further evaluated using sixty MSG images taken at different dates. The network detected not only bright thick clouds but also thin or less bright clouds. The analysis demonstrated the feasibility of using machine learning models of cloud detection in MSG SEVIRI imagery." "57193843574;56073123600;7401526171;7005052907;41761935400;12806771100;","Short-Term Precipitation Forecast Based on the PERSIANN System and LSTM Recurrent Neural Networks",2018,"10.1029/2018JD028375","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85056703759&doi=10.1029%2f2018JD028375&partnerID=40&md5=b4b67e14831344118752b577c2297cdb","Short-term Quantitative Precipitation Forecasting is important for flood forecasting, early flood warning, and natural hazard management. This study proposes a precipitation forecast model by extrapolating Cloud-Top Brightness Temperature (CTBT) using advanced Deep Neural Networks, and applying the forecasted CTBT into an effective rainfall retrieval algorithm to obtain the Short-term Quantitative Precipitation Forecasting (0–6 hr). To achieve such tasks, we propose a Long Short-Term Memory (LSTM) and the Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks (PERSIANN), respectively. The precipitation forecasts obtained from our proposed framework, (i.e., LSTM combined with PERSIANN) are compared with a Recurrent Neural Network (RNN), Persistency method, and Farneback optical flow each combined with PERSIANN algorithm and the numerical model results from the first version of Rapid Refresh (RAPv1.0) over three regions in the United States, including the states of Oregon, Oklahoma, and Florida. Our experiments indicate better statistics, such as correlation coefficient and root-mean-square error, for the CTBT forecasts from the proposed LSTM compared to the RNN, Persistency, and the Farneback method. The precipitation forecasts from the proposed LSTM and PERSIANN framework has demonstrated better statistics compared to the RAPv1.0 numerical forecasts and PERSIANN estimations from RNN, Persistency, and Farneback projections in terms of Probability of Detection, False Alarm Ratio, Critical Success Index, correlation coefficient, and root-mean-square error, especially in predicting the convective rainfalls. The proposed method shows superior capabilities in short-term forecasting over compared methods, and has the potential to be implemented globally as an alternative short-term forecast product. ©2018. American Geophysical Union. All Rights Reserved." "56428816500;","FORCE-Landsat + Sentinel-2 analysis ready data and beyond",2019,"10.3390/rs11091124","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065733301&doi=10.3390%2frs11091124&partnerID=40&md5=51b4e8644015d2f77e87978f8c84224a","Ever increasing data volumes of satellite constellations call for multi-sensor analysis ready data (ARD) that relieve users from the burden of all costly preprocessing steps. This paper describes the scientific software FORCE (Framework for Operational Radiometric Correction for Environmental monitoring), an 'all-in-one' solution for the mass-processing and analysis of Landsat and Sentinel-2 image archives. FORCE is increasingly used to support a wide range of scientific to operational applications that are in need of both large area, as well as deep and dense temporal information. FORCE is capable of generating Level 2 ARD, and higher-level products. Level 2 processing is comprised of state-of-the-art cloud masking and radiometric correction (including corrections that go beyondARDspecification, e.g., topographic or bidirectional reflectance distribution function correction). It further includes data cubing, i.e., spatial reorganization of the data into a non-overlapping grid system for enhanced efficiency and simplicity of ARD usage. However, the usage barrier of Level 2 ARD is still high due to the considerable data volume and spatial incompleteness of valid observations (e.g., clouds). Thus, the higher-level modules temporally condense multi-temporal ARD into manageable amounts of spatially seamless data. For data mining purposes, per-pixel statistics of clear sky data availability can be generated. FORCE provides functionality for compiling best-available-pixel composites and spectral temporal metrics, which both utilize all available observations within a defined temporal window using selection and statistical aggregation techniques, respectively. These products are immediately fit for common Earth observation analysis workflows, such as machine learning-based image classification, and are thus referred to as highly analysis ready data (hARD). FORCE provides data fusion functionality to improve the spatial resolution of (i) coarse continuous fields like land surface phenology and (ii) Landsat ARD using Sentinel-2 ARD as prediction targets. Quality controlled time series preparation and analysis functionality with a number of aggregation and interpolation techniques, land surface phenology retrieval, and change and trend analyses are provided. Outputs of this module can be directly ingested into a geographic information system (GIS) to fuel research questions without any further processing, i.e., hARD+. FORCE is open source software under the terms of the GNU General Public License v. > = 3, and can be downloaded from http://force.feut.de. © 2019 by the authors." "57193084560;57202233322;35743520500;7003824834;57202230109;7004291692;6601980015;","Estimation of vegetable crop parameter by multi-temporal UAV-borne images",2018,"10.3390/rs10050805","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047531441&doi=10.3390%2frs10050805&partnerID=40&md5=9ba751d1631a34b919970bc922210606","3D point cloud analysis of imagery collected by unmanned aerial vehicles (UAV) has been shown to be a valuable tool for estimation of crop phenotypic traits, such as plant height, in several species. Spatial information about these phenotypic traits can be used to derive information about other important crop characteristics, like fresh biomass yield, which could not be derived directly from the point clouds. Previous approaches have often only considered single date measurements using a single point cloud derived metric for the respective trait. Furthermore, most of the studies focused on plant species with a homogenous canopy surface. The aim of this study was to assess the applicability of UAV imagery for capturing crop height information of three vegetables (crops eggplant, tomato, and cabbage) with a complex vegetation canopy surface during a complete crop growth cycle to infer biomass. Additionally, the effect of crop development stage on the relationship between estimated crop height and field measured crop height was examined. Our study was conducted in an experimental layout at the University of Agricultural Science in Bengaluru, India. For all the crops, the crop height and the biomass was measured at five dates during one crop growth cycle between February and May 2017 (average crop height was 42.5, 35.5, and 16.0 cm for eggplant, tomato, and cabbage). Using a structure from motion approach, a 3D point cloud was created for each crop and sampling date. In total, 14 crop height metrics were extracted from the point clouds. Machine learning methods were used to create prediction models for vegetable crop height. The study demonstrates that the monitoring of crop height using an UAV during an entire growing period results in detailed and precise estimates of crop height and biomass for all three crops (R2 ranging from 0.87 to 0.97, bias ranging from -0.66 to 0.45 cm). The effect of crop development stage on the predicted crop height was found to be substantial (e.g., median deviation increased from 1% to 20% for eggplant) influencing the strength and consistency of the relationship between point cloud metrics and crop height estimates and, thus, should be further investigated. Altogether the results of the study demonstrate that point cloud generated fromUAV-based RGB imagery can be used to effectively measure vegetable crop biomass in larger areas (relative error = 17.6%, 19.7%, and 15.2% for eggplant, tomato, and cabbage, respectively) with a similar accuracy as biomass prediction models based on measured crop height (relative error = 21.6, 18.8, and 15.2 for eggplant, tomato, and cabbage). © 2018 by the authors." "7202244409;56579595600;57070621900;","Stacked Sparse Autoencoder Modeling Using the Synergy of Airborne LiDAR and Satellite Optical and SAR Data to Map Forest Above-Ground Biomass",2017,"10.1109/JSTARS.2017.2748341","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85030755958&doi=10.1109%2fJSTARS.2017.2748341&partnerID=40&md5=d01df9ad133728a85aaaf0f0f461c5f2","Timely, spatially complete, and reliable forest above-ground biomass (AGB) data are a prerequisite to support forest management and policy formulation. Traditionally, forest AGB is spatially estimated by integrating satellite images, in particular, optical data, with field plots from forest inventory programs. However, field data are limited in remote and unmanaged areas. In addition, optical reflectance usually saturates at high-density biomass level and is subject to cloud contaminations. Thus, this study aimed to develop a deep learning based workflow for mapping forest AGB by integrating Landsat 8 and Sentinel-1A images with airborne light detection and ranging (LiDAR) data. A reference AGB map was derived from the wall-to-wall LiDAR data and field measurements. The LiDAR plots - stratified random samples of forest biomass extracted from the LiDAR simulated strips in the reference map - were adopted as a surrogate for traditional field plots. In addition to the deep learning model, i.e., stacked sparse autoencoder network (SSAE), five different prediction techniques including multiple stepwise linear regressions, K-nearest neighbor, support vector machine, back propagation neural networks, and random forest were individually used to establish the relationship between LiDAR-derived forest biomass and the satellite predictors. Optical variables (Landsat 8 OLI), SAR variables (Sentinel-1A), and their combined variables were individually input to the six prediction models. Results showed that the SSAE model had the best performance for the forest biomass estimation. The combined optical and microwave dataset as explanatory variables improved the modeling performance compared to either the optical-only or microwave-only data, regardless of prediction algorithms. The best mapping accuracy was obtained by the SSAE model with inputs of optical and microwave integrated metrics that yielded R2 of 0.812, root mean squared error (RMSE) of 21.753 Mg/ha, and relative RMSE (RMSEr) of 14.457%. Overall, the SSAE model with inputs of combined Landsat 8 OLI and Sentinel-1A information could result in accurate estimation of forest biomass by using the stratification-sampled and LiDAR-derived AGB as ground reference data. The modeling workflow has the potential to promote future forest growth monitoring and carbon stock assessment across large areas. © 2008-2012 IEEE." "57207951736;7402291608;56414664700;","Big data driven lithium-ion battery modeling method based on SDAE-ELM algorithm and data pre-processing technology",2019,"10.1016/j.apenergy.2019.03.154","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063325218&doi=10.1016%2fj.apenergy.2019.03.154&partnerID=40&md5=9cecf10351ed051a7d050113bd33c45e","As one of the bottleneck technologies of electric vehicles (EVs), the battery hosts complex and hardly observable internal chemical reactions. Therefore, a precise mathematical model is crucial for the battery management system (BMS) to ensure the secure and stable operation of the battery in a multi-variable environment. First, a Cloud-based BMS (C-BMS) is established based on a database containing complete battery status information. Next, a data cleaning method based on machine learning is applied to the big data of batteries. Meanwhile, to improve the model stability under dynamic conditions, an F-divergence-based data distribution quality assessment method and a sampling-based data preprocess method is designed. Then, a lithium-ion battery temperature-dependent model is built based on Stacked Denoising Autoencoders- Extreme Learning Machine (SDAE-ELM) algorithm, and a new training method combined with data preprocessing is also proposed to improve the model accuracy. Finally, to improve reliability, a conjunction working mode between the C-BMS and the BMS in vehicles (V-BMS) is also proposed, providing as an applied case of the model. Using the battery data extracted from electric buses, the effectiveness and accuracy of the model are validated. The error of the estimated battery terminal voltage is within 2%, and the error of the estimated State of Charge (SoC) is within 3%. © 2019" "55249955400;25027558600;23491959900;56117362200;57200500772;57200498053;16069090700;","Deep Recurrent Neural Networks for Winter Vegetation Quality Mapping via Multitemporal SAR Sentinel-1",2018,"10.1109/LGRS.2018.2794581","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041386709&doi=10.1109%2fLGRS.2018.2794581&partnerID=40&md5=17dace48dd6941ba4f2087c83bf27566","Mapping winter vegetation quality is a challenging problem in remote sensing. This is due to cloud coverage in winter periods, leading to a more intensive use of radar rather than optical images. The aim of this letter is to provide a better understanding of the capabilities of Sentinel-1 radar images for winter vegetation quality mapping through the use of deep learning techniques. Analysis is carried out on a multitemporal Sentinel-1 data over an area around Charentes-Maritimes, France. This data set was processed in order to produce an intensity radar data stack from October 2016 to February 2017. Two deep recurrent neural network (RNN)-based classifiers were employed. Our work revealed that the results of the proposed RNN models clearly outperformed classical machine learning approaches (support vector machine and random forest). © 2004-2012 IEEE." "36701583900;57189619677;","A hybrid ICT-solution for smart meter data analytics",2016,"10.1016/j.energy.2016.05.068","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84996563893&doi=10.1016%2fj.energy.2016.05.068&partnerID=40&md5=5fb8114cab6c558cb7f574625753a0f3","Smart meters are increasingly used worldwide. Smart meters are the advanced meters capable of measuring energy consumption at a fine-grained time interval, e.g., every 15 min. Smart meter data are typically bundled with social economic data in analytics, such as meter geographic locations, weather conditions and user information, which makes the data sets very sizable and the analytics complex. Data mining and emerging cloud computing technologies make collecting, processing, and analyzing the so-called big data possible. This paper proposes an innovative ICT-solution to streamline smart meter data analytics. The proposed solution offers an information integration pipeline for ingesting data from smart meters, a scalable platform for processing and mining big data sets, and a web portal for visualizing analytics results. The implemented system has a hybrid architecture of using Spark or Hive for big data processing, and using the machine learning toolkit, MADlib, for doing in-database data analytics in PostgreSQL database. This paper evaluates the key technologies of the proposed ICT-solution, and the results show the effectiveness and efficiency of using the system for both batch and online analytics. © 2016 Elsevier Ltd" "38460962400;36131796000;57190733106;22035530600;","Earthquake prediction in California using regression algorithms and cloud-based big data infrastructure",2018,"10.1016/j.cageo.2017.10.011","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85032887044&doi=10.1016%2fj.cageo.2017.10.011&partnerID=40&md5=7ddaac61f9a22d59afc6d1f57b485042","Earthquake magnitude prediction is a challenging problem that has been widely studied during the last decades. Statistical, geophysical and machine learning approaches can be found in literature, with no particularly satisfactory results. In recent years, powerful computational techniques to analyze big data have emerged, making possible the analysis of massive datasets. These new methods make use of physical resources like cloud based architectures. California is known for being one of the regions with highest seismic activity in the world and many data are available. In this work, the use of several regression algorithms combined with ensemble learning is explored in the context of big data (1 GB catalog is used), in order to predict earthquakes magnitude within the next seven days. Apache Spark framework, H2O library in R language and Amazon cloud infrastructure were been used, reporting very promising results. © 2017 Elsevier Ltd" "55856164800;35743520500;57190380155;","A supervoxel-based spectro-spatial approach for 3D urban point cloud labelling",2016,"10.1080/01431161.2016.1211348","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84979555860&doi=10.1080%2f01431161.2016.1211348&partnerID=40&md5=8238271813111d2515cd941e3d0517f5","ABSTRACT: Three-dimensional (3D) point cloud labelling of airborne lidar (light detection and ranging) data has promising applications in urban city modelling. Automatic and efficient methods for semantic labelling of airborne urban point cloud data with multiple classes still remains a challenge. We propose a novel 3D object-based classification framework for labelling urban lidar point cloud using a computer vision technique, supervoxels. The supervoxel approach is promising for representing dense lidar point cloud in a compact manner for 3D segmentation and for improving the computational efficiency. Initially, supervoxels are generated by over-segmenting the coloured point cloud using the voxel-based cloud connectivity algorithm in the geometric space. The local connectivity established between supervoxels has been used to produce meaningful and realistic objects (segments). The segments are classified by different machine learning techniques based on several spectral and geometric features extracted from the segments. All the points within a labelled segment are assigned the same segment label. Furthermore, the effect of different feature vectors and varying point density on the classification accuracy has been studied. Results indicate an accurate labelling of points in realistic 3D space conforming to the boundaries of objects. An overall classification accuracy of 90% is achieved by the proposed method. The labelled 3D points can be used directly for the reconstruction of buildings and other man-made objects. © 2016 Informa UK Limited, trading as Taylor & Francis Group." "48561616000;42962601400;43361574900;57205898179;6701477064;7003939174;6603119574;7103102779;7102831991;","Autonomous gaussian decomposition",2015,"10.1088/0004-6256/149/4/138","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84926459445&doi=10.1088%2f0004-6256%2f149%2f4%2f138&partnerID=40&md5=e50fb9798c3a0375df16d9d03386fbfd","We present a new algorithm, named Autonomous Gaussian Decomposition (AGD), for automatically decomposing spectra into Gaussian components. AGD uses derivative spectroscopy and machine learning to provide optimized guesses for the number of Gaussian components in the data, and also their locations, widths, and amplitudes. We test AGD and find that it produces results comparable to human-derived solutions on 21 cm absorption spectra from the 21 cm SPectral line Observations of Neutral Gas with the EVLA (21-SPONGE) survey. We use AGD with Monte Carlo methods to derive the H i line completeness as a function of peak optical depth and velocity width for the 21-SPONGE data, and also show that the results of AGD are stable against varying observational noise intensity. The autonomy and computational efficiency of the method over traditional manual Gaussian fits allow for truly unbiased comparisons between observations and simulations, and for the ability to scale up and interpret the very large data volumes from the upcoming Square Kilometer Array and pathfinder telescopes. © 2015. The American Astronomical Society. All rights reserved.." "23491844400;56237235700;9036557400;56237086200;56224155200;56503083100;56037172400;","Detection of convective initiation using Meteorological Imager onboard Communication, Ocean, and Meteorological Satellite based on machine learning approaches",2015,"10.3390/rs70709184","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84937931173&doi=10.3390%2frs70709184&partnerID=40&md5=b791f16ac1bf0010993910b14f97c969","As convective clouds in Northeast Asia are accompanied by various hazards related with heavy rainfall and thunderstorms, it is very important to detect convective initiation (CI) in the region in order to mitigate damage by such hazards. In this study, a novel approach for CI detection using images from Meteorological Imager (MI), a payload of the Communication, Ocean, and Meteorological Satellite (COMS), was developed by improving the criteria of the interest fields of Rapidly Developing Cumulus Areas (RDCA) derivation algorithm, an official CI detection algorithm for Multi-functional Transport SATellite-2 (MTSAT-2), based on three machine learning approaches-decision trees (DT), random forest (RF), and support vector machines (SVM). CI was defined as clouds within a 16 × 16 km window with the first detection of lightning occurrence at the center. A total of nine interest fields derived from visible, water vapor, and two thermal infrared images of MI obtained 15-75 min before the lightning occurrence were used as input variables for CI detection. RF produced slightly higher performance (probability of detection (POD) of 75.5% and false alarm rate (FAR) of 46.2%) than DT (POD of 70.7% and FAR of 46.6%) for detection of CI caused by migrating frontal cyclones and unstable atmosphere. SVM resulted in relatively poor performance with very high FAR ~83.3%. The averaged lead times of CI detection based on the DT and RF models were 36.8 and 37.7 min, respectively. This implies that CI over Northeast Asia can be forecasted ~30-45 min in advance using COMS MI data. © 2015 by the authors." "7405490236;","Combining hyperspectral and lidar data for vegetation mapping in the Florida everglades",2014,"10.14358/PERS.80.8.733","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84924225421&doi=10.14358%2fPERS.80.8.733&partnerID=40&md5=36a8afb73cfc26cc068df4397c229bd0","This study explored a combination of hyperspectral and lidar systems for vegetation mapping in the Florida Everglades. A framework was designed to integrate two remotely sensed datasets and four data processing techniques. Lidar elevation and intensity features were extracted from the original point cloud data to avoid the errors and uncertainties in the raster-based lidar methods. Lidar significantly increased the classification accuracy compared with the application of hyperspectral data alone. Three lidar-derived features (elevation, intensity, and topography) had the same contributions in the classification. A synergy of hyperspectral imagery with all lidar-derived features achieved the best result with an overall accuracy of 86 percent and a Kappa value of 0.82 based on an ensemble analysis of three machine learning classifiers. Ensemble analysis did not significantly increase the classification accuracy, but it provided a complementary uncertainty map for the final classified map. The study shows the promise of the synergy of hyperspectral and lidar systems for mapping complex wetlands. © 2014 American Society for Photogrammetry and Remote Sensing." "36668835800;7004452985;6603861412;","Diurnal variation of the global electric circuit from clustered thunderstorms",2014,"10.1002/2013JA019593","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84897589554&doi=10.1002%2f2013JA019593&partnerID=40&md5=3200e15a31e9e888f907fe7b91305e2e","The diurnal variation of the global electric circuit is investigated using the World Wide Lightning Location Network (WWLLN), which has been shown to identify nearly all thunderstorms (using WWLLN data from 2005). To create an estimate of global electric circuit activity, a clustering algorithm is applied to the WWLLN data set to identify global thunderstorms from 2010 to 2013. Annual, seasonal, and regional thunderstorm activity is investigated in this new WWLLN thunderstorm data set in order to estimate the source behavior of the global electric circuit. Through the clustering algorithm, the total number of active thunderstorms are counted every 30 min creating a measure of the global electric circuit source function. The thunderstorm clusters are compared to precipitation radar data from the Tropical Rainfall Measurement Mission satellite and with case studies of thunderstorm evolution. The clustering algorithm reveals an average of 660±70 thunderstorms active at any given time with a peak-to-peak variation of 36%. The highest number of thunderstorms occurs in November (720±90), and the lowest number occurs in January (610±80). Thunderstorm cluster and electrified storm cloud activity are combined with thunderstorm overflight current measurements to estimate the global electric circuit thunderstorm contribution current to be 1090±70 A with a variation of 24%. By utilizing the global coverage and high time resolution of WWLLN, the total active thunderstorm count and current is shown to be less than previous estimates based on compiled climatologies. Key Points Clustering algorithms are applied to lightning data to locate thunderstorms Global electric circuit thunderstorm current is estimated from WWLLN A global average of 660 thunderstorms are estimated to produce 1090 A ©2014. American Geophysical Union. All Rights Reserved." "55962329600;7401758587;","Spatiotemporal image fusion in remote sensing",2019,"10.3390/rs11070818","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064013012&doi=10.3390%2frs11070818&partnerID=40&md5=c4df3791067b16bc307801fabf663b67","In this paper, we discuss spatiotemporal data fusion methods in remote sensing. These methods fuse temporally sparse fine-resolution images with temporally dense coarse-resolution images. This review reveals that existing spatiotemporal data fusion methods are mainly dedicated to blending optical images. There is a limited number of studies focusing on fusing microwave data, or on fusing microwave and optical images in order to address the problem of gaps in the optical data caused by the presence of clouds. Therefore, future efforts are required to develop spatiotemporal data fusion methods flexible enough to accomplish different data fusion tasks under different environmental conditions and using different sensors data as input. The review shows that additional investigations are required to account for temporal changes occurring during the observation period when predicting spectral reflectance values at a fine scale in space and time. More sophisticated machine learning methods such as convolutional neural network (CNN) represent a promising solution for spatiotemporal fusion, especially due to their capability to fuse images with different spectral values. © 2019 by the authors." "57192387537;57190963447;56203143700;23036019000;23006934800;7005742190;35611187000;7004609788;","Fast cloud segmentation using convolutional neural networks",2018,"10.3390/rs10111782","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057086755&doi=10.3390%2frs10111782&partnerID=40&md5=7d2ca03a9cf9c3a7863586d39e356613","Information about clouds is important for observing and predicting weather and climate as well as for generating and distributing solar power. Most existing approaches extract cloud information from satellite data by classifying individual pixels instead of using closely integrated spatial information, ignoring the fact that clouds are highly dynamic, spatially continuous entities. This paper proposes a novel cloud classification method based on deep learning. Relying on a Convolutional Neural Network (CNN) architecture for image segmentation, the presented Cloud Segmentation CNN (CS-CNN), classifies all pixels of a scene simultaneously rather than individually. We show that CS-CNN can successfully process multispectral satellite data to classify continuous phenomena such as highly dynamic clouds. The proposed approach produces excellent results on Meteosat Second Generation (MSG) satellite data in terms of quality, robustness, and runtime compared to other machine learning methods such as random forests. In particular, comparing CS-CNN with the CLAAS-2 cloud mask derived from MSG data shows high accuracy (0.94) and Heidke Skill Score (0.90) values. In contrast to a random forest, CS-CNN produces robust results and is insensitive to challenges created by coast lines and bright (sand) surface areas. Using GPU acceleration, CS-CNN requires only 25 ms of computation time for classification of images of Europe with 508 × 508 pixels. © 2018 by the authors." "35303493300;25644982200;22034344400;","Machine learning search for variable stars",2018,"10.1093/mnras/stx3222","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046107561&doi=10.1093%2fmnras%2fstx3222&partnerID=40&md5=f5e1954a63380d8e5d670d3b09beea03","Photometric variability detection is often considered as a hypothesis testing problem: an object is variable if the null hypothesis that its brightness is constant can be ruled out given the measurements and their uncertainties. The practical applicability of this approach is limited by uncorrected systematic errors. We propose a new variability detection technique sensitive to a wide range of variability types while being robust to outliers and underestimated measurement uncertainties. We consider variability detection as a classification problem that can be approached with machine learning. Logistic Regression (LR), Support Vector Machines (SVM), k Nearest Neighbours (kNN), Neural Nets (NN), Random Forests (RF), and Stochastic Gradient Boosting classifier (SGB) are applied to 18 features (variability indices) quantifying scatter and/or correlation between points in a light curve. We use a subset of Optical Gravitational Lensing Experiment phase two (OGLE-II) Large Magellanic Cloud (LMC) photometry (30 265 light curves) that was searched for variability using traditional methods (168 known variable objects) as the training set and then apply the NN to a new test set of 31 798 OGLE-II LMC light curves. Among 205 candidates selected in the test set, 178 are real variables, while 13 low-amplitude variables are new discoveries. The machine learning classifiers considered are found to be more efficient (select more variables and fewer false candidates) compared to traditional techniques using individual variability indices or their linear combination. The NN, SGB, SVM, and RF show a higher efficiency compared to LR and kNN. © 2018 The Author(s)." "13104305800;36561227300;56582279100;55850416000;57194786552;7005262903;","Forecasting the ongoing invasion of Lagocephalus sceleratus in the Mediterranean Sea",2018,"10.1016/j.ecolmodel.2018.01.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041452458&doi=10.1016%2fj.ecolmodel.2018.01.007&partnerID=40&md5=6bbba2c1831fe05980a52dfbb56d39eb","Invasive species from the Suez Canal, also named “Lessepsian species”, often have an ecological and financial impact on marine life, fisheries, human well-being and health in the Mediterranean Sea. Among these, the silver-cheeked toad-fish Lagocephalus sceleratus (Gmelin, 1789) has rapidly colonised the eastern Mediterranean basin and is currently moving westwards. This pufferfish has a highly opportunistic behaviour, it attacks fish captured in nets and lines and seriously damages fishing gears and catch. It is a highly-toxic species with no immediate economic value for the Mediterranean market, although it currently represents 4% of the weight of the total artisanal catches. Consequently, the possible effects on Mediterranean fisheries and health require to enhance our understanding about the future geographical distribution of this pufferfish in the whole basin. In this paper, an overall habitat suitability map and an effective geographical spread map for L. sceleratus at Mediterranean scale are produced by using cloud computing-based algorithms to merge seven machine learning approaches. Further, the potential impact of the species is estimated for several Mediterranean Sea subdivisions: The major fishing areas of the Food and Agriculture Organization of the United Nations, the Economic Exclusive Zones, and the subdivisions of the General Fisheries Commission for the Mediterranean Sea. Our results suggest that without an intervention, L. sceleratus will continue its rapid spread and will likely have a high impact on fisheries. The presented method is generic and can be applied to other invasive species. It is based on an Open Science approach and all processes are freely available as Web services. © 2018 The Author(s)" "56135196400;7403872687;6507378331;7005052907;7401526171;","Precipitation identification with bispectral satellite information using deep learning approaches",2017,"10.1175/JHM-D-16-0176.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85019030320&doi=10.1175%2fJHM-D-16-0176.1&partnerID=40&md5=558e83622cfe32be2e28350847eda415","In the development of a satellite-based precipitation product, two important aspects are sufficient precipitation information in the satellite-input data and proper methodologies, which are used to extract such information and connect it to precipitation estimates. In this study, the effectiveness of the state-of-the-art deep learning (DL) approaches to extract useful features from bispectral satellite information, infrared (IR), and water vapor (WV) channels, and to produce rain/no-rain (R/NR) detection is explored. To verify the methodologies, two models are designed and evaluated: the first model, referred to as the DL-IR only method, applies deep learning approaches to the IR data only; the second model, referred to as the DL-IR+WV method, incorporates WV data to further improve the precipitation identification performance. The radar stage IV data are the reference data used as ground observation. The operational product, Precipitation Estimation from Remotely Sensed Information Using Artificial Neural Networks-Cloud Classification System (PERSIANN-CCS), serves as a baseline model with which to compare the performances. The experiments show significant improvement for both models in R/NR detection. The overall performance gains in the critical success index (CSI) are 21.60% and 43.66% over the verification periods for the DL-IR only model and the DL-IR+WV model compared to PERSIANN-CCS, respectively. In particular, the performance gains in CSI are as high as 46.51% and 94.57% for the models for the winter season. Moreover, specific case studies show that the deep learning techniques and the WV channel information effectively help recover a large number of missing precipitation pixels under warm clouds while reducing false alarms under cold clouds. © 2017 American Meteorological Society." "57069455200;36064917000;55500134600;55972035800;55292994700;57189502828;","Discriminative-Dictionary-Learning-Based Multilevel Point-Cluster Features for ALS Point-Cloud Classification",2016,"10.1109/TGRS.2016.2599163","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85027405929&doi=10.1109%2fTGRS.2016.2599163&partnerID=40&md5=36584dd7e642cc8a128538dd41f72da3","Efficient presentation and recognition of on-ground objects from airborne laser scanning (ALS) point clouds are a challenging task. In this paper, we propose an approach that combines a discriminative-dictionary-learning-based sparse coding and latent Dirichlet allocation (LDA) to generate multilevel point-cluster features for ALS point-cloud classification. Our method takes advantage of the labels of training data and each dictionary item to enforce discriminability in sparse coding during the dictionary learning process and more accurately further represent point-cluster features. The multipath AdaBoost classifiers with the hierarchical point-cluster features are trained, and we apply them to the classification of unknown points by the heritance of the recognition results under different paths. Experiments are performed on different ALS point clouds; the experimental results have shown that the extracted point-cluster features combined with the multipath classifiers can significantly enhance the classification accuracy, and they have demonstrated the superior performance of our method over other techniques in point-cloud classification. © 2016 IEEE." "57194545411;35113156800;57194542466;55885020100;","SMART POINT CLOUD: DEFINITION and REMAINING CHALLENGES",2016,"10.5194/isprs-annals-IV-2-W1-119-2016","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85030697065&doi=10.5194%2fisprs-annals-IV-2-W1-119-2016&partnerID=40&md5=8a0a77850ac2d8fb8eb4fc72fff84180","Dealing with coloured point cloud acquired from terrestrial laser scanner, this paper identifies remaining challenges for a new data structure: the smart point cloud. This concept arises with the statement that massive and discretized spatial information from active remote sensing technology is often underused due to data mining limitations. The generalisation of point cloud data associated with the heterogeneity and temporality of such datasets is the main issue regarding structure, segmentation, classification, and interaction for an immediate understanding. We propose to use both point cloud properties and human knowledge through machine learning to rapidly extract pertinent information, using user-centered information (smart data) rather than raw data. A review of feature detection, machine learning frameworks and database systems indexed both for mining queries and data visualisation is studied. Based on existing approaches, we propose a new 3-block flexible framework around device expertise, analytic expertise and domain base reflexion. This contribution serves as the first step for the realisation of a comprehensive smart point cloud data structure." "7202467963;36676931900;7501959001;","Smart Information Reconstruction via Time-Space-Spectrum Continuum for Cloud Removal in Satellite Images",2015,"10.1109/JSTARS.2015.2400636","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85027921302&doi=10.1109%2fJSTARS.2015.2400636&partnerID=40&md5=1dfc64395036ea147f0c8c790bf8d447","Cloud contamination is a big obstacle when processing satellite images retrieved from visible and infrared spectral ranges for application. Although computational techniques including interpolation and substitution have been applied to recover missing information caused by cloud contamination, these algorithms are subject to many limitations. In this paper, a novel smart information reconstruction (SMIR) method is proposed, in order to reconstruct cloud contaminated pixel values from the time-space-spectrum continuum with the aid of a machine learning tool, namely extreme learning machine (ELM). For the purpose of demonstration, the performance of SMIR is evaluated by reconstructing the missing remote sensing reflectance values derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) on board the Terra satellite over Lake Nicaragua, where is a very cloudy area year round. For comparison, the traditional backpropagation neural network algorithms will also be implemented to reconstruct the missing values. Experimental results show that the ELM outperforms the BP algorithms by an enhanced machine learning capacity with simulated memory effect embedded in MODIS due to linking the complex time-space-spectrum continuum between cloud-free and cloudy pixels. The ELM-based SMIR practice presents a correlation coefficient of 0.88 with root mean squared error of 7.4 E-04 sr-1 between simulated and observed reflectance values. Finding suggests that the SMIR method is effective to reconstruct all the missing information providing visually logical and quantitatively assured images for further image processing and interpretation in environmental applications. © 2015 IEEE." "36920327800;15050899300;57214847722;","Processing tree point clouds using Gaussian Mixture Models",2013,"10.5194/isprsannals-II-5-W2-43-2013","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046681848&doi=10.5194%2fisprsannals-II-5-W2-43-2013&partnerID=40&md5=3181027f285ca7eaa521e918c6865ee1","While traditionally used for surveying and photogrammetric fields, laser scanning is increasingly being used for a wider range of more general applications. In addition to the issues typically associated with processing point data, such applications raise a number of new complications, such as the complexity of the scenes scanned, along with the sheer volume of data. Consequently, automated procedures are required for processing, and analysing such data. This paper introduces a method for modelling multi-modal, geometrically complex objects in terrestrial laser scanning point data; specifically, the modelling of trees. The model method comprises a number of geometric features in conjunction with a multi-modal machine learning technique. The model can then be used for contextually dependent region growing through separating the tree into its component part at the point level. Subsequently object analysis can be performed, for example, performing volumetric analysis of a tree by removing points associated with leaves. The workflow for this process is as follows: isolate individual trees within the scanned scene, train a Gaussian mixture model (GMM), separate clusters within the mixture model according to exemplar points determined by the GMM, grow the structure of the tree, and then perform volumetric analysis on the structure." "57189686942;36610817400;57190371939;36779146200;57216843548;7003505161;","Google Earth Engine for geo-big data applications: A meta-analysis and systematic review",2020,"10.1016/j.isprsjprs.2020.04.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084935887&doi=10.1016%2fj.isprsjprs.2020.04.001&partnerID=40&md5=cc4f9f853572e272dd7fe6cc7030316d","Google Earth Engine (GEE) is a cloud-based geospatial processing platform for large-scale environmental monitoring and analysis. The free-to-use GEE platform provides access to (1) petabytes of publicly available remote sensing imagery and other ready-to-use products with an explorer web app; (2) high-speed parallel processing and machine learning algorithms using Google's computational infrastructure; and (3) a library of Application Programming Interfaces (APIs) with development environments that support popular coding languages, such as JavaScript and Python. Together these core features enable users to discover, analyze and visualize geospatial big data in powerful ways without needing access to supercomputers or specialized coding expertise. The development of GEE has created much enthusiasm and engagement in the remote sensing and geospatial data science fields. Yet after a decade since GEE was launched, its impact on remote sensing and geospatial science has not been carefully explored. Thus, a systematic review of GEE that can provide readers with the “big picture” of the current status and general trends in GEE is needed. To this end, the decision was taken to perform a meta-analysis investigation of recent peer-reviewed GEE articles focusing on several features, including data, sensor type, study area, spatial resolution, application, strategy, and analytical methods. A total of 349 peer-reviewed articles published in 146 different journals between 2010 and October 2019 were reviewed. Publications and geographical distribution trends showed a broad spectrum of applications in environmental analyses at both regional and global scales. Remote sensing datasets were used in 90% of studies while 10% of the articles utilized ready-to-use products for analyses. Optical satellite imagery with medium spatial resolution, particularly Landsat data with an archive exceeding 40 years, has been used extensively. Linear regression and random forest were the most frequently used algorithms for satellite imagery processing. Among ready-to-use products, the normalized difference vegetation index (NDVI) was used in 27% of studies for vegetation, crop, land cover mapping and drought monitoring. The results of this study confirm that GEE has and continues to make substantive progress on global challenges involving process of geo-big data. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "35235984200;22134685800;53980452600;56536265500;","Mapping Monthly Air Temperature in the Tibetan Plateau from MODIS Data Based on Machine Learning Methods",2018,"10.1109/JSTARS.2017.2787191","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041667569&doi=10.1109%2fJSTARS.2017.2787191&partnerID=40&md5=36ae12b9a5b7584e6b45c6757518d361","Detailed knowledge of air temperature (Ta) is desired for various scientific applications. However, in the Tibetan Plateau (TP), the meteorologically observed Ta is limited due to the low density and uneven distribution of stations. This paper aims to develop a 1-km resolution monthly mean Ta dataset over the TP during 2001-2015 from remote sensing and auxiliary data. 11 environmental variables were extracted from Moderate Resolution Imaging Spectroradiometer (MODIS) data, Shuttle Radar Topography Mission (SRTM) digital elevation model (DEM) data and topographic index data. Ten machine learning algorithms were implemented and compared to determine the optimal model for Ta estimation in the TP. The Cubist algorithm outperformed other methods, having the highest accuracy and the lowest sensitivity to cloud contamination. To minimize the overfitting problem, a simple forward variable selection method was introduced and six variables were selected from the original 11 environmental variables. Among these six variables, nighttime land surface temperature (Ts) was the most important predictor, followed by elevation and solar radiance. The seasonal performance of the Cubist model was also assessed. The model had good accuracies in all four seasons, with the highest accuracy in winter (R2 = 0.98 and MAE = 0.63 °C) and the lowest accuracy in summer (R2 = 0.91 and MAE = 0.86 °C). Due to the gaps in MODIS data caused by cloud cover, there were 0.39% missing values in the estimated Ta. To improve the data integrity, Delaunay triangulation interpolation was applied to fill the missing Ta values. The final monthly (2001-2015) Ta dataset had an overall accuracy of RMSE = 1.00 °C and MAE = 0.73 °C. It provides valuable information for climate change assessment and other environmental studies in the TP. © 2018 IEEE." "56780482500;6602890253;7404062492;6701762451;","Improved identification of primary biological aerosol particles using single-particle mass spectrometry",2017,"10.5194/acp-17-7193-2017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85020869635&doi=10.5194%2facp-17-7193-2017&partnerID=40&md5=3403ae8973e712bfa0c458a73d7000c9","Measurements of primary biological aerosol particles (PBAP), especially at altitudes relevant to cloud formation, are scarce. Single-particle mass spectrometry (SPMS) has been used to probe aerosol chemical composition from ground and aircraft for over 20 years. Here we develop a method for identifying bioaerosols (PBAP and particles containing fragments of PBAP as part of an internal mixture) using SPMS. We show that identification of bioaerosol using SPMS is complicated because phosphorus-bearing mineral dust and phosphorus-rich combustion by-products such as fly ash produce mass spectra with peaks similar to those typically used as markers for bioaerosol. We have developed a methodology to differentiate and identify bioaerosol using machine learning statistical techniques applied to mass spectra of known particle types. This improved method provides far fewer false positives compared to approaches reported in the literature. The new method was then applied to two sets of ambient data collected at Storm Peak Laboratory and a forested site in Central Valley, California to show that 0.04-2ĝ€% of particles in the 200-3000ĝ€nm aerodynamic diameter range were identified as bioaerosol. In addition, 36-56ĝ€% of particles identified as biological also contained spectral features consistent with mineral dust, suggesting internal dust-biological mixtures. © Author(s) 2017." "57116856400;55642382500;6507337126;14825445000;","Mapping cultural ecosystem services 2.0 – Potential and shortcomings from unlabeled crowd sourced images",2019,"10.1016/j.ecolind.2018.08.035","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85053817844&doi=10.1016%2fj.ecolind.2018.08.035&partnerID=40&md5=2f5e4f1e24c333c20998c442c80f3bc7","The volume of accessible geotagged crowdsourced photos has increased. Such data include spatial, temporal, and thematic information on recreation and outdoor activities, thus can be used to quantify the demand for cultural ecosystem services (CES). So far photo content has been analyzed based on user-labeled tags or the manual labeling of photos. Both approaches are challenged with respect to consistency and cost-efficiency, especially for large-scale studies with an enormous volume of photos. In this study, we aim at developing a new method to analyze the content of large volumes of photos and to derive indicators of socio-cultural usage of landscapes. The method uses machine-learning and network analysis to identify clusters of photo content that can be used as an indicator of cultural services provided by landscapes. The approach was applied in the Mulde river basin in Saxony, Germany. All public Flickr photos (n = 12,635) belonging to the basin were tagged by deep convolutional neural networks through a cloud computing platform, Clarifai. The machine-predicted tags were analyzed by a network analysis that leads to nine hierarchical clusters. Those clusters were used to distinguish between photos related to CES (65%) and not related to CES (35%). Among the nine clusters, two clusters were related to CES: ‘landscape aesthetics’ and ‘existence’. This step allowed mapping of different aspects of CES and separation of non-relevant photos from further analysis. We further analyzed the impact of protected areas on the spatial pattern of CES and not-related CES photos. The presence of protected areas had a significant positive impact on the areas with both ‘landscape aesthetics’ and ‘existence’ photos: the total number of days in each mapping unit where at least one photo was taken by a user (‘photo-user-day’) increased with the share of protected areas around the location. The presented approach has shown its potential for reliable mapping of socio-cultural uses of landscapes. It is expected to scale well with large numbers of photos and to be easily transferable to different regions. © 2018 Elsevier Ltd" "7003992813;57006485700;55927897800;56082363600;26325493600;56540003600;55257235200;7003922988;56513856300;","Assessment of classifiers and remote sensing features of hyperspectral imagery and stereo-photogrammetric point clouds for recognition of tree species in a forest area of high species diversity",2018,"10.3390/rs10050714","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047571087&doi=10.3390%2frs10050714&partnerID=40&md5=3618d4d0734d5faa482bce0efbf9d9fd","Recognition of tree species and geospatial information on tree species composition is essential for forest management. In this study, tree species recognition was examined using hyperspectral imagery from visible to near-infrared (VNIR) and short-wave infrared (SWIR) camera sensors in combination with a 3D photogrammetric canopy surface model based on RGB camera stereo-imagery. An arboretum with a diverse selection of 26 tree species from 14 genera was used as a test area. Aerial hyperspectral imagery and high spatial resolution photogrammetric color imagery were acquired from the test area using unmanned aerial vehicle (UAV) borne sensors. Hyperspectral imagery was processed to calibrated reflectance mosaics and was tested along with the mosaics based on original image digital number values (DN). Two alternative classifiers, a k nearest neighbor method (k-nn), combined with a genetic algorithm and a random forest method, were tested for predicting the tree species and genus, as well as for selecting an optimal set of remote sensing features for this task. The combination of VNIR, SWIR, and 3D features performed better than any of the data sets individually. Furthermore, the calibrated reflectance values performed better compared to uncorrected DN values. These trends were similar with both tested classifiers. Of the classifiers, the k-nn combined with the genetic algorithm provided consistently better results than the random forest algorithm. The best result was thus achieved using calibrated reflectance features from VNIR and SWIR imagery together with 3D point cloud features; the proportion of correctly-classified trees was 0.823 for tree species and 0.869 for tree genus. © 2018 by the authors." "57192947904;6603354695;6603888005;","Convolutional neural networks for multispectral image cloud masking",2017,"10.1109/IGARSS.2017.8127438","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041842432&doi=10.1109%2fIGARSS.2017.8127438&partnerID=40&md5=f96bfae6583c3a225d194aedfb69ca1b","Convolutional neural networks (CNN) have proven to be state of the art methods for many image classification tasks and their use is rapidly increasing in remote sensing problems. One of their major strengths is that, when enough data is available, CNN perform an end-to-end learning without the need of custom feature extraction methods. In this work, we study the use of different CNN architectures for cloud masking of Proba-V multispectral images. We compare such methods with the more classical machine learning approach based on feature extraction plus supervised classification. Experimental results suggest that CNN are a promising alternative for solving cloud masking problems. © 2017 IEEE." "57196960023;21934985600;54793961000;57202621598;56646235100;8314173300;","Object-based classification of terrestrial laser scanning point clouds for landslide monitoring",2017,"10.1111/phor.12215","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85039073355&doi=10.1111%2fphor.12215&partnerID=40&md5=90b2267a37db8161fee1421d49aeb2c3","Terrestrial laser scanning (TLS) is often used to monitor landslides and other gravitational mass movements with high levels of geometric detail and accuracy. However, unstructured TLS point clouds lack semantic information, which is required to geomorphologically interpret the measured changes. Extracting meaningful objects in a complex and dynamic environment is challenging due to the objects' fuzziness in reality, as well as the variability and ambiguity of their patterns in a morphometric feature space. This work presents a point-cloud-based approach for classifying multitemporal scenes of a hillslope affected by shallow landslides. The 3D point clouds are segmented into morphologically homogeneous and spatially connected parts. These segments are classified into seven target classes (scarp, eroded area, deposit, rock outcrop and different classes of vegetation) in a two-step procedure: a supervised classification step with a machine-learning classifier using morphometric features, followed by a correction step based on topological rules. This improves the final object extraction considerably. © 2017 The Authors. The Photogrammetric Record © 2017 The Remote Sensing and Photogrammetry Society and John Wiley & Sons Ltd." "57204179498;7003494809;","Large area cropland extent mapping with Landsat data and a generalized classifier",2018,"10.1016/j.rse.2018.09.025","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054886361&doi=10.1016%2fj.rse.2018.09.025&partnerID=40&md5=dc45cc0e139c1e404360b31eaa0012ac","Accurate and up-to-date cropland maps play an important role in the study of food security. Traditional mapping of croplands using medium resolution (10–100 m) remote sensing imagery involving a “one-time, one-place” approach requires significant computing and labor resources. Although high mapping accuracies can be achieved using this approach, it is tedious and expensive to collect reference information to train the classifiers at each location and to apply over large areas, such as a continent. Moreover, large area cropland mapping presents additional challenges including a wide range of agricultural management practices, climatic conditions, and crop types. To overcome these challenges, here we report on a generalized image classifier to map cropland extent, which builds a classification model using training data from one location and time period, applied to other times and locations without the need for additional training data. The study was demonstrated across eight agro-ecological zones (AEZs) in Europe, the Middle East and North Africa using Landsat data acquired between 2009 and 2011. To reduce between-scene variability associated with image availability and cloud cover, input data were reduced to salient temporal statistics derived from enhanced vegetation index (EVI) combined with topographic variables. The generalized classifier was then tested across three levels of generalization: 1. individual - where training data were extracted from and applied to the same Landsat footprint; 2. AEZ where training data were extracted from a set of Landsat footprints within an AEZ and applied to any other Landsat footprint in the same AEZ; and 3. regional where training data were extracted from a set of Landsat footprints in the whole study area and applied to any other Landsat footprint inside the study area. Results showed that the generalized classifier is successful in identifying and mapping croplands with comparable success across all three levels of generalization with minimal cost: average loss in accuracy (as measured by overall accuracy) from the individual level (average overall accuracy of 80 ± 5%) to regional level (average overall accuracy of 74 ± 10%) is between 2 and 10% depending on the location. Results also show that generalization is not sensitive to the choice of the classification algorithm – the Linear Discriminant Analysis (LDA) model performs equally well compared to many popular machine learning algorithms found in the literature. This work suggests the generalization/signature extension framework has a great potential for rapid identification and mapping of croplands with reasonable accuracies over large areas using only easily computed vegetation indices with very little user input and ground information requirement. © 2018" "36024017200;56897235200;35221474100;57195399685;14522552500;","Ensemble classification of individual Pinus crowns from multispectral satellite imagery and airborne LiDAR",2018,"10.1016/j.jag.2017.09.016","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85036561343&doi=10.1016%2fj.jag.2017.09.016&partnerID=40&md5=943f3499dfc551e07a4bf47e53e6ea94","Distinguishing tree species is relevant in many contexts of remote sensing assisted forest inventory. Accurate tree species maps support management and conservation planning, pest and disease control and biomass estimation. This study evaluated the performance of applying ensemble techniques with the goal of automatically distinguishing Pinus sylvestris L. and Pinus uncinata Mill. Ex Mirb within a 1.3 km2 mountainous area in Barcelonnette (France). Three modelling schemes were examined, based on: (1) high-density LiDAR data (160 returns m−2), (2) Worldview-2 multispectral imagery, and (3) Worldview-2 and LiDAR in combination. Variables related to the crown structure and height of individual trees were extracted from the normalized LiDAR point cloud at individual-tree level, after performing individual tree crown (ITC) delineation. Vegetation indices and the Haralick texture indices were derived from Worldview-2 images and served as independent spectral variables. Selection of the best predictor subset was done after a comparison of three variable selection procedures: (1) Random Forests with cross validation (AUCRFcv), (2) Akaike Information Criterion (AIC) and (3) Bayesian Information Criterion (BIC). To classify the species, 9 regression techniques were combined using ensemble models. Predictions were evaluated using cross validation and an independent dataset. Integration of datasets and models improved individual tree species classification (True Skills Statistic, TSS; from 0.67 to 0.81) over individual techniques and maintained strong predictive power (Relative Operating Characteristic, ROC = 0.91). Assemblage of regression models and integration of the datasets provided more reliable species distribution maps and associated tree-scale mapping uncertainties. Our study highlights the potential of model and data assemblage at improving species classifications needed in present-day forest planning and management. © 2017 Elsevier B.V." "35784422100;9036557400;56402112700;35240287700;","Spatiotemporal downscaling approaches for monitoring 8-day 30 m actual evapotranspiration",2017,"10.1016/j.isprsjprs.2017.02.006","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85013347183&doi=10.1016%2fj.isprsjprs.2017.02.006&partnerID=40&md5=c53991d93b0b076d71535ff6a0a76a1f","Continuous monitoring of actual evapotranspiration (ET) is critical for water resources management at both regional and local scales. Although the MODIS ET product (MOD16A2) provides viable sources for ET monitoring at 8-day intervals, the spatial resolution (1 km) is too coarse for local scale applications. In this study, we propose a machine learning and spatial temporal fusion (STF)-integrated approach in order to generate 8-day 30 m ET based on both MOD16A2 and Landsat 8 data with three schemes. Random forest machine learning was used to downscale MODIS 1 km ET to 30 m resolution based on nine Landsat-derived indicators including vegetation indices (VIs) and land surface temperature (LST). STF-based models including Spatial and Temporal Adaptive Reflectance Fusion Model and Spatio-Temporal Image Fusion Model were used to derive synthetic Landsat surface reflectance (scheme 1)/VIs (scheme 2)/ET (scheme 3) on Landsat-unavailable dates. The approach was tested over two study sites in the United States. The results showed that fusion of Landsat VIs produced the best accuracy of predicted ET (R2 = 0.52–0.97, RMSE = 0.47–3.0 mm/8 days and rRMSE = 6.4–37%). High density of cloud-clear Landsat image acquisitions and low spatial heterogeneity of Landsat VIs benefit the ET prediction. The downscaled 30 m ET had good agreement with MODIS ET (RMSE = 0.42–3.4 mm/8 days, rRMSE = 3.2–26%). Comparison with the in situ ET measurements showed that the downscaled ET had higher accuracy than MODIS ET. © 2017 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57189617625;55577971100;57206859426;","Cloud extraction from Chinese high resolution satellite imagery by probabilistic latent semantic analysis and object-based machine learning",2016,"10.3390/rs8110963","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85017650513&doi=10.3390%2frs8110963&partnerID=40&md5=b847e2ad4b4d283412a540fd7149efaf","Automatic cloud extraction from satellite imagery is a vital process for many applications in optical remote sensing since clouds can locally obscure the surface features and alter the reflectance. Clouds can be easily distinguished by the human eyes in satellite imagery via remarkable regional characteristics, but finding a way to automatically detect various kinds of clouds by computer programs to speed up the processing efficiency remains a challenge. This paper introduces a new cloud detection method based on probabilistic latent semantic analysis (PLSA) and object-based machine learning. The method begins by segmenting satellite images into superpixels by Simple Linear Iterative Clustering (SLIC) algorithm while also extracting the spectral, texture, frequency and line segment features. Then, the implicit information in each superpixel is extracted from the feature histogram through the PLSA model by which the descriptor of each superpixel can be computed to form a feature vector for classification. Thereafter, the cloud mask is extracted by optimal thresholding and applying the Support Vector Machine (SVM) algorithm at the superpixel level. The GrabCut algorithm is then applied to extract more accurate cloud regions at the pixel level by assuming the cloud mask as the prior knowledge. When compared to different cloud detection methods in the literature, the overall accuracy of the proposed cloud detection method was up to 90 percent for ZY-3 and GF-1 images, which is about a 6.8 percent improvement over the traditional spectral-based methods. The experimental results show that the proposed method can automatically and accurately detect clouds using the multispectral information of the available four bands. © 2016 by the authors." "57194694881;6507122674;","Classification of 3D digital heritage",2019,"10.3390/RS11070847","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069810409&doi=10.3390%2fRS11070847&partnerID=40&md5=e18b3d22a234f60c7392dd30d73fad40","In recent years, the use of 3D models in cultural and archaeological heritage for documentation and dissemination purposes is increasing. The association of heterogeneous information to 3D data by means of automated segmentation and classification methods can help to characterize, describe and better interpret the object under study. Indeed, the high complexity of 3D data along with the large diversity of heritage assets themselves have constituted segmentation and classification methods as currently active research topics. Although machine learning methods brought great progress in this respect, few advances have been developed in relation to cultural heritage 3D data. Starting from the existing literature, this paper aims to develop, explore and validate reliable and efficient automated procedures for the classification of 3D data (point clouds or polygonal mesh models) of heritage scenarios. In more detail, the proposed solution works on 2D data (""texture-based"" approach) or directly on the 3D data (""geometry-based approach) with supervised or unsupervised machine learning strategies. The method was applied and validated on four different archaeological/architectural scenarios. Experimental results demonstrate that the proposed approach is reliable and replicable and it is effective for restoration and documentation purposes, providing metric information e.g. of damaged areas to be restored. © 2019 by the authors." "47461334700;57203927097;57186194700;57203928635;25929453700;57203929375;57203925154;36188365300;55949126400;","Integration of terrestrial and drone-borne hyperspectral and photogrammetric sensing methods for exploration mapping and mining monitoring",2018,"10.3390/rs10091366","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85053639766&doi=10.3390%2frs10091366&partnerID=40&md5=6f5c54a913780938273dd4dc2144b1be","Mapping lithology and geological structures accurately remains a challenge in difficult terrain or in active mining areas. We demonstrate that the integration of terrestrial and drone-borne multi-sensor remote sensing techniques significantly improves the reliability, safety, and efficiency of geological activities during exploration and mining monitoring. We describe an integrated workflow to produce a geometrically and spectrally accurate combination of a Structure-from-Motion Multi-View Stereo point cloud and hyperspectral data cubes in the visible to near-infrared (VNIR) and short-wave infrared (SWIR), as well as long-wave infrared (LWIR) ranges acquired by terrestrial and drone-borne imaging sensors. Vertical outcrops in a quarry in the Freiberg mining district, Saxony (Germany), featuring sulfide-rich hydrothermal zones in a granitoid host, are used to showcase the versatility of our approach. The image data are processed using spectroscopic and machine learning algorithms to generate meaningful 2.5D (i.e., surface) maps that are available to geologists on the ground just shortly after data acquisition. We validate the remote sensing data with thin section analysis and laboratory X-ray diffraction, as well as point spectroscopic data. The combination of ground- and drone-based photogrammetric and hyperspectral VNIR, SWIR, and LWIR imaging allows for safer and more efficient ground surveys, as well as a better, statistically sound sampling strategy for further structural, geochemical, and petrological investigations. © 2018 by the authors." "57040611100;8628726200;","Using reforecasts to improve forecasting of fog and visibility for aviation",2016,"10.1175/WAF-D-15-0108.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84965029158&doi=10.1175%2fWAF-D-15-0108.1&partnerID=40&md5=ec823e225611f2ffcacd50bfcaef58b9","Fifteen years of forecasts from the National Oceanic and Atmospheric Administration's Second-Generation Global Medium-Range Ensemble Reforecast (GEFS/R) dataset were used to develop a statistical model that generates probabilistic predictions of cloud ceiling and visibility. Four major airports-Seattle-Tacoma International Airport (KSEA), San Francisco International Airport (KSFO), Denver International Airport (KDEN), and George Bush Intercontinental Airport (KIAH) in Houston, Texas-were selected for model training and analysis. Numerous statistical model configurations, including the use of several different machine learning algorithms, input predictors, and internal parameters, were explored and verified through cross validation to develop skillful forecasts at each station. The final model was then compared with both probabilistic climatology-based forecasts and deterministic operational guidance. Results indicated significantly enhanced skill within both deterministic and probabilistic frameworks from the model trained in this study relative to both operational guidance and climatology at all stations. Probabilistic forecasts also showed substantially higher skill within the framework used than any deterministic forecast. Dewpoint depression and cloud cover forecast fields from the GEFS/R model were typically found to have the highest correspondence with observed flight rule conditions of the atmospheric fields examined. Often forecast values nearest the prediction station were not found to be the most important flight rule condition predictors, with forecast values along coastlines and immediately offshore, where applicable, often serving as superior predictors. The effect of training data length on model performance was also examined; it was determined that approximately 3 yr of training data from a dynamical model were required for the statistical model to robustly capture the relationships between model variables and observed flight rule conditions (FRCs). © 2016 American Meteorological Society." "57191637325;55515805700;7404264180;34770075600;36604310300;57213448151;","A location selection policy of live virtual machine migration for power saving and load balancing",2013,"10.1155/2013/492615","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84890092998&doi=10.1155%2f2013%2f492615&partnerID=40&md5=2ccad30867d407c216530190ae047e54","Green cloud data center has become a research hotspot of virtualized cloud computing architecture. And load balancing has also been one of the most important goals in cloud data centers. Since live virtual machine (VM) migration technology is widely used and studied in cloud computing, we have focused on location selection (migration policy) of live VM migration for power saving and load balancing. We propose a novel approach MOGA-LS, which is a heuristic and self-adaptive multiobjective optimization algorithm based on the improved genetic algorithm (GA). This paper has presented the specific design and implementation of MOGA-LS such as the design of the genetic operators, fitness values, and elitism. We have introduced the Pareto dominance theory and the simulated annealing (SA) idea into MOGA-LS and have presented the specific process to get the final solution, and thus, the whole approach achieves a long-term efficient optimization for power saving and load balancing. The experimental results demonstrate that MOGA-LS evidently reduces the total incremental power consumption and better protects the performance of VM migration and achieves the balancing of system load compared with the existing research. It makes the result of live VM migration more high-effective and meaningful. © 2013 Jia Zhao et al." "36614966600;55893847400;","Learning grammar rules of building parts from precise models and noisy observations",2011,"10.1016/j.isprsjprs.2010.10.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-79951682749&doi=10.1016%2fj.isprsjprs.2010.10.001&partnerID=40&md5=2cb6f0e867771861b9188abe9f80b571","The automatic interpretation of dense three-dimensional (3D) point clouds is still an open research problem. The quality and usability of the derived models depend to a large degree on the availability of highly structured models which represent semantics explicitly and provide a priori knowledge to the interpretation process. The usage of formal grammars for modelling man-made objects has gained increasing interest in the last few years. In order to cope with the variety and complexity of buildings, a large number of fairly sophisticated grammar rules are needed. As yet, such rules mostly have to be designed by human experts. This article describes a novel approach to machine learning of attribute grammar rules based on the Inductive Logic Programming paradigm. Apart from syntactic differences, logic programs and attribute grammars are basically the same language. Attribute grammars extend context-free grammars by attributes and semantic rules and provide a much larger expressive power. Our approach to derive attribute grammars is able to deal with two kinds of input data. On the one hand, we show how attribute grammars can be derived from precise descriptions in the form of examples provided by a human user as the teacher. On the other hand, we present the acquisition of models from noisy observations such as 3D point clouds. This includes the learning of geometric and topological constraints by taking measurement errors into account. The feasibility of our approach is proven exemplarily by stairs, and a generic framework for learning other building parts is discussed. Stairs aggregate an arbitrary number of steps in a manner which is specified by topological and geometric constraints and can be modelled in a recursive way. Due to this recursion, they pose a special challenge to machine learning. In order to learn the concept of stairs, only a small number of examples were required. Our approach represents and addresses the quality of the given observations and the derived constraints explicitly, using concepts from uncertain projective geometry for learning geometric relations and the Wakeby distribution together with decision trees for topological relations. © 2010 International Society for Photogrammetry and Remote Sensing, Inc.(ISPRS)." "56705310600;6603888005;6603354695;57192947904;16030007800;57205068357;8531925900;","Estimating crop primary productivity with Sentinel-2 and Landsat 8 using machine learning methods trained with radiative transfer simulations",2019,"10.1016/j.rse.2019.03.002","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063587909&doi=10.1016%2fj.rse.2019.03.002&partnerID=40&md5=b663838812a43b4242323b3045dcdd86","Satellite remote sensing has been widely used in the last decades for agricultural applications, both for assessing vegetation condition and for subsequent yield prediction. Existing remote sensing-based methods to estimate gross primary productivity (GPP), which is an important variable to indicate crop photosynthetic function and stress, typically rely on empirical or semi-empirical approaches, which tend to over-simplify photosynthetic mechanisms. In this work, we take advantage of all parallel developments in mechanistic photosynthesis modeling and satellite data availability for an advanced monitoring of crop productivity. In particular, we combine process-based modeling with the soil-canopy energy balance radiative transfer model (SCOPE) with Sentinel-2 and Landsat 8 optical remote sensing data and machine learning methods in order to estimate crop GPP. With this approach, we by-pass the need for an intermediate step to retrieve the set of vegetation biophysical parameters needed to accurately model photosynthesis, while still accounting for the complex processes of the original physically-based model. Several implementations of the machine learning models are tested and validated using simulated and flux tower-based GPP data. Our final neural network model is able to estimate GPP at the tested flux tower sites with r 2 of 0.92 and RMSE of 1.38 gC d −1 m −2 , which outperforms empirical models based on vegetation indices. The first test of applicability of this model to Landsat 8 data showed good results (r 2 of 0.82 and RMSE of 1.97 gC d −1 m −2 ), which suggests that our approach can be further applied to other sensors. Modeling and testing is restricted to C3 crops in this study, but can be extended to C4 crops by producing a new training dataset with SCOPE that accounts for the different photosynthetic pathways. Our model successfully estimates GPP across a variety of C3 crop types and environmental conditions even though it does not use any local information from the corresponding sites. This highlights its potential to map crop productivity from new satellite sensors at a global scale with the help of current Earth observation cloud computing platforms. © 2019" "37047349100;56324442700;35361477400;","Using google earth engine to map complex shade-grown coffee landscapes in northern Nicaragua",2018,"10.3390/rs10060952","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048966937&doi=10.3390%2frs10060952&partnerID=40&md5=9687e400fade9f7c040d09574c1c9f82","Shade-grown coffee (shade coffee) is an important component of the forested tropics, and is essential to the conservation of forest-dependent biodiversity. Despite its importance, shade coffee is challenging to map using remotely sensed data given its spectral similarity to forested land. This paper addresses this challenge in three districts of northern Nicaragua, here leveraging cloud-based computing techniques within Google Earth Engine (GEE) to integrate multi-seasonal Landsat 8 satellite imagery (30 m), and physiographic variables (temperature, topography, and precipitation). Applying a random forest machine learning algorithm using reference data from two field surveys produced a 90.5% accuracy across ten classes of land cover, with an 82.1% and 80.0% user's and producer's accuracy respectively for shade-grown coffee. Comparing classification accuracies obtained from five datasets exploring different combinations of non-seasonal and seasonal spectral data as well as physiographic data also revealed a trend of increasing accuracy when seasonal data were included in the model and a significant improvement (7.8-20.1%) when topographical data were integrated with spectral data. These results are significant in piloting an open-access and user-friendly approach to mapping heterogeneous shade coffee landscapes with high overall accuracy, even in locations with persistent cloud cover. © 2018 by the authors." "57191962971;6602413782;24503806900;","AUTOMATED CLASSIFICATION of HERITAGE BUILDINGS for AS-BUILT BIM USING MACHINE LEARNING TECHNIQUES",2017,"10.5194/isprs-annals-IV-2-W2-25-2017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85030261014&doi=10.5194%2fisprs-annals-IV-2-W2-25-2017&partnerID=40&md5=5887159bb7bb82fef2093c977985111a","Semantically rich three dimensional models such as Building Information Models (BIMs) are increasingly used in digital heritage. They provide the required information to varying stakeholders during the different stages of the historic buildings life cyle which is crucial in the conservation process. The creation of as-built BIM models is based on point cloud data. However, manually interpreting this data is labour intensive and often leads to misinterpretations. By automatically classifying the point cloud, the information can be proccesed more effeciently. A key aspect in this automated scan-to-BIM process is the classification of building objects. In this research we look to automatically recognise elements in existing buildings to create compact semantic information models. Our algorithm efficiently extracts the main structural components such as floors, ceilings, roofs, walls and beams despite the presence of significant clutter and occlusions. More specifically, Support Vector Machines (SVM) are proposed for the classification. The algorithm is evaluated using real data of a variety of existing buildings. The results prove that the used classifier recognizes the objects with both high precision and recall. As a result, entire data sets are reliably labelled at once. The approach enables experts to better document and process heritage assets. © Authors 2017. CC BY 4.0 License." "57190012327;7401911971;55971004500;23008015000;","A machine learning based reconstruction method for satellite remote sensing of soil moisture images with in situ observations",2017,"10.3390/rs9050484","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85019936862&doi=10.3390%2frs9050484&partnerID=40&md5=5bc667fefbb354454699bc91e60efb60","Surface soil moisture is an important environment variable that is dominant in a variety of research and application areas. Acquiring spatiotemporal continuous soil moisture observations is therefore of great importance. Weather conditions can contaminate optical remote sensing observations on soil moisture, and the absence of remote sensors causes gaps in regional soil moisture observation time series. Therefore, reconstruction is highly motivated to overcome such contamination and to fill in such gaps. In this paper, we propose a novel image reconstruction algorithm that improved upon the Satellite and In situ sensor Collaborated Reconstruction (SICR) algorithm provided by our previous publication. Taking artificial neural networks as a model, complex and highly variable relationships between in situ observations and remote sensing soil moisture is better projected. With historical data for the network training, feedforward neural networks (FNNs) project in situ soil moisture to remote sensing soil moisture at better performances than conventional models. Consequently, regional soil moisture observations can be reconstructed under full cloud contamination or under a total absence of remote sensors. Experiments confirmed better reconstruction accuracy and precision with this improvement than with SICR. The new algorithm enhances the temporal resolution of high spatial resolution remote sensing regional soil moisture observations with good quality and can benefit multiple soil moisture-based applications and research. © 2017 by the authors." "7403045983;56420772600;57191953810;","Multiple Kernel Sparse Representation for Airborne LiDAR Data Classification",2017,"10.1109/TGRS.2016.2619384","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84995514061&doi=10.1109%2fTGRS.2016.2619384&partnerID=40&md5=4a159b4ddef2423da63d7c910d94c4c6","To effectively learn heterogeneous features extracted from raw LiDAR point cloud data for landcover classification, a multiple kernel sparse representation classification (MKSRC) framework is proposed in this paper. In the MKSRC, multiple kernel learning (MKL) is embedded into sparse representation classification (SRC). The heterogeneous features are first extracted from the raw LiDAR point cloud data before classification. These features contain useful information from different dimensions, including single point features and neighbor features. Based on feature extraction, on the one hand, MKL is reasonably integrated into the SRC, namely, different base kernels that are constructed with each heterogeneous feature separately are utilized in the process of sparse representation. Furthermore, joint sparsity model is also introduced into the MKSRC framework and multiple kernel joint SRC (MKJSRC) is then proposed. On the other hand, improved kernel alignment (IKA) methods are proposed to more effectively determine the weights of base kernels in both of MKSRC and MKJSRC. Experiments are conducted on three real airborne LiDAR data sets. The experimental results demonstrate that MKSRC and MKJSRC frameworks can effectively learn the heterogeneous features for LiDAR point cloud classification and outperforms the other state-of-the-art sparse representation-based classifiers and the recent MKL algorithm. Moreover, the proposed IKA is helpful to better determine the 'optimal' weights of the base kernels in both MKSRC and MKJSRC than in the existing kernel alignment method. © 1980-2012 IEEE." "55308134300;6701525647;57195650018;7006763621;57206132735;12808541000;8271528700;8870015300;12041384200;57210713987;25623421700;8744419200;30267916000;55370113700;8712352400;","Testing the chemical tagging technique with open clusters",2015,"10.1051/0004-6361/201425232","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84929206771&doi=10.1051%2f0004-6361%2f201425232&partnerID=40&md5=dd651bcafd5b2eeb4fa803a33ae32180","Context. Stars are born together from giant molecular clouds and, if we assume that the priors were chemically homogeneous and well-mixed, we expect them to share the same chemical composition. Most of the stellar aggregates are disrupted while orbiting the Galaxy and most of the dynamic information is lost, thus the only possibility of reconstructing the stellar formation history is to analyze the chemical abundances that we observe today. Aims. The chemical tagging technique aims to recover disrupted stellar clusters based merely on their chemical composition. We evaluate the viability of this technique to recover co-natal stars that are no longer gravitationally bound. Methods. Open clusters are co-natal aggregates that have managed to survive together. We compiled stellar spectra from 31 old and intermediate-age open clusters, homogeneously derived atmospheric parameters, and 17 abundance species, and applied machine learning algorithms to group the stars based on their chemical composition. This approach allows us to evaluate the viability and efficiency of the chemical tagging technique. Results. We found that stars at different evolutionary stages have distinct chemical patterns that may be due to NLTE effects, atomic diffusion, mixing, and biases. When separating stars into dwarfs and giants, we observed that a few open clusters show distinct chemical signatures while the majority show a high degree of overlap. This limits the recovery of co-natal aggregates by applying the chemical tagging technique. Nevertheless, there is room for improvement if more elements are included and models are improved. © ESO, 2015." "55893253300;35234314600;24476946300;57195935722;26428593300;","UAV-based structural damage mapping: A review",2019,"10.3390/ijgi9010014","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077571589&doi=10.3390%2fijgi9010014&partnerID=40&md5=708dc44e5a994554c9b93087d0c269bd","Structural disaster damage detection and characterization is one of the oldest remote sensing challenges, and the utility of virtually every type of active and passive sensor deployed on various air- and spaceborne platforms has been assessed. The proliferation and growing sophistication of unmanned aerial vehicles (UAVs) in recent years has opened up many new opportunities for damage mapping, due to the high spatial resolution, the resulting stereo images and derivatives, and the flexibility of the platform. This study provides a comprehensive review of how UAV-based damage mapping has evolved from providing simple descriptive overviews of a disaster science, to more sophisticated texture and segmentation-based approaches, and finally to studies using advanced deep learning approaches, as well as multi-temporal and multi-perspective imagery to provide comprehensive damage descriptions. The paper further reviews studies on the utility of the developed mapping strategies and image processing pipelines for first responders, focusing especially on outcomes of two recent European research projects, RECONASS (Reconstruction and Recovery Planning: Rapid and Continuously Updated Construction Damage, and Related Needs Assessment) and INACHUS (Technological and Methodological Solutions for Integrated Wide Area Situation Awareness and Survivor Localization to Support Search and Rescue Teams). Finally, recent and emerging developments are reviewed, such as recent improvements in machine learning, increasing mapping autonomy, damage mapping in interior, GPS-denied environments, the utility of UAVs for infrastructure mapping and maintenance, as well as the emergence of UAVs with robotic abilities. © 2019 by the authors." "57202646249;57192169932;57203311348;22950341000;57209333386;","Estimating above-ground biomass of maize using features derived from UAV-based RGB imagery",2019,"10.3390/rs11111261","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067396948&doi=10.3390%2frs11111261&partnerID=40&md5=d40f66f3c9a65fff6b032c97a262de59","The rapid, accurate, and economical estimation of crop above-ground biomass at the farm scale is crucial for precision agricultural management. The unmanned aerial vehicle (UAV) remote-sensing system has a great application potential with the ability to obtain remote-sensing imagery with high temporal-spatial resolution. To verify the application potential of consumer-grade UAV RGB imagery in estimating maize above-ground biomass, vegetation indices and plant height derived from UAV RGB imagery were adopted. To obtain a more accurate observation, plant height was directly derived from UAV RGB point clouds. To search the optimal estimation method, the estimation performances of the models based on vegetation indices alone, based on plant height alone, and based on both vegetation indices and plant height were compared. The results showed that plant height directly derived from UAV RGB point clouds had a high correlation with ground-truth data with an R2 value of 0.90 and an RMSE value of 0.12 m. The above-ground biomass exponential regression models based on plant height alone had higher correlations for both fresh and dry above-ground biomass with R2 values of 0.77 and 0.76, respectively, compared to the linear regression model (both R2 values were 0.59). The vegetation indices derived from UAV RGB imagery had great potential to estimate maize above-ground biomass with R2 values ranging from 0.63 to 0.73. When estimating the above-ground biomass of maize by using multivariable linear regression based on vegetation indices, a higher correlation was obtained with an R2 value of 0.82. There was no significant improvement of the estimation performance when plant height derived from UAV RGB imagery was added into the multivariable linear regression model based on vegetation indices. When estimating crop above-ground biomass based on UAV RGB remote-sensing system alone, looking for optimized vegetation indices and establishing estimation models with high performance based on advanced algorithms (e.g., machine learning technology) may be a better way. © 2019 by the authors." "55348158300;6602817904;","Untangling the Galaxy. I. Local Structure and Star Formation History of the Milky Way",2019,"10.3847/1538-3881/ab339a","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072702923&doi=10.3847%2f1538-3881%2fab339a&partnerID=40&md5=6cb4c6b45bd05eb4e88afeab5e30e13f","Gaia DR2 provides unprecedented precision in measurements of the distance and kinematics of stars in the solar neighborhood. Through applying unsupervised machine learning on DR2's 5D data set (3D position + 2D velocity), we identify a number of clusters, associations, and comoving groups within 1 kpc and (many of which have not been previously known). We estimate their ages with the precision of ∼0.15 dex. Many of these groups appear to be filamentary or string-like, oriented in parallel to the Galactic plane, and some span hundreds of parsec in length. Most of these string lack a central cluster, indicating that their filamentary structure is primordial, rather than the result of tidal stripping or dynamical processing. The youngest strings (<100 Myr) are orthogonal to the Local Arm. The older ones appear to be remnants of several other arm-like structures that cannot be presently traced by dust and gas. The velocity dispersion measured from the ensemble of groups and strings increase with age, suggesting a timescale for dynamical heating of ∼300 Myr. This timescale is also consistent with the age at which the population of strings begins to decline, while the population in more compact groups continues to increase, suggesting that dynamical processes are disrupting the weakly bound string populations, leaving only individual clusters to be identified at the oldest ages. These data shed a new light on the local galactic structure and a large-scale cloud collapse. © 2019. The American Astronomical Society. All rights reserved." "56543514000;23493587700;56386000100;57202621598;6602209960;","Ground and multi-class classification of Airborne Laser Scanner point clouds using Fully Convolutional Networks",2018,"10.3390/rs10111723","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057137810&doi=10.3390%2frs10111723&partnerID=40&md5=14ac73392dd6f4539e1a9474ab4c7979","Various classification methods have been developed to extract meaningful information from Airborne Laser Scanner (ALS) point clouds. However, the accuracy and the computational efficiency of the existing methods need to be improved, especially for the analysis of large datasets (e.g., at regional or national levels). In this paper, we present a novel deep learning approach to ground classification for Digital Terrain Model (DTM) extraction as well as for multi-class land-cover classification, delivering highly accurate classification results in a computationally efficient manner. Considering the top-down acquisition angle of ALS data, the point cloud is initially projected on the horizontal plane and converted into a multi-dimensional image. Then, classification techniques based on Fully Convolutional Networks (FCN) with dilated kernels are designed to perform pixel-wise image classification. Finally, labels are transferred from pixels to the original ALS points. We also designed a Multi-Scale FCN (MS-FCN) architecture to minimize the loss of information during the point-to-image conversion. In the ground classification experiment, we compared our method to a Convolutional Neural Network (CNN)-based method and LAStools software. We obtained a lower total error on both the International Society for Photogrammetry and Remote Sensing (ISPRS) filter test benchmark dataset and AHN-3 dataset in the Netherlands. In the multi-class classification experiment, our method resulted in higher precision and recall values compared to the traditional machine learning technique using Random Forest (RF); it accurately detected small buildings. The FCN achieved precision and recall values of 0.93 and 0.94 when RF obtained 0.91 and 0.92, respectively. Moreover, our strategy significantly improved the computational efficiency of state-of-the-art CNN-based methods, reducing the point-to-image conversion time from 47 h to 36 min in our experiments on the ISPRS filter test dataset. Misclassification errors remained in situations that were not included in the training dataset, such as large buildings and bridges, or contained noisy measurements. © 2018 by the authors." "57202228599;14419019600;25924302800;6603033059;","A cloud-based multi-temporal ensemble classifier to map smallholder farming systems",2018,"10.3390/rs10050729","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047518739&doi=10.3390%2frs10050729&partnerID=40&md5=bc838262e9481ee12d9c92ff48239f66","Smallholder farmers cultivate more than 80% of the cropland area available in Africa. The intrinsic characteristics of such farms include complex crop-planting patterns, and small fields that are vaguely delineated. These characteristics pose challenges to mapping crops and fields from space. In this study, we evaluate the use of a cloud-based multi-temporal ensemble classifier to map smallholder farming systems in a case study for southern Mali. The ensemble combines a selection of spatial and spectral features derived from multi-spectral Worldview-2 images, field data, and five machine learning classifiers to produce a map of the most prevalent crops in our study area. Different ensemble sizes were evaluated using two combination rules, namely majority voting and weighted majority voting. Both strategies outperform any of the tested single classifiers. The ensemble based on the weighted majority voting strategy obtained the higher overall accuracy (75.9%). This means an accuracy improvement of 4.65% in comparison with the average overall accuracy of the best individual classifier tested in this study. The maximum ensemble accuracy is reached with 75 classifiers in the ensemble. This indicates that the addition of more classifiers does not help to continuously improve classification results. Our results demonstrate the potential of ensemble classifiers to map crops grown byWest African smallholders. The use of ensembles demands high computational capability, but the increasing availability of cloud computing solutions allows their efficient implementation and even opens the door to the data processing needs of local organizations. © 2018 by the authors." "57193585555;7005457386;23974497200;8901407100;8528696400;56124073100;","Bayesian and classical machine learning methods: A Comparison for tree species classification with LiDAR waveform signatures",2018,"10.3390/rs10010039","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85040699889&doi=10.3390%2frs10010039&partnerID=40&md5=a882bae04d7833eb9166a9455fd85216","A plethora of information contained in full-waveform (FW) Light Detection and Ranging (LiDAR) data offers prospects for characterizing vegetation structures. This study aims to investigate the capacity of FW LiDAR data alone for tree species identification through the integration of waveform metrics with machine learning methods and Bayesian inference. Specifically, we first conducted automatic tree segmentation based on the waveform-based canopy height model (CHM) using three approaches including TreeVaW, watershed algorithms and the combination of TreeVaW and watershed (TW) algorithms. Subsequently, the Random forests (RF) and Conditional inference forests (CF) models were employed to identify important tree-level waveform metrics derived from three distinct sources, such as raw waveforms, composite waveforms, the waveform-based point cloud and the combined variables from these three sources. Further, we discriminated tree (gray pine, blue oak, interior live oak) and shrub species through the RF, CF and Bayesian multinomial logistic regression (BMLR) using important waveform metrics identified in this study. Results of the tree segmentation demonstrated that the TW algorithms outperformed other algorithms for delineating individual tree crowns. The CF model overcomes waveform metrics selection bias caused by the RF model which favors correlated metrics and enhances the accuracy of subsequent classification. We also found that composite waveforms are more informative than raw waveforms and waveform-based point cloud for characterizing tree species in our study area. Both classical machine learning methods (the RF and CF) and the BMLR generated satisfactory average overall accuracy (74% for the RF, 77% for the CF and 81% for the BMLR) and the BMLR slightly outperformed the other two methods. However, these three methods suffered from low individual classification accuracy for the blue oak which is prone to being misclassified as the interior live oak due to the similar characteristics of blue oak and interior live oak. Uncertainty estimates from the BMLR method compensate for this downside by providing classification results in a probabilistic sense and rendering users with more confidence in interpreting and applying classification results to real-world tasks such as forest inventory. Overall, this study recommends the CF method for feature selection and suggests that BMLR could be a superior alternative to classical machining learning methods." "8632797000;15828981100;55742229000;7401526171;7005052907;","Satellite-based precipitation estimation using watershed segmentation and growing hierarchical self-organizing map",2006,"10.1080/01431160600763428","https://www.scopus.com/inward/record.uri?eid=2-s2.0-34248361145&doi=10.1080%2f01431160600763428&partnerID=40&md5=646c84fb36ae1fdf676c9d5bae651bef","This paper outlines the development of a multi-satellite precipitation estimation methodology that draws on techniques from machine learning and morphology to produce high-resolution, short-duration rainfall estimates in an automated fashion. First, cloud systems are identified from geostationary infrared imagery using morphology based watershed segmentation algorithm. Second, a novel pattern recognition technique, growing hierarchical self-organizing map (GHSOM), is used to classify clouds into a number of clusters with hierarchical architecture. Finally, each cloud cluster is associated with co-registered passive microwave rainfall observations through a cumulative histogram matching approach. The network was initially trained using remotely sensed geostationary infrared satellite imagery and hourly ground-radar data in lieu of a dense constellation of polar-orbiting spacecraft such as the proposed global precipitation measurement (GPM) mission. Ground-radar and gauge rainfall measurements were used to evaluate this technique for both warm (June 2004) and cold seasons (December 2004-February 2005) at various temporal (daily and monthly) and spatial (0.04 and 0.25) scales. Significant improvements of estimation accuracy are found classifying the clouds into hierarchical sub-layers rather than a single layer. Furthermore, 2-year (2003-2004) satellite rainfall estimates generated by the current algorithm were compared with gauge-corrected Stage IV radar rainfall at various time scales over continental United States. This study demonstrates the usefulness of the watershed segmentation and the GHSOM in satellite-based rainfall estimations." "57196198872;57195530363;34881956500;57201646570;7202856872;","Assessing the pasturelands and livestock dynamics in Brazil, from 1985 to 2017: A novel approach based on high spatial resolution imagery and Google Earth Engine cloud computing",2019,"10.1016/j.rse.2019.111301","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068378546&doi=10.1016%2fj.rse.2019.111301&partnerID=40&md5=3987b46699dd85689966246971c915e1","The livestock activity accounts for a large part of the transformations in land cover in the world, with pasture areas being the main land use in Brazil and the main livelihood of the largest commercial herd in the world. In this sense, a better understanding of the spatial-temporal dynamics of pasture areas is of fundamental importance for a better occupation and territorial governance. Moreover, because they provide different ecosystem services, pastures play a key role in mitigating climate change and in meeting GHG emission reduction targets. Within this context, and based on Landsat image processing via machine learning methods in a cloud computing platform (Google Earth Engine), this work has mapped, annually and in an unprecedented way, the totality of the Brazilian pastures, from 1985 to 2017. With an overall accuracy of about 90%, the 33 maps produced indicated the pasture area varying from ~118 Mha ±3.41% (1985) to ~178 Mha ±2.53% (2017), with this expansion occurring mostly in the northern region of the country and to a lesser extent in the midwest. Temporarily, most of this expansion occurred in the first half of the period evaluated (i.e. between 1985 and 2002), with an increase in Brazilian pasture areas of ~57 mha in just 17 years. After 2002, this area remained relatively stable, varying between ~175 mha ±2.48% and ~178 mha ±2.53% by 2017. In 33 years, about 87% of the mapped areas experienced zero, one, two, or three land-cover / land-use transitions; overall, of the ~178 mha ±2.53% of existing pastures in 2017, ~52 mha are at least 33 years old, ~66 mha were formed after 1985, and ~33 mha may have undergone some reform action in the period under consideration. The dynamics revealed in this study reinforce the thesis of pasture utilization as a land reserve, and demonstrate the importance of these areas in the economic, social, and environmental aspects of Brazil. © 2019 Elsevier Inc." "22634069200;57204103920;56366080900;57037171600;57192115362;57204632561;","Inpainting of Remote Sensing SST Images with Deep Convolutional Generative Adversarial Network",2019,"10.1109/LGRS.2018.2870880","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054490519&doi=10.1109%2fLGRS.2018.2870880&partnerID=40&md5=0ff4fcf58721fe495f4103bcae883555","Cloud occlusion is a common problem in the satellite remote sensing (RS) field and poses great challenges for image processing and object detection. Most existing methods for cloud occlusion recovery extract the surrounding information from the single corrupted image rather than the historical RS image records. Moreover, the existing algorithms can only handle small and regular-shaped obnubilation regions. This letter introduces a deep convolutional generative adversarial network to recover the RS sea surface temperature images with cloud occlusion from the big historical image records. We propose a new loss function for the inpainting network, which adds a supervision term to solve our specific problem. Given a trained generative model, we search for the closest encoding of the corrupted image in the low-dimensional space using our inpainting loss function. This encoding is then passed through the generative model to infer the missing content. We conduct experiments on the RS image data set from the national oceanic and atmospheric administration. Compared with traditional and machine learning methods, both qualitative and quantitative results show that our method has advantages over existing methods. © 2004-2012 IEEE." "7202467963;36676931900;7501959001;","Integrating multisensor satellite data merging and image reconstruction in support of machine learning for better water quality management",2017,"10.1016/j.jenvman.2017.06.045","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85021280402&doi=10.1016%2fj.jenvman.2017.06.045&partnerID=40&md5=435861d3218065c139a0316bb38199e5","Monitoring water quality changes in lakes, reservoirs, estuaries, and coastal waters is critical in response to the needs for sustainable development. This study develops a remote sensing-based multiscale modeling system by integrating multi-sensor satellite data merging and image reconstruction algorithms in support of feature extraction with machine learning leading to automate continuous water quality monitoring in environmentally sensitive regions. This new Earth observation platform, termed “cross-mission data merging and image reconstruction with machine learning” (CDMIM), is capable of merging multiple satellite imageries to provide daily water quality monitoring through a series of image processing, enhancement, reconstruction, and data mining/machine learning techniques. Two existing key algorithms, including Spectral Information Adaptation and Synthesis Scheme (SIASS) and SMart Information Reconstruction (SMIR), are highlighted to support feature extraction and content-based mapping. Whereas SIASS can support various data merging efforts to merge images collected from cross-mission satellite sensors, SMIR can overcome data gaps by reconstructing the information of value-missing pixels due to impacts such as cloud obstruction. Practical implementation of CDMIM was assessed by predicting the water quality over seasons in terms of the concentrations of nutrients and chlorophyll-a, as well as water clarity in Lake Nicaragua, providing synergistic efforts to better monitor the aquatic environment and offer insightful lake watershed management strategies. © 2017 Elsevier Ltd" "57196243510;35363587600;","Introducing two Random Forest based methods for cloud detection in remote sensing images",2018,"10.1016/j.asr.2018.04.030","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046732706&doi=10.1016%2fj.asr.2018.04.030&partnerID=40&md5=376ada9d5601629d5e48dcd7f230d8e4","Cloud detection is a necessary phase in satellite images processing to retrieve the atmospheric and lithospheric parameters. Currently, some cloud detection methods based on Random Forest (RF) model have been proposed but they do not consider both spectral and textural characteristics of the image. Furthermore, they have not been tested in the presence of snow/ice. In this paper, we introduce two RF based algorithms, Feature Level Fusion Random Forest (FLFRF) and Decision Level Fusion Random Forest (DLFRF) to incorporate visible, infrared (IR) and thermal spectral and textural features (FLFRF) including Gray Level Co-occurrence Matrix (GLCM) and Robust Extended Local Binary Pattern (RELBP_CI) or visible, IR and thermal classifiers (DLFRF) for highly accurate cloud detection on remote sensing images. FLFRF first fuses visible, IR and thermal features. Thereafter, it uses the RF model to classify pixels to cloud, snow/ice and background or thick cloud, thin cloud and background. DLFRF considers visible, IR and thermal features (both spectral and textural) separately and inserts each set of features to RF model. Then, it holds vote matrix of each run of the model. Finally, it fuses the classifiers using the majority vote method. To demonstrate the effectiveness of the proposed algorithms, 10 Terra MODIS and 15 Landsat 8 OLI/TIRS images with different spatial resolutions are used in this paper. Quantitative analyses are based on manually selected ground truth data. Results show that after adding RELBP_CI to input feature set cloud detection accuracy improves. Also, the average cloud kappa values of FLFRF and DLFRF on MODIS images (1 and 0.99) are higher than other machine learning methods, Linear Discriminate Analysis (LDA), Classification And Regression Tree (CART), K Nearest Neighbor (KNN) and Support Vector Machine (SVM) (0.96). The average snow/ice kappa values of FLFRF and DLFRF on MODIS images (1 and 0.85) are higher than other traditional methods. The quantitative values on Landsat 8 images show similar trend. Consequently, while SVM and K-nearest neighbor show overestimation in predicting cloud and snow/ice pixels, our Random Forest (RF) based models can achieve higher cloud, snow/ice kappa values on MODIS and thin cloud, thick cloud and snow/ice kappa values on Landsat 8 images. Our algorithms predict both thin and thick cloud on Landsat 8 images while the existing cloud detection algorithm, Fmask cannot discriminate them. Compared to the state-of-the-art methods, our algorithms have acquired higher average cloud and snow/ice kappa values for different spatial resolutions. © 2018 COSPAR" "56472615400;6603673004;15132432700;","Discovery of new dipper stars with K2: A window into the inner disc region of T Tauri stars",2018,"10.1093/MNRAS/STY328","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047318806&doi=10.1093%2fMNRAS%2fSTY328&partnerID=40&md5=44f0f6835fc20e5da608cf859d2ea8e6","In recent years, a new class of young stellar object (YSO) has been defined, referred to as dippers, where large transient drops in flux are observed. These dips are too large to be attributed to stellar variability, last from hours to days and can reduce the flux of a star by 10-50 per cent. This variability has been attributed to occultations by warps or accretion columns near the inner edge of circumstellar discs. Here, we present 95 dippers in the Upper Scorpius association and ρ Ophiuchus cloud complex found in K2 Campaign 2 data using supervised machine learning with a random forest classifier. We also present 30 YSOs that exhibit brightening events on the order of days, known as bursters. Not all dippers and bursters are known members, but all exhibit infrared excesses and are consistent with belonging to either of the two young star-forming regions. We find 21.0 ± 5.5 per cent of stars with discs are dippers for both regions combined. Our entire dipper sample consists only of late-type (KM) stars, but we show that biases limit dipper discovery for earlier spectral types. Using the dipper properties as a proxy, we find that the temperature at the inner disc edge is consistent with interferometric results for similar and earlier type stars. © 2017 The Authors." "57190867837;14048744800;56111060800;57149551600;56735366800;55687238300;56954125400;","Cloud Type Classification of Total-Sky Images Using Duplex Norm-Bounded Sparse Coding",2017,"10.1109/JSTARS.2017.2669206","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85016500979&doi=10.1109%2fJSTARS.2017.2669206&partnerID=40&md5=5d5ebc8757300f0c69474233dff48381","Cloud type classification plays an essential role in ground-based cloud observation. However, it is a challenge to accurately identify the cloud categories that involve a large variety of structural patterns and visual appearances. Image representation and classifier are the crucial factors for cloud classification, though they are individually investigated in the literature. This paper proposes a new cloud type classification method using duplex norm-bounded sparse coding (DNSC), which designs image representation and classifier under the same framework, i.e., norm-bounded sparse coding (NSC). NSC is not only used to encode local descriptors, but also well explored to develop an effective classifier. NSC takes both locality and sparseness into account, and it can be benefit to capture discriminative patterns for image representation and have discriminative power for classifier. Furthermore, NSC has closed-form solution and can be computed efficiently. More specifically, DNSC first extracts local descriptor from an input cloud image, and then DNSC forms a holistic representation leveraging NSC and max-pooling strategy. Finally, a classifier is built on the holistic representation using NSC. The proposed DNSC is evaluated on the total-sky cloud image set, and the experimental results demonstrate that DNSC outperforms the state-of-the-art methods and its accuracy increases by about 7% compared with baselines. In addition, the categorywise performance improvement is particularly pronounced over the complex categories, such as Cirriform and Mixed cloudiness. © 2017 IEEE." "57213286480;55387269600;35268384500;","Classifying structures in the interstellar medium with support vector machines: The G16.05-0.57 supernova remnant",2011,"10.1088/0004-637X/741/1/14","https://www.scopus.com/inward/record.uri?eid=2-s2.0-80155212931&doi=10.1088%2f0004-637X%2f741%2f1%2f14&partnerID=40&md5=a447078746cfc17ccbef11b4f9dd1c48","We apply Support Vector Machines (SVMs) - a machine learning algorithm - to the task of classifying structures in the interstellar medium (ISM). As a case study, we present a position-position-velocity (PPV) data cube of 12CO J = 3-2 emission toward G16.05-0.57, a supernova remnant that lies behind the M17 molecular cloud. Despite the fact that these two objects partially overlap in PPV space, the two structures can easily be distinguished by eye based on their distinct morphologies. The SVM algorithm is able to infer these morphological distinctions, and associate individual pixels with each object at >90% accuracy. This case study suggests that similar techniques may be applicable to classifying other structures in the ISM - a task that has thus far proven difficult to automate. © 2011 The American Astronomical Society. All rights reserved." "24168143800;55894060300;57193168426;6602798575;","Towards national-scale characterization of grassland use intensity from integrated Sentinel-2 and Landsat time series",2020,"10.1016/j.rse.2019.03.017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064150657&doi=10.1016%2fj.rse.2019.03.017&partnerID=40&md5=78899ee9822ada8c1fb8c74f4867fbcc","The increased availability of systematically acquired high spatial and temporal resolution optical imagery improves the characterization of dynamic land surface processes such as agriculture. The use of time series phenology can help overcome limitations of conventional classification-based mapping approaches encountered when, for example, attempting to characterize grassland use intensity. In Europe, permanent grasslands account for more than one third of all agricultural land and a considerable share of the EU Common Agricultural Policy (CAP) budget is devoted to grasslands. The frequency and timing of mowing events is an important proxy for grassland use intensity and methods that allow characterizing grassland use intensity at the parcel level and over large areas are urgently needed. Here we present a novel algorithm that allows detecting and quantifying the number and timing of mowing events in central European grasslands. The algorithm utilizes all imagery acquired by Sentinel-2 MSI and Landsat-8 OLI for one entire year as available from the NASA Harmonized Landsat-Sentinel dataset. Cloud-free observations from both sensors are first synthesized through compositing at 10-day interval. Machine learning algorithms are then used to derive a grassland stratum. The intra-annual growing season profiles of NDVI values are subsequently assessed and compared to an idealized growing season trajectory. Residuals between the idealized trajectory and a polynomial model fit to the observed NDVI values are then evaluated to detect potential mowing events. We demonstrate and evaluate the performance of our algorithm and utilize its large area analysis capabilities by mapping the frequency and timing of grassland mowing events in 2016 on the national-scale across Germany. Our results suggest that 25% of the grassland area is not used for mowing. Validation results however suggest a relatively high omission error of the algorithm for areas that only experienced a single mowing event. The date ranges of detected mowing events compare overall well to a sample of interpreted time series points and to farm level reports on mowing dates. The mapped mowing patterns depict typical management regimes across Germany. Overall, our results exemplify the value of multi-sensor time series applications for characterizing land use intensity across large areas. © 2019 Elsevier Inc." "57208529533;36020977200;6508003494;55927784300;","Remote sensing of snow cover using spaceborne SAR: A review",2019,"10.3390/rs11121456","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068115357&doi=10.3390%2frs11121456&partnerID=40&md5=6518425541f2a99cc28db05692561af7","The importance of snow cover extent (SCE) has been proven to strongly link with various natural phenomenon and human activities; consequently, monitoring snow cover is one the most critical topics in studying and understanding the cryosphere. As snow cover can vary significantly within short time spans and often extends over vast areas, spaceborne remote sensing constitutes an effcient observation technique to track it continuously. However, as optical imagery is limited by cloud cover and polar darkness, synthetic aperture radar (SAR) attracted more attention for its ability to sense day-and-night under any cloud and weather condition. In addition to widely applied backscattering-based method, thanks to the advancements of spaceborne SAR sensors and image processing techniques, many new approaches based on interferometric SAR (InSAR) and polarimetric SAR (PolSAR) have been developed since the launch of ERS-1 in 1991 to monitor snow cover under both dry and wet snow conditions. Critical auxiliary data including DEM, land cover information, and local meteorological data have also been explored to aid the snow cover analysis. This review presents an overview of existing studies and discusses the advantages, constraints, and trajectories of the current developments. © 2019 by the authors." "57192312900;57203279842;","Accuracies achieved in classifying five leading world crop types and their growth stages using optimal earth observing-1 hyperion hyperspectral narrowbands on Google Earth Engine",2018,"10.3390/rs10122027","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85058870617&doi=10.3390%2frs10122027&partnerID=40&md5=45d296c9c73f827cf2669407b19522a6","As the global population increases, we face increasing demand for food and nutrition. Remote sensing can help monitor food availability to assess global food security rapidly and accurately enough to inform decision-making. However, advances in remote sensing technology are still often limited to multispectral broadband sensors. Although these sensors havemany applications, they can be limited in studying agricultural crop characteristics such as differentiating crop types and their growth stages with a high degree of accuracy and detail. In contrast, hyperspectral data contain continuous narrowbands that provide data in terms of spectral signatures rather than a few data points along the spectrum, and hence can help advance the study of crop characteristics. To better understand and advance this idea, we conducted a detailed study of five leading world crops (corn, soybean, winter wheat, rice, and cotton) that occupy 75% and 54% of principal crop areas in the United States and the world respectively. The study was conducted in seven agroecological zones of the United States using 99 Earth Observing-1 (EO-1) Hyperion hyperspectral images from 2008-2015 at 30 m resolution. The authors first developed a first-of-its-kind comprehensive Hyperion-derived Hyperspectral Imaging Spectral Library of Agricultural crops (HISA) of these crops in the US based on USDA Cropland Data Layer (CDL) reference data. Principal Component Analysis was used to eliminate redundant bands by using factor loadings to determine which bands most influenced the first few principal components. This resulted in the establishment of 30 optimal hyperspectral narrowbands (OHNBs) for the study of agricultural crops. The rest of the 242 Hyperion HNBs were redundant, uncalibrated, or noisy. Crop types and crop growth stages were classified using linear discriminant analysis (LDA) and support vector machines (SVM) in the Google Earth Engine cloud computing platform using the 30 optimal HNBs (OHNBs). The best overall accuracies were between 75% to 95% in classifying crop types and their growth stages, which were achieved using 15-20 HNBs in the majority of cases. However, in complex cases (e.g., 4 or more crops in a Hyperion image) 25-30 HNBs were required to achieve optimal accuracies. Beyond 25-30 bands, accuracies asymptote. This research makes a significant contribution towards understanding modeling, mapping, and monitoring agricultural crops using data from upcoming hyperspectral satellites, such as NASA's Surface Biology and Geology mission (formerly HyspIRI mission) and the recently launched HysIS (Indian Hyperspectral Imaging Satellite, 55 bands over 400-950 nm in VNIR and 165 bands over 900-2500 nm in SWIR), and contributions in advancing the building of a novel, first-of-its-kind global hyperspectral imaging spectral-library of agricultural crops (GHISA: www.usgs.gov/WGSC/GHISA). © 2018 by the authors." "57203802638;35796556100;28268124500;22933900100;7403375011;6505807960;7004580125;","Robust quantification of riverine land cover dynamics by high-resolution remote sensing",2018,"10.1016/j.rse.2018.08.035","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85053020163&doi=10.1016%2fj.rse.2018.08.035&partnerID=40&md5=32fe03568064b4d225aaf22a56caab2f","Floodplain areas belong to the most diverse, dynamic and complex ecological habitats of the terrestrial portion of the Earth. Spatial and temporal quantification of floodplain dynamics is needed for assessing the impacts of hydromorphological controls on river ecosystems. However, estimation of land cover dynamics in a post-classification setting is hindered by a high contribution of classification errors. A possible solution relies on the selection of specific information of the change map, instead of increasing the overall classification accuracy. In this study, we analyze the capabilities of Unmanned Aerial Systems (UAS), the associated classification processes and their respective accuracies to extract a robust estimate of floodplain dynamics. We show that an estimation of dynamics should be built on specific land cover interfaces to be robust against classification errors and should include specific features depending on the season-sensor coupling. We use five different sets of features and determine the optimal combination to use information largely based on blue and infrared bands with the support of texture and point cloud metrics at leaf-off conditions. In this post-classification setting, the best observation of dynamics can be achieved by focusing on the gravel-water interface. The semi-supervised approach generated error of 10% of observed changes along highly dynamic reaches using these two land cover classes. The results show that a robust quantification of floodplain land cover dynamics can be achieved by high-resolution remote sensing. © 2018 Elsevier Inc." "55561559600;57201373924;57201378591;57202193525;6507534695;","Classification of aerial photogrammetric 3D point clouds",2018,"10.14358/PERS.84.5.287","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047395345&doi=10.14358%2fPERS.84.5.287&partnerID=40&md5=f0eee31acdd513bf06de58aa4cadf01c","We present a powerful method to extract per-point semantic class labels from aerial photogrammetry data. Labeling this kind of data is important for tasks such as environmental modeling, object classification, and scene understanding. Unlike previous point cloud classification methods that rely exclusively on geometric features, we show that incorporating color information yields a significant increase in accuracy in detecting semantic classes. We test our classification method on four real-world photogrammetry datasets that were generated with Pix4Dmapper, and with varying point densities. We show that off-the-shelf machine learning techniques coupled with our new features allow us to train highly accurate classifiers that generalize well to unseen data, processing point clouds containing 10 million points in less than three minutes on a desktop computer. We also demonstrate that our approach can be used to generate accurate Digital Terrain Models, outperforming approaches based on more simple heuristics such as Maximally Stable Extremal Regions. © 2018 American Society for Photogrammetry and Remote Sensing." "56203143700;23006934800;7005742190;","A hybrid approach for fog retrieval based on a combination of satellite and ground truth data",2018,"10.3390/rs10040628","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045998717&doi=10.3390%2frs10040628&partnerID=40&md5=5a55942d0fa14fd40a9c54d48e198537","Fog has a substantial influence on various ecosystems and it impacts economy, traffic systems and human life in many ways. In order to be able to deal with the large number of influence factors, a spatially explicit high-resoluted data set of fog frequency distribution is needed. In this study, a hybrid approach for fog retrieval based on Meteosat Second Generation (MSG) data and ground truth data is presented. The method is based on a random forest (RF) machine learning model that is trained with cloud base altitude (CBA) observations from Meteorological Aviation Routine Weather Reports (METAR) as well as synoptic weather observations (SYNOP). Fog is assumed where the model predicts CBA values below a dynamically derived threshold above the terrain elevation. Cross validation results show good accordance with observation data with a mean absolute error of 298 m in CBA values and an average Heidke Skill Score of 0.58 for fog occurrence. Using this technique, a 10 year fog baseline climatology with a temporal resolution of 15 min was derived for Europe for the period from 2006 to 2015. Spatial and temporal variations in fog frequency are analyzed. Highest average fog occurrences are observed in mountainous regions with maxima in spring and summer. Plains and lowlands show less overall fog occurrence but strong positive anomalies in autumn and winter. © 2018 by the authors." "57111518400;9036557400;57188729460;57190373951;56503083100;35756335100;","Icing detection over East Asia from geostationary satellite data using machine learning approaches",2018,"10.3390/rs10040631","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045974773&doi=10.3390%2frs10040631&partnerID=40&md5=6e0cc386ad159c44e8b05033ee6cae0b","Even though deicing or airframe coating technologies continue to develop, aircraft icing is still one of the critical threats to aviation. While the detection of potential icing clouds has been conducted using geostationary satellite data in the US and Europe, there is not yet a robust model that detects potential icing areas in East Asia. In this study, we proposed machine-learning-based icing detection models using data from two geostationary satellites-the Communication, Ocean, and Meteorological Satellite (COMS) Meteorological Imager (MI) and the Himawari-8 Advanced Himawari Imager (AHI)-over Northeast Asia. Two machine learning techniques-random forest (RF) and multinomial log-linear (MLL) models-were evaluated with quality-controlled pilot reports (PIREPs) as the reference data. The machine-learning-based models were compared to the existing models through five-fold cross-validation. The RF model for COMS MI produced the best performance, resulting in a mean probability of detection (POD) of 81.8%, a mean overall accuracy (OA) of 82.1%, and mean true skill statistics (TSS) of 64.0%. One of the existing models, flight icing threat (FIT), produced relatively poor performance, providing a mean POD of 36.4%, a mean OA of 61.0, and a mean TSS of 9.7%. The Himawari-8 based models also produced performance comparable to the COMS models. However, it should be noted that very limited PIREP reference data were available especially for the Himawari-8 models, which requires further evaluation in the future with more reference data. The spatio-temporal patterns of the icing areas detected using the developed models were also visually examined using time-series satellite data. © 2018 by the authors." "35726158400;56152411400;36093063000;24825034100;","Cloud based metalearning system for predictive modeling of biomedical data",2014,"10.1155/2014/859279","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84900023007&doi=10.1155%2f2014%2f859279&partnerID=40&md5=972e351ccdab23feb3dde40bca77d3b2","Rapid growth and storage of biomedical data enabled many opportunities for predictive modeling and improvement of healthcare processes. On the other side analysis of such large amounts of data is a difficult and computationally intensive task for most existing data mining algorithms. This problem is addressed by proposing a cloud based system that integrates metalearning framework for ranking and selection of best predictive algorithms for data at hand and open source big data technologies for analysis of biomedical data. © 2014 Milan Vukićević et al." "57202441820;57147072600;","Improving public data for building segmentation from Convolutional Neural Networks (CNNs) for fused airborne lidar and image data using active contours",2019,"10.1016/j.isprsjprs.2019.05.013","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066730376&doi=10.1016%2fj.isprsjprs.2019.05.013&partnerID=40&md5=52d19dfff7dc4ebfacad229debb7002f","Robust and reliable automatic building detection and segmentation from aerial images/point clouds has been a prominent field of research in remote sensing, computer vision and point cloud processing for a number of decades. One of the largest issues associated with deep learning methods is the high quantity of data required for training. To help address this we present a method to improve public GIS building footprint labels by using Morphological Geodesic Active Contours (MorphGACs). We demonstrate by improving the quality of building footprint labels for detection and semantic segmentation, more robust and reliable models can be obtained. We evaluate these methods over a large UK-based dataset of 24556 images containing 169835 building instances. This is achieved by training several Mask/Faster R-CNN and RetinaNet deep convolutional neural networks. Networks are supplied with both RGB and fused RGB-lidar data. We offer quantitative analysis on the benefits of the inclusion of depth data for building segmentation. By employing both methods we achieve a detection accuracy of 0.92 (mAP@0.5) and segmentation f1 scores of 0.94 over a 4911 test images ranging from urban to rural scenes. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "23981063100;57193132723;6701735773;","Ongoing Breakthroughs in Convective Parameterization",2019,"10.1007/s40641-019-00127-w","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065178633&doi=10.1007%2fs40641-019-00127-w&partnerID=40&md5=01c67e2f3d8310902fefc773153ed866","Purpose of Review: While the increase of computer power mobilizes a part of the atmospheric modeling community toward models with explicit convection or based on machine learning, we review the part of the literature dedicated to convective parameterization development for large-scale forecast and climate models. Recent Findings: Many developments are underway to overcome endemic limitations of traditional convective parameterizations, either in unified or multiobject frameworks: scale-aware and stochastic approaches, new prognostic equations or representations of new components such as cold pools. Understanding their impact on the emergent properties of a model remains challenging, due to subsequent tuning of parameters and the limited understanding given by traditional metrics. Summary: Further effort still needs to be dedicated to the representation of the life cycle of convective systems, in particular their mesoscale organization and associated cloud cover. The development of more process-oriented metrics based on new observations is also needed to help quantify model improvement and better understand the mechanisms of climate change. © 2019, Springer Nature Switzerland AG." "57207870010;56682032300;13405658600;36867775200;6603156461;","Machine learning to predict the global distribution of aerosol mixing state metrics",2018,"10.3390/atmos9010015","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85040237561&doi=10.3390%2fatmos9010015&partnerID=40&md5=aa275e13faf9e96510446aa4135b5001","Atmospheric aerosols are evolving mixtures of chemical species. In global climate models (GCMs), this ""aerosol mixing state"" is represented in a highly simplified manner. This can introduce errors in the estimates of climate-relevant aerosol properties, such as the concentration of cloud condensation nuclei. The goal for this study is to determine a global spatial distribution of aerosol mixing state with respect to hygroscopicity, as quantified by the mixing state metric χ. In this way, areas can be identified where the external or internal mixture assumption is more appropriate. We used the output of a large ensemble of particle-resolved box model simulations in conjunction with machine learning techniques to train a model of the mixing state metric c. This lower-order model for χ uses as inputs only variables known to GCMs, enabling us to create a global map of χ based on GCM data. We found that χ varied between 20% and nearly 100%, and we quantified how this depended on particle diameter, location, and time of the year. This framework demonstrates how machine learning can be applied to bridge the gap between detailed process modeling and a large-scale climate model. © 2017 by the author." "26646794300;57209065066;56649853900;35446475800;35509463200;7003755200;","Mapping burned areas in tropical forests using a novel machine learning framework",2018,"10.3390/rs10010069","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85040666800&doi=10.3390%2frs10010069&partnerID=40&md5=bb782516080ad9f710c80906ee5f3d93","This paper presents an application of a novel machine-learning framework on MODIS (moderate-resolution imaging spectroradiometer) data to map burned areas over tropical forests of South America and South-east Asia. The RAPT (RAre Class Prediction in the absence of True labels) framework is able to build data adaptive classification models using noisy training labels. It is particularly suitable when expert annotated training samples are difficult to obtain as in the case of wild fires in the tropics. This framework has been used to build burned area maps from MODIS surface reflectance data as features and Active Fire hotspots as training labels that are known to have high commission and omission errors due to the prevalence of cloud cover and smoke, especially in the tropics. Using the RAPT framework we report burned areas for 16 MODIS tiles from 2001 to 2014. The total burned area detected in the tropical forests of South America and South-east Asia during these years is 2,071,378 MODIS (500 m) pixels (approximately 520 K sq. km.), which is almost three times compared to the estimates from collection 5 MODIS MCD64A1 (783,468 MODIS pixels). An evaluation using Landsat-based reference burned area maps indicates that our product has an average user's accuracy of 53% and producer's accuracy of 55% while collection 5 MCD64A1 burned area product has an average user's accuracy of 61% and producer's accuracy of 27%. Our analysis also indicates that the two products can be complimentary and a combination of the two approaches is likely to provide a more comprehensive assessment of tropical fires. Finally, we have created a publicly accessible web-based viewer that helps the community to visualize the burned area maps produced using RAPT and examine various validation sources corresponding to every detected MODIS pixel." "29867490900;8213128600;6603354695;34869963500;8213128500;6603888005;","Randomized kernels for large scale Earth observation applications",2017,"10.1016/j.rse.2017.02.009","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85015678699&doi=10.1016%2fj.rse.2017.02.009&partnerID=40&md5=9d1180a016796546adca1fbbf10857a4","Current remote sensing applications of bio-geophysical parameter estimation and image classification have to deal with an unprecedented big amount of heterogeneous and complex data sources. New satellite sensors involving a high number of improved time, space and wavelength resolutions give rise to challenging computational problems. Standard physical inversion techniques cannot cope efficiently with this new scenario. Dealing with land cover classification of the new image sources has also turned to be a complex problem requiring large amount of memory and processing time. In order to cope with these problems, statistical learning has greatly helped in the last years to develop statistical retrieval and classification models that can ingest large amounts of Earth observation data. Kernel methods constitute a family of powerful machine learning algorithms, which have found wide use in remote sensing and geosciences. However, kernel methods are still not widely adopted because of the high computational cost when dealing with large scale problems, such as the inversion of radiative transfer models or the classification of high spatial-spectral-temporal resolution data. This paper introduces to the remote sensing community an efficient kernel method for fast statistical retrieval of atmospheric and biophysical parameters and image classification problems. We rely on a recently presented approximation to shift-invariant kernels using projections on random Fourier features. The method proposes an explicit mapping function defined through a set of projections randomly sampled from the Fourier domain. It is proved to approximate the implicit mapping of a kernel function. This allows to deal with large-scale data but taking advantage of kernel methods. The method is simple, computationally very efficient in both memory and processing costs, and easily parallelizable. We show that kernel regression and classification is now possible for datasets with millions of samples. Examples on atmospheric parameter retrieval from hyperspectral infrared sounders like IASI/Metop; large scale emulation and inversion of the familiar PROSAIL radiative transfer model on Sentinel-2 data; and the identification of clouds over landmarks in time series of MSG/Seviri images show the efficiency and effectiveness of the proposed technique. © 2017 Elsevier Inc." "35279503800;9241987300;","Correcting biased observation model error in data assimilation",2017,"10.1175/MWR-D-16-0428.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85021304929&doi=10.1175%2fMWR-D-16-0428.1&partnerID=40&md5=bc16a5c64ad46b09a26bf5d7d18be34a","While the formulation of most data assimilation schemes assumes an unbiased observation model error, in real applications model error with nontrivial biases is unavoidable. A practical example is errors in the radiative transfer model (which is used to assimilate satellite measurements) in the presence of clouds. Together with the dynamical model error, the result is that many (in fact 99%) of the cloudy observed measurements are not being used although they may contain useful information. This paper presents a novel nonparametric Bayesian scheme that is able to learn the observation model error distribution and correct the bias in incoming observations. This scheme can be used in tandem with any data assimilation forecasting system. The proposed model error estimator uses nonparametric likelihood functions constructed with data-driven basis functions based on the theory of kernel embeddings of conditional distributions developed in the machine learning community. Numerically, positive results are shown with two examples. The first example is designed to produce a bimodality in the observation model error (typical of ""cloudy"" observations) by introducing obstructions to the observations that occur randomly in space and time. The second example, which is physically more realistic, is to assimilate cloudy satellite brightness temperature-like quantities, generated from a stochastic multicloud model for tropical convection and a simple radiative transfer model. © 2017 American Meteorological Society." "57205740630;57205739434;57208528577;8605292800;55630272400;7404975854;","Mapping tidal flats with landsat 8 images and google Earth Engine: A case study of the China's Eastern coastal zone circa 2015",2019,"10.3390/rs11080959","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065012905&doi=10.3390%2frs11080959&partnerID=40&md5=7571595c16469386903dba1c0f407d9a","Accurate and up-to-date tidal flat mapping is of much importance to learning how coastal ecosystems work in a time of anthropogenic disturbances and rising sea levels, which will provide scientific instruction for sustainable management and ecological assessments. For large-scale and high spatial-resolution mapping of tidal flats, it is diffcult to obtain accurate tidal flat maps without multi-temporal observation data. In this study, we aim to investigate the potential and advantages of the freely accessible Landsat 8 Operational Land Imager (OLI) imagery archive and Google Earth Engine (GEE) for accurate tidal flats mapping. A novel approach was proposed, including multi-temporal feature extraction, machine learning classification using GEE and morphological post-processing. The 50 km buffer of the coastline from Hangzhou Bay to Yalu River in China's eastern coastal zone was taken as the study area. From the perspective of natural attributes and unexploited status of tidal flats, we delineated a broader extent comprising intertidal flats, supratidal barren flats and vegetated flats, since intertidal flats are major component of the tidal flats. The overall accuracy of the resultant map was about 94.4% from a confusion matrix for accuracy assessment. The results showed that the use of time-series images can greatly eliminate the effects of tidal level, and improve the mapping accuracy. This study also proved the potential and advantage of combining the GEE platform with time-series Landsat images, due to its powerful cloud computing platform, especially for large scale and longtime tidal flats mapping. © 2019 by the authors." "56211542300;57195673046;36173974000;56497323100;55224911200;7202941773;","Improving ocean color data coverage through machine learning",2019,"10.1016/j.rse.2018.12.023","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059813485&doi=10.1016%2fj.rse.2018.12.023&partnerID=40&md5=85eb60339c0f5ef2d803ff5531a4927e","Oceanic chlorophyll-a concentration (Chl, mg m −3 ) maps derived from satellite ocean color measurements are the only data source which provides synoptic information of phytoplankton abundance on global scale. However, after excluding data collected under non-optimal observing conditions such as strong sun glint, clouds, thick aerosols, straylight, and large viewing angles, only ~5% of MODIS ocean measurements lead to valid Chl retrievals, regardless of the fact that about 25–30% of the global ocean is cloud free. A recently developed ocean color index (CI) is effective in deriving relative ocean color patterns under most non-optimal observing conditions to improve coverage, but these patterns cannot be interpreted as Chl. In this study, we combine the advantage of the high-quality, low-coverage Chl and lower-quality, higher-coverage CI to improve spatial and temporal coverage of Chl through machine learning, specifically via a random forest based regression ensemble (RFRE) approach. For every MODIS scene, the machine learning requires CI, Rayleigh-corrected reflectance (R rc (λ = 469, 555, 645 nm), dimensionless), and high-quality low-coverage Chl from the common pixels where they all have valid data to develop an RFRE-based model to convert CI and R rc (λ) to Chl. The model is then applied to all valid CI pixels of the same scene to derive Chl. This process is repeated for each scene, and the model parameterization is optimized for each scene independently. The approach has been tested for the Yellow Sea and East China Sea (YSECS) where non-optimal observing conditions frequently occur. Validations using extensive field measurements and image-based statistics for 2017 show very promising results, where coverage in the new Chl maps is increased by ~3.5 times without noticeable degradation in quality as compared with the original Chl data products. The improvement in Chl coverage without compromising data quality is not only critical in revealing otherwise unknown bloom patterns, but also important in reducing uncertainties in time-series analysis. Tests of the RFRE approach for several other regions such as the East Caribbean, Arabian Sea, and Gulf of Mexico suggest its general applicability in improving Chl coverage of other regions. © 2018 Elsevier Inc." "57203719531;7202844175;","The use of terrestrial laser scanning for the characterization of a cliff-talus system in the Thompson River Valley, British Columbia, Canada",2019,"10.1016/j.geomorph.2018.11.022","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85058077408&doi=10.1016%2fj.geomorph.2018.11.022&partnerID=40&md5=3b2bfadb9fa631eab9bba6900e122bd9","A postglacial river terrace along the Thompson River in Interior British Columbia, Canada has been monitored using terrestrial laser scanning (TLS) and high-resolution photography for almost a 3-year study to observe the deformation and failure processes, which result in changes in the slope morphology. Change detection using Multiscale Model to Model Cloud Comparison (M3C2) and a multi-scale dimensionality analysis (CANUPO) were performed on the 3-dimensional point cloud data to track the deposition patterns occurring in this active cliff talus system. Changes documented in the analysis of TLS data were verified using the high-resolution photography. Over 1.5 m of valley parallel retreat was captured in a section of the cliff face related to instability of a cobble and boulder horizon beneath a thick fluvial gravel unit. Because of the high-resolution remote sensing data, it was possible to observe a longitudinal sorting of grain sizes (i.e. fall sorting) in this cliff-talus system, whereby the size of individual particles controls the position on the slope. The overall mapped distribution of particle sizes on the slope remained constant for the almost 3-year study period. Flows of granular debris were observed in TLS change detection and the CANUPO analysis was able to display the longitudinal and lateral sorting of grain sizes that occurs during flow. This case history demonstrates that high resolution remote sensing data of large slopes permits us to link the geomorphic processes occurring in the cliff face with mass movement and deposition occurring on the talus slope below. © 2018 Elsevier B.V." "56425743000;35565008000;7003749049;22134864100;","Building extraction from LiDAR data applying deep convolutional neural networks",2019,"10.1109/LGRS.2018.2867736","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85053325964&doi=10.1109%2fLGRS.2018.2867736&partnerID=40&md5=74892af2f37cd115be2c2e353fe98603","Deep learning paradigm has been shown to be a very efficient classification framework for many application scenarios, including the analysis of Light Detection and Ranging (LiDAR) data for building detection. In fact, deep learning acts as a set of mathematical transformations, encoding the raw input data into appropriate forms of representations that maximize the classification performance. However, it is clear that mathematical computations alone, even highly nonlinear, are not adequate to model the physical properties of a problem, distinguishing, for example, the building structures from vegetation. In this letter, we address this difficulty by augmenting the raw LiDAR data with features coming from a physical interpretation of the information. Then, we exploit a deep learning paradigm based on a convolutional neural network model to find out the best input representations suitable for the classification. As test sites, three complex urban study areas with various kinds of building structures through the LiDAR data set of Vaihingen, Germany were selected. Our method has been evaluated in the context of 'ISPRS Test Project on Urban Classification and 3-D Building Reconstruction.' Comparisons with traditional methods, such as artificial neural networks and support vector machine-based classifiers, indicate the outperformance of the proposed approach in terms of robustness and efficiency. © 2004-2012 IEEE." "56117999100;56518310300;7006375009;","A machine-learning approach to forecasting remotely sensed vegetation health",2018,"10.1080/01431161.2017.1410296","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046946925&doi=10.1080%2f01431161.2017.1410296&partnerID=40&md5=50fe9e40179124b0b82e629dcd26f216","Drought threatens food and water security around the world, and this threat is likely to become more severe under climate change. High resolution predictive information can help farmers, water managers, and others to manage the effects of drought. We have created an open-source tool to produce short-term forecasts of vegetation health at high spatial resolution, using data that are global in coverage. The tool automates downloading and processing Moderate Resolution Imaging Spectroradiometer (MODIS) data sets and training gradient boosted machine models on hundreds of millions of observations to predict future values of the enhanced vegetation index. We compared the predictive power of different sets of variables (MODIS surface reflectance data and Level-3 MODIS products) in two regions with distinct agro-ecological systems, climates, and cloud coverage: Sri Lanka and California. Performance in California is higher because of more cloud-free days and less missing data. In both regions, the correlation between the actual and model predicted vegetation health values in agricultural areas is above 0.75. Predictive power more than doubles in agricultural areas compared to a baseline model. © 2017 The Author(s)." "56737079700;29867490900;34870201500;6603888005;","Remote Sensing Image Classification with Large-Scale Gaussian Processes",2018,"10.1109/TGRS.2017.2758922","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85032447623&doi=10.1109%2fTGRS.2017.2758922&partnerID=40&md5=e4c962a88d36c634ee09df2f1da11d54","Current remote sensing image classification problems have to deal with an unprecedented amount of heterogeneous and complex data sources. Upcoming missions will soon provide large data streams that will make land cover/use classification difficult. Machine-learning classifiers can help at this, and many methods are currently available. A popular kernel classifier is the Gaussian process classifier (GPC), since it approaches the classification problem with a solid probabilistic treatment, thus yielding confidence intervals for the predictions as well as very competitive results to the state-of-the-art neural networks and support vector machines. However, its computational cost is prohibitive for large-scale applications, and constitutes the main obstacle precluding wide adoption. This paper tackles this problem by introducing two novel efficient methodologies for GP classification. We first include the standard random Fourier features approximation into GPC, which largely decreases its computational cost and permits large-scale remote sensing image classification. In addition, we propose a model which avoids randomly sampling a number of Fourier frequencies and alternatively learns the optimal ones within a variational Bayes approach. The performance of the proposed methods is illustrated in complex problems of cloud detection from multispectral imagery and infrared sounding data. Excellent empirical results support the proposal in both computational cost and accuracy. © 1980-2012 IEEE." "57196061602;57195202700;57144638800;6602918196;57196066156;8368714300;","Automatic Cloud-Type Classification Based On the Combined Use of a Sky Camera and a Ceilometer",2017,"10.1002/2017JD027131","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85031706802&doi=10.1002%2f2017JD027131&partnerID=40&md5=ea38eca4d49665b0209408eaee6833bd","A methodology, aimed to be fully operational, for automatic cloud classification based on the synergetic use of a sky camera and a ceilometer is presented. The random forest machine learning algorithm was used to train the classifier with 19 input features: 12 extracted from the sky camera images and 7 from the ceilometer. The method was developed and tested based on a set of 717 images collected at the radiometric stations of the Univ. of Jaén (Spain). Up to nine different types of clouds (plus clear sky) were considered (clear sky, cumulus, stratocumulus, nimbostratus, altocumulus, altostratus, stratus, cirrocumulus, cirrostratus, and cirrus) plus an additional category multicloud, aiming to account for the frequent cases in which the sky is covered by several cloud types. A total of eight experiments was conducted by (1) excluding/including the ceilometer information, (2) including/excluding the multicloud category, and (3) using six or nine different cloud types, aside from the clear-sky and multicloud category. The method provided accuracies ranging from 45% to 78%, being highly dependent on the use of the ceilometer information. This information showed to be particularly relevant for accurately classifying “cumuliform” clouds and to account for the multicloud category. In this regard, the camera information alone was found to be not suitable to deal with this category. Finally, while the use of the ceilometer provided an overall superior performance, some limitations were found, mainly related to the classification of clouds with similar cloud base height and geometric thickness. ©2017. American Geophysical Union. All Rights Reserved." "7202937042;56071618400;57193525599;","Northern conifer forest species classification using multispectral data acquired from an unmanned aerial vehicle",2017,"10.14358/PERS.83.7.501","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85023765352&doi=10.14358%2fPERS.83.7.501&partnerID=40&md5=981d5426e69397aa19857274da300d22","Object-based image analysis and machine learning classification procedures, after field calibration and photogrammetric processing of consumer-grade unmanned aerial vehicle (UAV) digital camera data, were implemented to classify tree species in a conifer forest in the Great Lakes/St Lawrence Lowlands Ecoregion, Ontario, Canada. A red-green-blue (RGB) digital camera yielded approximately 72 percent classification accuracy for three commercial tree species and one conifer shrub. Accuracy improved approximately 15 percent, to 87 percent overall, with higher radiometric quality data acquired separately using a digital camera that included near infrared observations (at a lower spatial resolution). Interpretation of the point cloud, spectral, texture and object (tree crown) classification Variable Importance (VI) selected by a machine learning algorithm suggested a good correspondence with the traditional aerial photointerpretation cues used in the development of well-established large-scale photography northern conifer elimination keys, which use three-dimensional crown shape, spectral response (tone), texture derivatives to quantify branching characteristics, and crown size, development and outline features. These results suggest that commonly available consumer-grade UAV-based digital cameras can be used with object-based image analysis to obtain acceptable conifer species classification accuracy to support operational forest inventory applications. © 2017 American Society for Photogrammetry and Remote Sensing." "57038880600;52263978100;56342251100;24391589100;","Optical cloud pixel recovery via machine learning",2017,"10.3390/rs9060527","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85021153126&doi=10.3390%2frs9060527&partnerID=40&md5=ae5f76ae343252067da1e4e983f44d74","Remote sensing derived Normalized Difference Vegetation Index (NDVI) is a widely used index to monitor vegetation and land use change. NDVI can be retrieved from publicly available data repositories of optical sensors such as Landsat, Moderate Resolution Imaging Spectro-radiometer (MODIS) and several commercial satellites. Studies that are heavily dependent on optical sensors are subject to data loss due to cloud coverage. Specifically, cloud contamination is a hindrance to long-term environmental assessment when using information from satellite imagery retrieved from visible and infrared spectral ranges. Landsat has an ongoing high-resolution NDVI record starting from 1984. Unfortunately, this long time series NDVI data suffers from the cloud contamination issue. Though both simple and complex computational methods for data interpolation have been applied to recover cloudy data, all the techniques have limitations. In this paper, a novel Optical Cloud Pixel Recovery (OCPR) method is proposed to repair cloudy pixels from the time-space-spectrum continuum using a Random Forest (RF) trained and tested with multi-parameter hydrologic data. The RF-based OCPR model is compared with a linear regression model to demonstrate the capability of OCPR. A case study in Apalachicola Bay is presented to evaluate the performance of OCPR to repair cloudy NDVI reflectance. The RF-based OCPR method achieves a root mean squared error of 0.016 between predicted and observed NDVI reflectance values. The linear regression model achieves a root mean squared error of 0.126. Our findings suggest that the RF-based OCPR method is effective to repair cloudy pixels and provides continuous and quantitatively reliable imagery for long-term environmental analysis. © 2017 by the authors." "56489062200;57212315922;56224155200;9036557400;56402112700;","Detection of tropical cyclone genesis via quantitative satellite ocean surface wind pattern and intensity analyses using decision trees",2016,"10.1016/j.rse.2016.06.006","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84973343153&doi=10.1016%2fj.rse.2016.06.006&partnerID=40&md5=6f9839ff73cd5c0e9647eef126095b2c","Microwave remote sensing can be used to measure ocean surface winds, which can be used to detect tropical cyclone (TC) formation in an objective and quantitative way. This study develops a new model using WindSat data and a machine learning approach. Dynamic and hydrologic indices are quantified from WindSat wind and rainfall snapshot images over 352 developing and 973 non-developing tropical disturbances from 2005 to 2009. The degree of cyclonic circulation symmetry near the system center is quantified using circular variances, and the degree of strong wind aggregation (heavy rainfall) is defined using a spatial pattern analysis program tool called FRAGSTATS. In addition, the circulation strength and convection are defined based on the areal averages of wind speed and rainfall. An objective TC formation detection model is then developed by applying those indices to a machine-learning decision tree algorithm using calibration data from 2005 to 2007. Results suggest that the circulation symmetry and intensity are the most important parameters that characterize developing tropical disturbances. Despite inherent sampling issues associated with the polar orbiting satellite, a validation from 2008 to 2009 shows that the model produced a positive detection rate of approximately 95.3% and false alarm rate of 28.5%, which is comparable with the pre-existing objective methods based on cloud-pattern recognition. This study suggests that the quantitative microwave-sensed dynamic ocean surface wind pattern and intensity recognition model provides a new method of detecting TC formation. © 2016 Elsevier Inc." "57211317672;16246477000;7404871794;","CloudFCN: Accurate and robust cloud detection for satellite imagery with deep learning",2019,"10.3390/rs11192312","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073477360&doi=10.3390%2frs11192312&partnerID=40&md5=12b1a4f8fd9d375f0315d879d63ac9c6","Cloud masking is of central importance to the Earth Observation community. This paper deals with the problem of detecting clouds in visible and multispectral imagery from high-resolution satellite cameras. Recently, Machine Learning has offered promising solutions to the problem of cloud masking, allowing for more flexibility than traditional thresholding techniques, which are restricted to instruments with the requisite spectral bands. However, few studies use multi-scale features (as in, a combination of pixel-level and spatial) whilst also offering compelling experimental evidence for real-world performance. Therefore, we introduce CloudFCN, based on a Fully Convolutional Network architecture, known as U-net, which has become a standard Deep Learning approach to image segmentation. It fuses the shallowest and deepest layers of the network, thus routing low-level visible content to its deepest layers. We offer an extensive range of experiments on this, including data from two high-resolution sensors-Carbonite-2 and Landsat 8-and several complementary tests. Owing to a variety of performance-enhancing design choices and training techniques, it exhibits state-of-the-art performance where comparable to other methods, high speed, and robustness to many different terrains and sensor types. © 2019 by the authors." "57188762538;6603267637;57208526154;","Assessing PM 2.5 concentrations in Tehran, Iran, from space using MAIAC, deep blue, and dark target AOD and machine learning algorithms",2019,"10.1016/j.apr.2018.12.017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065037196&doi=10.1016%2fj.apr.2018.12.017&partnerID=40&md5=b4abb457fe976ace7c0f96a3a17488e5","This study aims to explore the spatial estimation of fine particulate matter (PM 2.5 ) using 10-km merged dark target and deep blue (DB_DT) Aerosol Optical Depth (AOD) and 1-km Multi-Angle Implementation of Atmospheric Correction (MAIAC) AOD over Tehran. The ability of four Machine Learning Algorithms (MLAs) to predict PM 2.5 concentrations is also investigated. Results show that the association of satellite AOD with surface PM significantly increases after considering the contribution of relative humidity in PM mass concentration and normalization of AOD to Planetary boundary layer height (PBLH). The examination of derived aerosol layer height (ALH) from 159 Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observation (CALIPSO) profiles shows that PBLH could successfully represent the top of aerosol-laden layers. Surprisingly, the highest correlation was found between normalized 10-km DB_DT AOD and corrected PM 2.5 measurements. Consequently, random forest (RF) fed by this AOD product has yielded the best performance (R 2 = 0.68, RMSE = 17.52 and MRE = 27.46%). Importance analysis of variables reveals that DB_DT and meteorological fields are of highest and least importance among selected variables, respectively. The RF performance is less satisfactory during summer which is assumed to be caused by the omission of unknown features representing the formation of secondary aerosols. The inferior accuracy of estimation in the north and east of Tehran is also linked to lacking features which could feed the transportation of PM 2.5 from west to the east of the study area into MLAs. © 2019 Turkish National Committee for Air Pollution Research and Control" "55798569000;6602169970;7102831991;23060360800;43361574900;","Calibrating the HISA temperature: Measuring the temperature of the Riegel-Crutcher cloud",2018,"10.1093/mnras/sty1384","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054723530&doi=10.1093%2fmnras%2fsty1384&partnerID=40&md5=fde3b481733dde86c745e3d9e74355b9","HI self-absorption (HISA) clouds are clumps of cold neutral hydrogen (H I) visible in front of warm background gas, which makes them ideal places to study the properties of the cold atomic component of the interstellar medium. The Riegel-Crutcher (R-C) cloud is the most striking HISA feature in the Galaxy. It is one of the closest HISA clouds to us and is located in the direction of the Galactic Centre, which provides a bright background. High-resolution interferometric measurements have revealed the filamentary structure of this cloud; however, it is difficult to accurately determine the temperature and the density of the gas without optical depth measurements. In this paper, we present new H I absorption observations with the Australia Telescope Compact Array against 46 continuum sources behind the R-C cloud to directly measure the optical depth of the cloud. We decompose the complex H I absorption spectra into Gaussian components using an automated machine learning algorithm. We find 300 Gaussian components, from which 67 are associated with the R-C cloud (0 < vLSR < 10 km s-1, full width at half maximum < 10 km s-1). Combining the new HI absorption data with HI emission data from previous surveys, we calculate the spin temperature and find it to be between 20 and 80 K. Our measurements uncover a temperature gradient across the cloud with spin temperatures decreasing towards positive Galactic latitudes. We also find three new OH absorption lines associated with the cloud, which support the presence of molecular gas. © 2018 The Author(s)." "36680040200;7004177153;57197811867;6603238769;36635241900;","A Machine Learning approach for automatic land cover mapping from DSLR images over the Maltese Islands",2018,"10.1016/j.envsoft.2017.09.014","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85035037271&doi=10.1016%2fj.envsoft.2017.09.014&partnerID=40&md5=8360363e0d132bc2c84e53cb4da79525","High resolution raster data for land cover mapping or change analysis are normally acquired through satellite or aerial imagery. Apart from the incurred costs, the available files might not have the required temporal resolution. Moreover, cloud cover and atmospheric absorptions may limit the applicability of existing algorithms or reduce their accuracy. This paper presents a novel technique that is capable of mapping garrigue and/or phrygana vegetation as well as karst or ground-armour terrain in photos captured by a digital camera. By including a reference pattern in every frame, the automated method estimates the total area covered by each land type. Pixel based classification is performed by supervised decision tree methods. Although the intention is not to replace traditional surface cover analysis, the proposed technique allows accurate land cover mapping with quantitative estimates to be obtained. Since no expensive hardware or specialised personnel are required, vegetation monitoring of local sites can be carried out cheaply and frequently. The developed proof of concept is tested on photos taken in thirteen different sites across the Maltese Islands. © 2017 Elsevier Ltd" "36170984100;57208984389;","Object-based habitat mapping using very high spatial resolution multispectral and hyperspectral imagery with LiDAR data",2017,"10.1016/j.jag.2017.03.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048075963&doi=10.1016%2fj.jag.2017.03.007&partnerID=40&md5=69868acce260bf698bb3b6671021ad72","This study investigated the combined use of multispectral/hyperspectral imagery and LiDAR data for habitat mapping across parts of south Cumbria, North West England. The methodology adopted in this study integrated spectral information contained in pansharp QuickBird multispectral/AISA Eagle hyperspectral imagery and LiDAR-derived measures with object-based machine learning classifiers and ensemble analysis techniques. Using the LiDAR point cloud data, elevation models (such as the Digital Surface Model and Digital Terrain Model raster) and intensity features were extracted directly. The LiDAR-derived measures exploited in this study included Canopy Height Model, intensity and topographic information (i.e. mean, maximum and standard deviation). These three LiDAR measures were combined with spectral information contained in the pansharp QuickBird and Eagle MNF transformed imagery for image classification experiments. A fusion of pansharp QuickBird multispectral and Eagle MNF hyperspectral imagery with all LiDAR-derived measures generated the best classification accuracies, 89.8 and 92.6% respectively. These results were generated with the Support Vector Machine and Random Forest machine learning algorithms respectively. The ensemble analysis of all three learning machine classifiers for the pansharp QuickBird and Eagle MNF fused data outputs did not significantly increase the overall classification accuracy. Results of the study demonstrate the potential of combining either very high spatial resolution multispectral or hyperspectral imagery with LiDAR data for habitat mapping. © 2017 Elsevier B.V." "56237086200;9036557400;57190373951;56402112700;56224155200;56503083100;","Detection of tropical overshooting cloud tops using himawari-8 imagery",2017,"10.3390/rs9070685","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85022320006&doi=10.3390%2frs9070685&partnerID=40&md5=8abd1b6588842d17c2da0dfab6b4b077","Overshooting convective cloud Top (OT)-accompanied clouds can cause severe weather conditions, such as lightning, strong winds, and heavy rainfall. The distribution and behavior of OTs can affect regional and global climate systems. In this paper, we propose a new approach for OT detection by using machine learning methods with multiple infrared images and their derived features. Himawari-8 satellite images were used as the main input data, and binary detection (OT or nonOT) with class probability was the output of the machine learning models. Three machine learning techniques-random forest (RF), extremely randomized trees (ERT), and logistic regression (LR)-were used to develop OT classification models to distinguish OT from non-OT. The hindcast validation over the Southeast Asia andWest Pacific regions showed that RF performed best, resulting in a mean probabilities of detection (POD) of 77.06% and a mean false alarm ratio (FAR) of 36.13%. Brightness temperature at 11.2 μm (Tb11) and its standard deviation (STD) in a 3 × 3 window size were identified as the most contributing variables for discriminating OT and nonOT classes. The proposed machine learning-based OT detection algorithms produced promising results comparable to or even better than the existing approaches, which are the infrared window (IRW)-texture and water vapor (WV) minus IRW brightness temperature difference (BTD) methods. © 2017 by the authors." "57194655355;57194656372;","The future of geospatial intelligence",2017,"10.1080/10095020.2017.1337318","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85021448233&doi=10.1080%2f10095020.2017.1337318&partnerID=40&md5=6d4d5e00deb9eca7085fa98fbce7f28a","For centuries, humans’ capacity to capture and depict physical space has played a central role in industrial and societal development. However, the digital revolution and the emergence of networked devices and services accelerate geospatial capture, coordination, and intelligence in unprecedented ways. Underlying the digital transformation of industry and society is the fusion of the physical and digital worlds–‘perceptality’–where geospatial perception and reality merge. This paper analyzes the myriad forces that are driving perceptality and the future of geospatial intelligence and presents real-world implications and examples of its industrial application. Applications of sensors, robotics, cameras, machine learning, encryption, cloud computing and other software, and hardware intelligence are converging, enabling new ways for organizations and their equipment to perceive and capture reality. Meanwhile, demands for performance, reliability, and security are pushing compute ‘to the edge’ where real-time processing and coordination are vital. Big data place new restraints on economics, as pressures abound to actually use these data, both in real-time and for longer term strategic analysis and decision-making. These challenges require orchestration between information technology (IT) and operational technology (OT) and synchronization of diverse systems, data-sets, devices, environments, workflows, and people. © 2017 Wuhan University. Published by Taylor & Francis Group." "56566276600;36988356600;6507663359;","Aerial laser scanning and imagery data fusion for road detection in city scale",2015,"10.1109/IGARSS.2015.7326746","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84962556656&doi=10.1109%2fIGARSS.2015.7326746&partnerID=40&md5=a8778be0d95d82203ce2cc9770f373c8","This paper presents a workflow including a novel algorithm for road detection from dense LiDAR fused with high-resolution aerial imagery data. Using a supervised machine learning approach point clouds are firstly classified into one of three groups: building, ground, or unassigned. Ground points are further processed by a novel algorithm to extract a road network. The algorithm exploits the high variance of slope and height of the point data in the direction orthogonal to the road boundaries. Applying the proposed approach on a 40 million point dataset successfully extracted a complex road network with an F-measure of 76.9%. © 2015 IEEE." "56402694100;56985759200;7005725869;57134023000;","Learning from synthetic models for roof style classification in point clouds",2014,"10.1145/2666310.2666407","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84961201554&doi=10.1145%2f2666310.2666407&partnerID=40&md5=d977549ed8b3fc27ff26ab7f786ad49b","Automatic roof style classification using point clouds is useful and can be used as a prior knowledge in various applications, such as the construction of 3D models of real-world buildings. Previous classification approaches usually employ heuristic rules to recognize roof style and are limited to a few roof styles. In this paper, the recognition of roof style is done by a roof style classifier which is trained based on bag of words features extracted from a point cloud. In the computation of bag of words features, a key challenge is the generation of the codebook. Unsupervised learning is often misguided easily by the data and detects uninteresting patterns within the data. In contrast, we propose to integrate existing knowledge of roof structure and cluster the points of target roof styles into several semantic classes which can then be used as code words in the bag of words model. We use synthetic variants of these code words to train a semantics point classifier. We evaluate our approach on two datasets with different levels of degradations. We compare the results of our approach with two unsupervised learning algorithms: K-Means and Gaussian Mixture Model. We show that our approach achieve higher accuracy in classification of the roof styles and maintains consistent performance among different datasets. Copyright 2014 ACM." "22235255500;8213128500;57208222251;56609110900;6603888005;57208121325;","Fusing optical and SAR time series for LAI gap fillingwith multioutput Gaussian processes",2019,"10.1016/j.rse.2019.111452","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074396423&doi=10.1016%2fj.rse.2019.111452&partnerID=40&md5=4cfbca4b6a1c9a997c5f4fdc98f930fe","The availability of satellite optical information is often hampered by the natural presence of clouds, which can be problematic for many applications. Persistent clouds over agricultural fields can mask key stages of crop growth, leading to unreliable yield predictions. Synthetic Aperture Radar (SAR) provides all-weather imagery which can potentially overcome this limitation, but given its high and distinct sensitivity to different surface properties, the fusion of SAR and optical data still remains an open challenge. In this work, we propose the use of Multi-Output Gaussian Process (MOGP) regression, a machine learning technique that learns automatically the statistical relationships among multisensor time series, to detect vegetated areas over which the synergy between SAR-optical imageries is profitable. For this purpose, we use the Sentinel-1 Radar Vegetation Index (RVI) and Sentinel-2 Leaf Area Index (LAI) time series over a study area in north west of the Iberian peninsula. Through a physical interpretation of MOGP trained models, we show its ability to provide estimations of LAI even over cloudy periods using the information shared with RVI, which guarantees the solution keeps always tied to real measurements. Results demonstrate the advantage of MOGP especially for long data gaps, where optical-based methods notoriously fail. The leave-one-image-out assessment technique applied to the whole vegetation cover shows MOGP predictions improve standard GP estimations over short-time gaps (R2 of 74% vs 68%, RMSE of 0.4 vs 0.44 [m2m−2]) and especially over long-time gaps (R2 of 33% vs 12%, RMSE of 0.5 vs 1.09 [m2m−2]). A second assessment is focused on crop-specific regions, clustering pixels fulfilling specific model conditions where the synergy is profitable. Results reveal the MOGP performance is crop type and crop stage dependent. For long time gaps, best R2 are obtained over maize, ranging from 0.1 (tillering) to 0.36 (development) up to 0.81 (maturity); for moderate time gap, R2 = 0.93 (maturity) is obtained. Crops such as wheat, oats, rye and barley, can profit from the LAI-RVI synergy, with R2 varying between 0.4 and 0.6. For beet or potatoes, MOGP provides poorer results, but alternative descriptors to RVI should be tested for these specific crops in the future before discarding synergy real benefits. In conclusion, active-passive sensor fusion with MOGP represents a novel and promising approach to cope with crop monitoring over cloud-dominated areas. © 2019 Elsevier Inc." "7401931279;57201113391;57194601941;55696622200;","SEN12MS – A CURATED DATASET of GEOREFERENCED MULTI-SPECTRAL SENTINEL-1/2 IMAGERY for DEEP LEARNING and DATA FUSION",2019,"10.5194/isprs-annals-IV-2-W7-153-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084684118&doi=10.5194%2fisprs-annals-IV-2-W7-153-2019&partnerID=40&md5=a594132c1ed18fcccd585a917a8d3bba","The availability of curated large-scale training data is a crucial factor for the development of well-generalizing deep learning methods for the extraction of geoinformation from multi-sensor remote sensing imagery. While quite some datasets have already been published by the community, most of them suffer from rather strong limitations, e.g. regarding spatial coverage, diversity or simply number of available samples. Exploiting the freely available data acquired by the Sentinel satellites of the Copernicus program implemented by the European Space Agency, as well as the cloud computing facilities of Google Earth Engine, we provide a dataset consisting of 180,662 triplets of dual-pol synthetic aperture radar (SAR) image patches, multi-spectral Sentinel-2 image patches, and MODIS land cover maps. With all patches being fully georeferenced at a 10 m ground sampling distance and covering all inhabited continents during all meteorological seasons, we expect the dataset to support the community in developing sophisticated deep learning-based approaches for common tasks such as scene classification or semantic segmentation for land cover mapping. © 2019 ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. All rights reserved." "55176818100;7004479957;","Spatially Extended Tests of a Neural Network Parametrization Trained by Coarse-Graining",2019,"10.1029/2019MS001711","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070870530&doi=10.1029%2f2019MS001711&partnerID=40&md5=8de2eb403fdab22ecc47983fc4c761ae","General circulation models (GCMs) typically have a grid size of 25–200 km. Parametrizations are used to represent diabatic processes such as radiative transfer and cloud microphysics and account for subgrid-scale motions and variability. Unlike traditional approaches, neural networks (NNs) can readily exploit recent observational data sets and global cloud-system resolving model (CRM) simulations to learn subgrid variability. This article describes an NN parametrization trained by coarse-graining a near-global CRM simulation with a 4-km horizontal grid spacing. The NN predicts the residual heating and moistening averaged over (160 km)2 grid boxes as a function of the coarse-resolution fields within the same atmospheric column. This NN is coupled to the dynamical core of a GCM with the same 160-km resolution. A recent study described how to train such an NN to be stable when coupled to specified time-evolving advective forcings in a single-column model, but feedbacks between NN and GCM components cause spatially extended simulations to crash within a few days. Analyzing the linearized response of such an NN reveals that it learns to exploit a strong synchrony between precipitation and the atmospheric state above 10 km. Removing these variables from the NN's inputs stabilizes the coupled simulations, which predict the future state more accurately than a coarse-resolution simulation without any parametrizations of subgrid-scale variability, although the mean state slowly drifts. ©2019. The Authors." "26639121800;57204325611;6506993025;36172338400;6701323899;12767844100;7102775303;55198994700;55735863100;","Census of ρ Ophiuchi candidate members from Gaia Data Release 2",2019,"10.1051/0004-6361/201935321","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068310949&doi=10.1051%2f0004-6361%2f201935321&partnerID=40&md5=7ec112407e652513099939aa30e2e848","Context. The Ophiuchus cloud complex is one of the best laboratories to study the earlier stages of the stellar and protoplanetary disc evolution. The wealth of accurate astrometric measurements contained in the Gaia Data Release 2 can be used to update the census of Ophiuchus member candidates. Aims. We seek to find potential new members of Ophiuchus and identify those surrounded by a circumstellar disc. Methods. We constructed a control sample composed of 188 bona fide Ophiuchus members. Using this sample as a reference we applied three different density-based machine learning clustering algorithms (DBSCAN, OPTICS, and HDBSCAN) to a sample drawn from the Gaia catalogue centred on the Ophiuchus cloud. The clustering analysis was applied in the five astrometric dimensions defined by the three-dimensional Cartesian space and the proper motions in right ascension and declination. Results. The three clustering algorithms systematically identify a similar set of candidate members in a main cluster with astrometric properties consistent with those of the control sample. The increased flexibility of the OPTICS and HDBSCAN algorithms enable these methods to identify a secondary cluster. We constructed a common sample containing 391 member candidates including 166 new objects, which have not yet been discussed in the literature. By combining the Gaia data with 2MASS and WISE photometry, we built the spectral energy distributions from 0.5 to 22 μm for a subset of 48 objects and found a total of 41 discs, including 11 Class II and 1 Class III new discs. Conclusions. Density-based clustering algorithms are a promising tool to identify candidate members of star forming regions in large astrometric databases. By combining the Gaia data with infrared catalogues, it is possible to discover new protoplanetary discs. If confirmed, the candidate members discussed in this work would represent an increment of roughly 40-50% of the current census of Ophiuchus. © ESO 2019." "57201549485;7003656992;","Creating 3D city models with building footprints and LIDAR point cloud classification: A machine learning approach",2019,"10.1016/j.compenvurbsys.2019.01.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85060525776&doi=10.1016%2fj.compenvurbsys.2019.01.004&partnerID=40&md5=56d4e0f32197f36009e19d338a596220","The increasing availability of open geospatial data, such as building footprint vector data and LiDAR (Light Detection and Ranging) point clouds, has provided opportunities to generate large-scale 3D city models at low cost. However, using unclassified point clouds with building footprints to estimate building heights may yield erroneous results due to potential errors and anomalies in both datasets and their integration. Some of the points within footprints often reflect irrelevant objects other than roofs, leading to biases in height estimation, and few studies have developed systematic methods to filter them out. In this paper, a LiDAR point classification methodology is proposed that extracts only rooftop points for building height estimation. The LiDAR points are characterized by point, footprint, and neighborhood-based features and classified by the Random Forest (RF) algorithm into four classes – rooftop, wall, ground, and high outlier. The percentage of correctly classified points among 15,577 sample points in Columbus, Ohio, amounts to 96.5%. Conducting this classification separately for different building types (commercial, residential, skyscraper, and small constructions) does not significantly change the overall accuracy. The footprint-based features contribute most to predicting the classes correctly. Height validation results based on a sample of 498 buildings show that (1) using average or median heights with classified points provides the best estimates, minimizing the disparities between computed heights and ground truth and (2) the RF method yields outcomes much closer to ground truth than earlier classification approaches. Some outcomes are visualized in 3D format using Google Earth 3D Imagery and ArcScene. © 2019 Elsevier Ltd" "57208165544;57218145290;57204964406;56479914700;6701446204;6701518328;7007038828;56841699000;57202561981;","Developing an advanced PM2.5 exposure model in Lima, Peru",2019,"10.3390/rs11060641","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063997197&doi=10.3390%2frs11060641&partnerID=40&md5=87b2eccc9f1ecd7f6d5f0220b2e291de","It is well recognized that exposure to fine particulate matter (PM2.5) affects health adversely, yet few studies from South America have documented such associations due to the sparsity of PM2.5 measurements. Lima's topography and aging vehicular fleet results in severe air pollution with limited amounts of monitors to effectively quantify PM2.5 levels for epidemiologic studies. We developed an advanced machine learning model to estimate daily PM2.5 concentrations at a 1 km2 spatial resolution in Lima, Peru from 2010 to 2016. We combined aerosol optical depth (AOD), meteorological fields from the European Centre for Medium-Range Weather Forecasts (ECMWF), parameters from theWeather Research and Forecasting model coupled with Chemistry (WRF-Chem), and land use variables to fit a random forest model against ground measurements from 16 monitoring stations. Overall cross-validation R2 (and root mean square prediction error, RMSE) for the random forest model was 0.70 (5.97 μg/m3). Mean PM2.5 for ground measurements was 24.7 μg/m3 while mean estimated PM2.5 was 24.9 μg/m3 in the cross-validation dataset. The mean difference between ground and predicted measurements was -0.09 μg/m3 (Std.Dev. = 5.97 μg/m3), with 94.5% of observations falling within 2 standard deviations of the difference indicating good agreement between ground measurements and predicted estimates. Surface downwards solar radiation, temperature, relative humidity, and AOD were the most important predictors, while percent urbanization, albedo, and cloud fraction were the least important predictors. Comparison of monthly mean measurements between ground and predicted PM2.5 shows good precision and accuracy from our model. Furthermore, mean annual maps of PM2.5 show consistent lower concentrations in the coast and higher concentrations in the mountains, resulting from prevailing coastal winds blown from the Pacific Ocean in the west. Our model allows for construction of long-term historical daily PM2.5 measurements at 1 km2 spatial resolution to support future epidemiological studies. © 2019 by the authors." "48861439400;55207460700;57214256006;57205419896;35262347300;","Gap-Filling of MODIS fractional snow cover products via non-local spatio-temporal filtering based on machine learning techniques",2019,"10.3390/rs11010090","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059941339&doi=10.3390%2frs11010090&partnerID=40&md5=44b4433ac9b197ad862177a3ec86337c","Cloud obscuration leaves significant gaps in MODIS snow cover products. In this study, an innovative gap-filling method based on the concept of non-local spatio-temporal filtering (NSTF) is proposed to reconstruct the cloud gaps in MODIS fractional snow cover (SCF) products. The ground information of a gap pixel was estimated by using the appropriate similar pixels in the remaining known part of an image via an automatic machine learning technique. We take the MODIS SCF product cloud gap filling data from 2001 to 2016 in Northern Xinjiang, China as an example. The results demonstrate that the methodology can generate almost continuous spatio-temporal, daily MODIS SCF images, and it leaves only 0.52% of cloud gaps long-term, on average. The validation results based on ""cloud assumption"" exhibit high accuracy, with a higher R 2 exceeding 0.8, a lower RMSE of 0.1, an overestimated error of 1.13%, an underestimated error of 1.4%, and a spatial efficiency (SPAEF) of 0.78. The validation based on 50 in situ snow depth observations demonstrates the superiority of the methodology in terms of accuracy and consistency. The overall accuracy is 93.72%. The average omission and commission error have increased approximately 1.16 and 0.53% compared with the original MODIS SCF products under a clear sky term. © 2019 by the authors." "56151374100;56227666500;6602574676;7006954443;7403441497;57195007299;7404614089;7006393267;","New neural network cloud mask algorithm based on radiative transfer simulations",2018,"10.1016/j.rse.2018.09.029","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054464404&doi=10.1016%2fj.rse.2018.09.029&partnerID=40&md5=b8391a6d7d3f1c58c5104a9f998e71cc","Cloud detection and screening constitute critically important first steps required to derive many satellite data products. Traditional threshold-based cloud mask algorithms require a complicated design process and fine tuning for each sensor, and they have difficulties over areas partially covered with snow/ice. Exploiting advances in machine learning techniques and radiative transfer modeling of coupled environmental systems, we have developed a new, threshold-free cloud mask algorithm based on a neural network classifier driven by extensive radiative transfer simulations. Statistical validation results obtained by using collocated CALIOP and MODIS data show that its performance is consistent over different ecosystems and significantly better than the MODIS Cloud Mask (MOD35 C6) during the winter seasons over snow-covered areas in the mid-latitudes. Simulations using a reduced number of satellite channels also show satisfactory results, indicating its flexibility to be configured for different sensors. Compared to threshold-based methods and previous machine-learning approaches, this new cloud mask (i) does not rely on thresholds, (ii) needs fewer satellite channels, (iii) has superior performance during winter seasons in mid-latitude areas, and (iv) can easily be applied to different sensors. © 2018 Elsevier Inc." "24177644800;22234792700;6506680683;","Modelling LiDAR derived tree canopy height from Landsat TM, ETM+ and OLI satellite imagery—A machine learning approach",2018,"10.1016/j.jag.2018.08.013","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063916703&doi=10.1016%2fj.jag.2018.08.013&partnerID=40&md5=3b6f946bd5cbec0b1491f03638e26c9c","Understanding ecological changes in native vegetation communities often requires information over long time periods (multiple decades). Tropical cyclones can have a major impact on woody vegetation structure across northern Australia, however understanding the impacts on woody vegetation structure is limited. Woody vegetation structural attributes such as height are used in ecological studies to identify long term changes and trends. LiDAR has been used to measure woody vegetation structure, however LiDAR datasets cover relatively small areas and historical coverage is restricted, limiting the use of this technology for monitoring long-term change. The Landsat archive spans multiple decades and is suitable for regional/continental assessment. Advances in predictive modelling using machine learning algorithms have enabled complex relationships between dependent and independent variables to be identified. The aim of this study is to develop a predictive model to estimate woody vegetation height from Landsat imagery to assist in understanding change through space and time. A LiDAR canopy height model was produced covering a range of vegetation communities in northern Australia (Darwin region) for use as the dependent variable. A random forest regression model was developed to predict mean LiDAR canopy height (30 m spatial resolution) from Landsat-5 Thematic Mapper (TM). Validation of the random forest model was undertaken on independent data (n = 30,500) resulting in an overall R2 = 0.53, RMSE of 2.8 m. Assessment of the RMSE within four broad vegetation communities ranged from 2.5 to 3.7 m with the two dominant communities in the study area Mangrove forests and Eucalyptus communities recording an RMSE value of 2.9 m and 2.5 m respectively. The model was also applied to Landsat-7 Enhanced Thematic Mapper Plus (ETM+) resulting in an R2 of 0.49, RMSE of 2.8 m. The model was then applied to all cloud free Landsat-5 TM, Landsat-7 ETM+ and Landsat-8 Operational Land Imager (OLI) imagery (106/69 path/row) available between the months April, May and June for 1987 to 2016 to produce annual estimates (29 years) of canopy height. A number of time traces were produced to illustrate tree canopy height through time in the Darwin region which was severely impacted by cyclone (hurricane) Tracy on the 25th December 1974. © 2018 Elsevier B.V." "56735478500;55628589750;56531367400;","Building a cloud in the southeast Atlantic: Understanding low-cloud controls based on satellite observations with machine learning",2018,"10.5194/acp-18-16537-2018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057180840&doi=10.5194%2facp-18-16537-2018&partnerID=40&md5=aecd9d62a838cfc0f2d81a6c24255e02","Understanding the processes that determine low-cloud properties and aerosol-cloud interactions (ACIs) is crucial for the estimation of their radiative effects. However, the covariation of meteorology and aerosols complicates the determination of cloud-relevant influences and the quantification of the aerosol-cloud relation. This study identifies and analyzes sensitivities of cloud fraction and cloud droplet effective radius to their meteorological and aerosol environment in the atmospherically stable southeast Atlantic during the biomass-burning season based on an 8-day-averaged data set. The effect of geophysical parameters on clouds is investigated based on a machine learning technique, gradient boosting regression trees (GBRTs), using a combination of satellite and reanalysis data as well as trajectory modeling of air-mass origins. A comprehensive, multivariate analysis of important drivers of cloud occurrence and properties is performed and evaluated. The statistical model reveals marked subregional differences of relevant drivers and processes determining low clouds in the southeast Atlantic. Cloud fraction is sensitive to changes of lower tropospheric stability in the oceanic, southwestern subregion, while in the northeastern subregion it is governed mostly by surface winds. In the pristine, oceanic subregion large-scale dynamics and aerosols seem to be more important for changes of cloud droplet effective radius than in the polluted, near-shore subregion, where free tropospheric temperature is more relevant. This study suggests the necessity to consider distinct ACI regimes in cloud studies in the southeast Atlantic. © Author(s) 2018." "57196180188;23089590700;7006518879;55709751800;6603285262;","Monitoring the dynamics of surface water fraction from MODIS time series in a Mediterranean environment",2018,"10.1016/j.jag.2017.11.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85055434285&doi=10.1016%2fj.jag.2017.11.007&partnerID=40&md5=9f4b9054b8173cc172a300fd96a8a6c0","Detailed spatial information of changes in surface water extent is needed for water management and biodiversity conservation, particularly in drier parts of the globe where small, temporally-variant wetlands prevail. Although global surface water histories are now generated from 30 m Landsat data, for many locations they contain large temporal gaps particularly for longer periods (>10 years) due to revisit intervals and cloud cover. Daily Moderate Resolution Imaging Spectrometer (MODIS) imagery has potential to fill such gaps, but its relatively coarse spatial resolution may not detect small water bodies, which can be of great ecological importance. To address this problem, this study proposes and tests options for estimating the surface water fraction from MODIS 16-day 500 m Bidirectional Reflectance Distribution Function (BRDF) corrected surface reflectance image composites. The spatial extent of two Landsat tiles over Spain were selected as test areas. We obtained a 500 m reference dataset on surface water fraction by spatially aggregating 30 m binary water masks obtained from the Landsat-derived C-version of Function of Mask (CFmask), which themselves were evaluated against high-resolution Google Earth imagery. Twelve regression tree models were developed with two approaches, Random Forest and Cubist, using spectral metrics derived from MODIS data and topographic parameters generated from a 30 m spatial resolution digital elevation model. Results showed that accuracies were higher when we included annual summary statistics of the spectral metrics as predictor variables. Models trained on a single Landsat tile were ineffective in mapping surface water in the other tile, but global models trained with environmental conditions from both tiles can provide accurate results for both study areas. We achieved the highest accuracy with Cubist global model (R2 = 0.91, RMSE = 11.05%, MAE = 7.67%). Our method was not only effective for mapping permanent water fraction, but also in accurately capturing temporal fluctuations of surface water. Based on this good performance, we produced surface water fraction maps at 16-day interval for the 2000–2015 MODIS archive. Our approach is promising for monitoring surface water fraction at high frequency time intervals over much larger regions provided that training data are collected across the spatial domain for which the model will be applied. © 2017 Elsevier B.V." "56595213400;57215912664;55619939300;57191878392;","Machine learning algorithms: A background artifact",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067271010&partnerID=40&md5=b3ef76e81f10132f7ca842789650f9d7","With the greater development of technology and automation human history is predominantly updated. The technology movement shifted from large mainframes to PCs to cloud when computing the available data for a larger period. This has happened only due to the advent of many tools and practices, that elevated the next generation in computing. A large number of techniques has been developed so far to automate such computing. Research dragged towards training the computers to behave similar to human intelligence. Here the diversity of machine learning came into play for knowledge discovery. Machine Learning (ML) is applied in many areas such as medical, market-ing, telecommunications, and stock, health care and so on. This paper presents reviews about machine learning algorithm foundations, its types and flavors together with R code and Python scripts possibly for each machine learning techniques. © 2018 Authors." "55561559600;57201378591;57201373924;57201372901;6507534695;","CLASSIFICATION of AERIAL PHOTOGRAMMETRIC 3D POINT CLOUDS",2017,"10.5194/isprs-annals-IV-1-W1-3-2017","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85029759044&doi=10.5194%2fisprs-annals-IV-1-W1-3-2017&partnerID=40&md5=8bd2fae7e1dc4b8dd2c9f931e214a7da","We present a powerful method to extract per-point semantic class labels from aerial photogrammetry data. Labelling this kind of data is important for tasks such as environmental modelling, object classification and scene understanding. Unlike previous point cloud classification methods that rely exclusively on geometric features, we show that incorporating color information yields a significant increase in accuracy in detecting semantic classes. We test our classification method on three real-world photogrammetry datasets that were generated with Pix4Dmapper Pro, and with varying point densities. We show that off-the-shelf machine learning techniques coupled with our new features allow us to train highly accurate classifiers that generalize well to unseen data, processing point clouds containing 10 million points in less than 3 minutes on a desktop computer. © 2017 Copernicus GmbH. All rights reserved." "54381297800;7006336226;55612371100;55251096900;56052000200;57206190262;","A security-awareness virtual machine management scheme based on Chinese wall policy in cloud computing",2014,"10.1155/2014/805923","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84896884892&doi=10.1155%2f2014%2f805923&partnerID=40&md5=4282467bf04a386e9a40c0e90803fc25","Cloud computing gets increasing attention for its capacity to leverage developers from infrastructure management tasks. However, recent works reveal that side channel attacks can lead to privacy leakage in the cloud. Enhancing isolation between users is an effective solution to eliminate the attack. In this paper, to eliminate side channel attacks, we investigate the isolation enhancement scheme from the aspect of virtual machine (VM) management. The security-awareness VMs management scheme (SVMS), a VMs isolation enhancement scheme to defend against side channel attacks, is proposed. First, we use the aggressive conflict of interest relation (ACIR) and aggressive in ally with relation (AIAR) to describe user constraint relations. Second, based on the Chinese wall policy, we put forward four isolation rules. Third, the VMs placement and migration algorithms are designed to enforce VMs isolation between the conflict users. Finally, based on the normal distribution, we conduct a series of experiments to evaluate SVMS. The experimental results show that SVMS is efficient in guaranteeing isolation between VMs owned by conflict users, while the resource utilization rate decreases but not by much. © 2014 Si Yu et al." "55995014200;55553737066;57189042582;","Prediction based proactive thermal virtual machine scheduling in green clouds",2014,"10.1155/2014/208983","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84897558406&doi=10.1155%2f2014%2f208983&partnerID=40&md5=6214e3a2d2a0a09b7e4a2426789d9f16","Cloud computing has rapidly emerged as a widely accepted computing paradigm, but the research on Cloud computing is still at an early stage. Cloud computing provides many advanced features but it still has some shortcomings such as relatively high operating cost and environmental hazards like increasing carbon footprints. These hazards can be reduced up to some extent by efficient scheduling of Cloud resources. Working temperature on which a machine is currently running can be taken as a criterion for Virtual Machine (VM) scheduling. This paper proposes a new proactive technique that considers current and maximum threshold temperature of Server Machines (SMs) before making scheduling decisions with the help of a temperature predictor, so that maximum temperature is never reached. Different workload scenarios have been taken into consideration. The results obtained show that the proposed system is better than existing systems of VM scheduling, which does not consider current temperature of nodes before making scheduling decisions. Thus, a reduction in need of cooling systems for a Cloud environment has been obtained and validated. © 2014 Supriya Kinger et al." "7801425389;6602999062;16300591200;6603109490;","Risk assessment of atmospheric emissions using machine learning",2008,"10.5194/nhess-8-991-2008","https://www.scopus.com/inward/record.uri?eid=2-s2.0-52149105987&doi=10.5194%2fnhess-8-991-2008&partnerID=40&md5=b0eeddd7665d2540a6649c1031021a34","Supervised and unsupervised machine learning algorithms are used to perform statistical and logical analysis of several transport and dispersion model runs which simulate emissions from a fixed source under different atmospheric conditions. First, a clustering algorithm is used to automatically group the results of different transport and dispersion simulations according to specific cloud characteristics. Then, a symbolic classification algorithm is employed to find complex non-linear relationships between the meteorological input conditions and each cluster of clouds. The patterns discovered are provided in the form of probabilistic measures of contamination, thus suitable for result interpretation and dissemination. The learned patterns can be used for quick assessment of the areas at risk and of the fate of potentially hazardous contaminants released in the atmosphere." "57216873082;23989756500;57201837594;","Landsat-8 and Sentinel-2 based Forest fire burn area mapping using machine learning algorithms on GEE cloud platform over Uttarakhand, Western Himalaya",2020,"10.1016/j.rsase.2020.100324","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085095632&doi=10.1016%2fj.rsase.2020.100324&partnerID=40&md5=6ef3715da79e3f600ba9a6287b18293a","The accurate quantitative and qualitative estimation of burn-area are crucial to analyze the impact of fire on forest. The medium resolution optical-satellite imagery of Landsat-8 and Sentinel-2 are employed covering the period 2016 to 2019 for forest fire patches identification on Google Earth Engine (GEE). The most indispensable season of Forest Fire (FF) is pre-monsoon in Uttarakhand, western Himalaya, India. Bi-temporal (pre and post fire) reflectance contrast of burn-sensitive spectral bands was used to compute differential spectral indices, namely, Normalized Burn Ratio (dNBR), Normalized Difference Vegetation Index (dNDVI), Normalized Difference Water Index (dNDWI), and Short-Wave Infrared (dSWIR). The differential spectral-indices composite is further used as an input to unsupervised Weka clustering algorithms for capturing the shape and pattern of fire patches. Sample training-data of burn and unburn classes were collected with reference to thermal and optical spectral principle. Classification Regression Tree (CART), Random Forest (RF), and Support Vector Machine (SVM) algorithms have been employed to identify FF. The key findings revealed that CART and RF algorithms displayed similar forest fire patches with an overall accuracy of 97–100%. The classification accuracy is slightly lower in SVM and its underestimating forest fire patches detections. Landsat-8 OLI derived burn area was fitted better with fire product of Climate Change Initiative (Fire-CCI of ESA) and MCD64A1 of MODIS burn area product with R-square of 0.71–0.93 and 0.62–0.91, respectively which attributed to better spectral bands of Landsat-8 than the Sentinel-2. However, Sentinel-2 bands have the potential to capture fire patches during post-fire events. This study has demonstrated the potential utilities of combined effort of unsupervised and supervised algorithms on Landsat-8 and Sentinel-2 on GEE to identify fire patches. © 2020" "36938152500;16021918800;57199644951;57215652608;","Super-resolution of PROBA-V images using convolutional neural networks",2019,"10.1007/s42064-019-0059-8","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081573720&doi=10.1007%2fs42064-019-0059-8&partnerID=40&md5=d7e98b1dab609f8e07ea71566a986583","European Space Aqency (ESA)’s PROBA-V Earth observation (EO) satellite enables us to monitor our planet at a large scale to study the interaction between vegetation and climate, and provides guidance for important decisions on our common global future. However, the interval at which high-resolution images are recorded spans over several days, in contrast to the availability of lower-resolution images which is often daily. We collect an extensive dataset of both high- and low-resolution images taken by PROBA-V instruments during monthly periods to investigate Multi Image Super-resolution, a technique to merge several low-resolution images into one image of higher quality. We propose a convolutional neural network (CNN) that is able to cope with changes in illumination, cloud coverage, and landscape features which are introduced by the fact that the different images are taken over successive satellite passages at the same region. Given a bicubic upscaling of low resolution images taken under optimal conditions, we find the Peak Signal to Noise Ratio of the reconstructed image of the network to be higher for a large majority of different scenes. This shows that applied machine learning has the potential to enhance large amounts of previously collected EO data during multiple satellite passes. © 2019, Tsinghua University Press." "16203966600;57201290072;35576373500;55989619100;42161076200;41762495400;57209161159;57217643830;57214381300;57201030837;57190971137;35291520900;57212136051;57212145681;57217348546;57193029474;57202232564;57211316438;45661429600;24462074700;55324316900;6505843293;16202693300;12544553500;57202261421;","Land Cover Mapping in Data Scarce Environments: Challenges and Opportunities",2019,"10.3389/fenvs.2019.00150","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076030442&doi=10.3389%2ffenvs.2019.00150&partnerID=40&md5=589463082cd55231574205d523fc1152","Land cover maps are a critical component to make informed policy, development, planning, and resource management decisions. However, technical, capacity, and institutional challenges inhibit the creation of consistent and relevant land cover maps for use in developing regions. Many developing regions lack coordinated capacity, infrastructure, and technologies to produce a robust land cover monitoring system that meets land management needs. Local capacity may be replaced by external consultants or methods which lack long-term sustainability. In this study, we characterize and respond to the key land cover mapping gaps and challenges encountered in the Lower Mekong (LMR) and Hindu Kush-Himalaya (HKH) region through a needs assessment exercise and a collaborative system design. Needs were assessed using multiple approaches, including focus groups, user engagement workshops, and online surveys. Efforts to understand existing limitations and stakeholder needs resulted in a co-developed and modular land cover monitoring system which utilizes state-of-the-art cloud computing and machine learning which leverages freely available Earth observations. This approach meets the needs of diverse actors and is a model for transnational cooperation. © Copyright © 2019 Saah, Tenneson, Matin, Uddin, Cutter, Poortinga, Nguyen, Patterson, Johnson, Markert, Flores, Anderson, Weigel, Ellenberg, Bhargava, Aekakkararungroj, Bhandari, Khanal, Housman, Potapov, Tyukavina, Maus, Ganz, Clinton and Chishtie." "57188582312;55936805100;57195935005;23005893600;","Classification of ALS Point Clouds Using End-to-End Deep Learning",2019,"10.1007/s41064-019-00073-0","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070969790&doi=10.1007%2fs41064-019-00073-0&partnerID=40&md5=a71b783eac179d5b5cea0b3ab8daab04","Deep learning, referring to artificial neural networks with multiple layers, is widely used for classification tasks in many disciplines including computer vision. The most popular type is the Convolutional Neural Network (CNN), commonly applied to 2D image data. However, CNNs are difficult to adapt to irregular data like point clouds. PointNet, on the other hand, has enabled the derivation of features based on the geometric distribution of a set of points in nD-space utilising a neural network. We use PointNet on multiple scales to automatically learn a representation of local neighbourhoods in an end-to-end fashion, which is optimised for semantic labelling on 3D point clouds acquired by Airborne Laser Scanning (ALS). The results are comparable to those using manually crafted features, suggesting a successful representation of these neighbourhoods. On the ISPRS 3D Semantic Labelling benchmark, we achieve 80.6% overall accuracy, a mid-field result. Investigation on a bigger dataset, namely the 2011 ALS point cloud of the federal state of Vorarlberg, shows overall accuracies of up to 95.8% over large-scale built-up areas. Lower accuracy is achieved for the separation of low vegetation and ground points, presumably because of invalid assumptions about the distribution of classes in space, especially in high alpine regions. We conclude that the method of the end-to-end system, allowing training on a big variety of classification problems without the need for expert knowledge about neighbourhood features can also successfully be applied to single-point-based classification of ALS point clouds. © 2019, Deutsche Gesellschaft für Photogrammetrie, Fernerkundung und Geoinformation (DGPF) e.V." "35772399000;57208483290;","Cloud forecasting system for monitoring and alerting of energy use by home appliances",2019,"10.1016/j.apenergy.2019.04.063","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064952748&doi=10.1016%2fj.apenergy.2019.04.063&partnerID=40&md5=f1c4e431f7892476bd35969d5fac6535","In recent years, energy information systems have had an important role in the operational optimization of intelligent buildings to provide such benefits as high efficiency, energy savings and smart services. Interest in the intelligent management of home energy consumption using data mining and time series analysis is increasing. Therefore, this work develops an efficient web-based energy information management system for the power consumption of home appliances that monitors the energy load of a home, analyzes its energy consumption based on machine learning, and then sends information to various stakeholders. It interacts with the end-user through energy dashboards and emails. The web-based system includes a novel hybrid artificial intelligence model to improve its prediction of energy usage. An automatic warning function is also developed to identify anomalous energy consumption in a home in real time. The cloud system automatically sends a message to the user's email whenever a warning is necessary. End-users of this system can use forecast information and anomalous data to enhance the efficiency of energy usage in their buildings especially during peak times by adjusting the operating schedule of their appliances and electrical equipment. © 2019 Elsevier Ltd" "55966388900;35115334500;57202621598;6602209960;9737845400;23970956600;56604418600;7004260140;","Semantic segmentation of road furniture in mobile laser scanning data",2019,"10.1016/j.isprsjprs.2019.06.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066941388&doi=10.1016%2fj.isprsjprs.2019.06.001&partnerID=40&md5=787817b0aea04edb19aa188d69e00186","Road furniture recognition has become a prevalent issue in the past few years because of its great importance in smart cities and autonomous driving. Previous research has especially focussed on pole-like road furniture, such as traffic signs and lamp posts. Published methods have mainly classified road furniture as individual objects. However, most road furniture consists of a combination of classes, such as a traffic sign mounted on a street light pole. To tackle this problem, we propose a framework to interpret road furniture at a more detailed level. Instead of being interpreted as single objects, mobile laser scanning data of road furniture is decomposed in elements individually labelled as poles, and objects attached to them, such as, street lights, traffic signs and traffic lights. In our framework, we first detect road furniture from unorganised mobile laser scanning point clouds. Then detected road furniture is decomposed into poles and attachments (e.g. traffic signs). In the interpretation stage, we extract a set of features to classify the attachments by utilising a knowledge-driven method and four representative types of machine learning classifiers, which are random forest, support vector machine, Gaussian mixture model and naïve Bayes, to explore the optimal method. The designed features are the unary features of attachments and the spatial relations between poles and their attachments. Two experimental test sites in Enschede dataset and Saunalahti dataset were applied, and Saunalahti dataset was collected in two different epochs. In the experimental results, the random forest classifier outperforms the other methods, and the overall accuracy acquired is higher than 80% in Enschede test site and higher than 90% in both Saunalahti epochs. The designed features play an important role in the interpretation of road furniture. The results of two epochs in the same area prove the high reliability of our framework and demonstrate that our method achieves good transferability with an accuracy over 90% through employing the training data of one epoch to test the data in another epoch. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "24469129600;9279987000;15061019900;7006071339;56650915200;7201722804;7006347612;7006392180;6507886390;6508238265;","Reducing the effects of vegetation phenology on change detection in tropical seasonal biomes",2019,"10.1080/15481603.2018.1550245","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057616433&doi=10.1080%2f15481603.2018.1550245&partnerID=40&md5=27c2bf92856226887d1ec5432d07c7d1","Tropical seasonal biomes (TSBs), such as the savannas (Cerrado) and semi-arid woodlands (Caatinga) of Brazil, are vulnerable ecosystems to human-induced disturbances. Remote sensing can detect disturbances such as deforestation and fires, but the analysis of change detection in TSBs is affected by seasonal modifications in vegetation indices due to phenology. To reduce the effects of vegetation phenology on changes caused by deforestation and fires, we developed a novel object-based change detection method. The approach combines both the spatial and spectral domains of the normalized difference vegetation index (NDVI), using a pair of Operational Land Imager (OLI)/Landsat-8 images acquired in 2015 and 2016. We used semivariogram indices (SIs) as spatial features and descriptive statistics as spectral features (SFs). We tested the performance of the method using three machine-learning algorithms: support vector machine (SVM), artificial neural network (ANN) and random forest (RF). The results showed that the combination of spatial and spectral information improved change detection by correctly classifying areas with seasonal changes in NDVI caused by vegetation phenology and areas with NDVI changes caused by human-induced disturbances. The use of semivariogram indices reduced the effects of vegetation phenology on change detection. The performance of the classifiers was generally comparable, but the SVM presented the highest overall classification accuracy (92.27%) when using the hybrid set of NDVI-derived spectral-spatial features. From the vegetated areas, 18.71% of changes were caused by human-induced disturbances between 2015 and 2016. The method is particularly useful for TSBs where vegetation exhibits strong seasonality and regularly spaced time series of satellite images are difficult to obtain due to persistent cloud cover. © 2018, © 2018 Informa UK Limited, trading as Taylor & Francis Group." "57200606418;57193801698;6701316969;57208699587;34882105800;","Joint estimation of Plant Area Index (PAI) and wet biomass in wheat and soybean from C-band polarimetric SAR data",2019,"10.1016/j.jag.2019.02.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85062866797&doi=10.1016%2fj.jag.2019.02.007&partnerID=40&md5=3acc4f3b931683bc9bfa5a539c90e706","Retrieval of the Plant Area Index (PAI) and wet biomass from polarimetric SAR (PolSAR) data is of paramount importance for in-season monitoring of crop growth. Notably, the joint estimation of biophysical parameters might be effective instead of an individual parameter due to their inherent relationships (possibly nonlinear). The semi-empirical water cloud model (WCM) can be suitably utilized to estimate biophysical parameters from PolSAR data. Nevertheless, instability problems could occur during the model inversion process using traditional inversion approaches. Iterative optimization (IO) can have difficulty in finding the global minima while look up table (LUT) searches have a lower generalization capability. These challenges reduce the transferability of IO and LUT search inversions in computational efficiency and seldom account for the inter-correlation among the parameters. Alternatively, a machine learning regression technique with a regularization routine may provide a stable and optimum solution for ill-posed problems related to the inversion of the WCM. In the present work, the crop biophysical parameters viz. PAI and wet biomass are estimated simultaneously using the multi-target Random Forest Regression (MTRFR) technique. The accuracy of the retrieval method is analyzed using the in-situ measurements and quad-pol RADARSAT-2 data acquired during the SMAPVEX16 campaign over Manitoba, Canada. The inversion process is tested with different polarization combinations of SAR data for wheat and soybean. The validation used ground measured biophysical parameters for various crops, indicating promising results with a correlation coefficient (r) in the range of 0.6–0.8. In addition, the relationship between PAI and wet biomass using the multi-target and single output model is also assessed based on in-situ measurements. The results confirm that the inter-correlation between biophysical parameters is well preserved in the MTRFR based joint inversion technique for both wheat and soybean. © 2019 Elsevier B.V." "35226593200;56993749800;","New feature classes for acoustic habitat mapping—a multibeam echosounder point cloud analysis for mapping submerged aquatic vegetation (SAV)",2019,"10.3390/geosciences9050235","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067623891&doi=10.3390%2fgeosciences9050235&partnerID=40&md5=71235b43753fd6d8f77c1e029b005560","A new method for multibeam echosounder (MBES) data analysis is presented with the aim of improving habitat mapping, especially when considering submerged aquatic vegetation (SAV). MBES data were acquired with 400 kHz in 1–8 m water depth with a spatial resolution in the decimeter scale. The survey area was known to be populated with the seagrass Zostera marina and the bathymetric soundings were highly influenced by this habitat. The depth values often coincide with the canopy of the seagrass. Instead of classifying the data with a digital terrain model and the given derivatives, we derive predictive features from the native point cloud of the MBES soundings in a similar way to terrestrial LiDAR data analysis. We calculated the eigenvalues to derive nine characteristic features, which include linearity, planarity, and sphericity. The features were calculated for each sounding within a cylindrical neighborhood of 0.5 m radius and holding 88 neighboring soundings, on average, during our survey. The occurrence of seagrass was ground-truthed by divers and aerial photography. A data model was constructed and we applied a random forest machine learning supervised classification to predict between the two cases of “seafloor” and “vegetation”. Prediction by linearity, planarity, and sphericity resulted in 88.5% prediction accuracy. After constructing the higher-order eigenvalue derivatives and having the nine features available, the model resulted in 96% prediction accuracy. This study outlines for the first time that valuable feature classes can be derived from MBES point clouds—an approach that could substantially improve bathymetric measurements and habitat mapping. © 2019 by the authors. Licensee MDPI, Basel, Switzerland." "57204807990;55973699300;57204803315;16052116000;","Mapping pasture biomass in Mongolia using Partial Least Squares, Random Forest regression and Landsat 8 imagery",2019,"10.1080/01431161.2018.1541110","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057279946&doi=10.1080%2f01431161.2018.1541110&partnerID=40&md5=5c9cc2865b9ed75ed97a0e6c1cc5d59d","The aim of this study was to develop a robust methodology to estimate pasture biomass across the huge land surface of Mongolia (1.56 × 10 6 km 2 ) using high-resolution Landsat 8 satellite data calibrated against field-measured biomass samples. Two widely used regression models were compared and adopted for this study: Partial Least Squares (PLS) and Random Forest (RF). Both methods were trained to predict pasture biomass using a total of 17 spectral indices derived from Landsat 8 multi-temporal satellite imagery as predictor variables. For training, reference biomass data from a field survey of 553 sites were available. PLS results showed a satisfactory correlation between field measured and estimated biomass with coefficient of determination (R 2 ) = 0.750 and Root Mean Square Error (RMSE) = 101.10 kg ha −1 . The RF regression gave similar results with R 2 = 0.764, RMSE = 98.00 kg ha −1 . An examination of feature importance found the following vegetation indices to be the most relevant: Green Chlorophyll Index (CL green ), Simple Ratio (SR), Wide Dynamic Range Vegetation Index (WDRVI), Enhanced Vegetation Index EVI 1 and Normalized Difference Vegetation Index (NDVI) indices. With respect to the spectral reflectances, Red and Short Wavelength Infra-Red2 (SWIR 2 ) bands showed the strongest correlation with biomass. Using the developed PLS models, a spatial map of pasture biomass covering Mongolia at a spatial resolution of 30 m was generated. Our study confirms the high potential of RF and PLS regression (PLSR) models to predict pasture biomass. The computationally simpler PLSR model is preferred for applications involving large regions. This method can be implemented easily, provided that sufficient reference data and cloud-free observations are available. © 2018, © 2018 Informa UK Limited, trading as Taylor & Francis Group." "57207450471;12786455500;49962003300;10640157100;7202445940;57203526009;35220028600;57201333517;57207461906;","UAV-based biomass estimation for rice-combining spectral, TIN-based structural and meteorological features",2019,"10.3390/RS11070890","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069851052&doi=10.3390%2fRS11070890&partnerID=40&md5=9dab4eae4d00fcc1dd3a04947d4a838a","Accurate estimation of above ground biomass (AGB) is very important for crop growth monitoring. The objective of this study was to estimate rice biomass by utilizing structural and meteorological features with widely used spectral features. Structural features were derived from the triangulated irregular network (TIN), which was directly built from structure from motion (SfM) point clouds. Growing degree days (GDD) was used as the meteorological feature. Three models were used to estimate rice AGB, including the simple linear regression (SLR) model, simple exponential regression (SER) model, and machine learning model (random forest). Compared to models that do not use structural and meteorological features (NDRE, R2 = 0.64, RMSE = 286.79 g/m2, MAE= 236.49 g/m2), models that include such features obtained better estimation accuracy (NDRE*Hcv/GDD, R2 = 0.86, RMSE = 178.37 g/m2, MAE = 127.34 g/m2). This study suggests that the estimation accuracy of rice biomass can benefit from the utilization of structural and meteorological features. © 2018 by the authors." "56527202200;7404087101;57194068533;55950054200;55175154700;7006499360;","A machine learning approach to map tropical selective logging",2019,"10.1016/j.rse.2018.11.044","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057719718&doi=10.1016%2fj.rse.2018.11.044&partnerID=40&md5=a7d9cf58512b21ba0846acd64ee1a9d4","Hundreds of millions of hectares of tropical forest have been selectively logged, either legally or illegally. Methods for detecting and monitoring tropical selective logging using satellite data are at an early stage, with current methods only able to detect more intensive timber harvest (>20 m3 ha−1). The spatial resolution of widely available datasets, like Landsat, have previously been considered too coarse to measure the subtle changes in forests associated with less intensive selective logging, yet most present-day logging is at low intensity. We utilized a detailed selective logging dataset from over 11,000 ha of forest in Rondônia, southern Brazilian Amazon, to develop a Random Forest machine-learning algorithm for detecting low-intensity selective logging (<15 m3 ha−1). We show that Landsat imagery acquired before the cessation of logging activities (i.e. the final cloud-free image of the dry season during logging) was better at detecting selective logging than imagery acquired at the start of the following dry season (i.e. the first cloud-free image of the next dry season). Within our study area the detection rate of logged pixels was approximately 90% (with roughly 20% commission and 8% omission error rates) and approximately 40% of the area inside low-intensity selective logging tracts were labelled as logged. Application of the algorithm to 6152 ha of selectively logged forest at a second site in Pará northeast Brazilian Amazon, resulted in the detection of 2316 ha (38%) of selective logging (with 20% commission and 7% omission error rates). This suggests that our method can detect low-intensity selective logging across large areas of the Amazon. It is thus an important step forward in developing systems for detecting selective logging pan-tropically with freely available data sets, and has key implications for monitoring logging and implementing carbon-based payments for ecosystem service schemes. © 2018 The Authors" "56816836100;56410160500;54924178900;57208225790;57208376176;7402706393;35361180100;6701726073;7102400127;7403044017;","Sorghum biomass prediction using uav-based remote sensing data and crop model simulation",2018,"10.1109/IGARSS.2018.8519034","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064161999&doi=10.1109%2fIGARSS.2018.8519034&partnerID=40&md5=a05d0f4142033efd93e8037ac3d88a9a","Accurate phenotyping with unmanned aerial vehicles is a remote sensing application that has received recent attention as plant breeders seek to automate the expensive and time consuming traditional manual acquisition of measurements of plant traits. This paper focuses on the prediction of sorghum biomass utilizing high temporal and spatial resolution remote sensing data. Two methods are investigated for biomass prediction. The first uses nonlinear regression models to predict biomass directly from remote sensing data, based on features from Light Detection And Ranging (LiDAR) point clouds and hyperspectral data. The second strategy focuses on the biophysical sorghum crop model, APSIM, first, using remote sensing data to parametrize the crop model, and then simulating the biomass. Results from both approaches are provided and evaluated for an agricultural test field at the Agronomy Center for Research and Education (ACRE) at Purdue University. © 2018 IEEE" "56668666900;6507214354;57024178400;57203412821;","Detailed land cover mapping from multitemporal Landsat-8 data of different cloud cover",2018,"10.3390/rs10081214","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051637763&doi=10.3390%2frs10081214&partnerID=40&md5=f4b870ee390d538e8e405d41cb1d73f6","Detailed, accurate and frequent land cover mapping is a prerequisite for several important geospatial applications and the fulfilment of current sustainable development goals. This paper introduces a methodology for the classification of annual high-resolution satellite data into several detailed land cover classes. In particular, a nomenclature with 27 different classes was introduced based on CORINE Land Cover (CLC) Level-3 categories and further analysing various crop types. Without employing cloud masks and/or interpolation procedures, we formed experimental datasets of Landsat-8 (L8) images with gradually increased cloud cover in order to assess the influence of cloud presence on the reference data and the resulting classification accuracy. The performance of shallow kernel-based and deep patch-based machine learning classification frameworks was evaluated. Quantitatively, the resulting overall accuracy rates differed within a range of less than 3%; however, maps produced based on Support Vector Machines (SVM) were more accurate across class boundaries and the respective framework was less computationally expensive compared to the applied patch-based deep Convolutional Neural Network (CNN). Further experimental results and analysis indicated that employing all multitemporal images with up to 30% cloud cover delivered relatively higher overall accuracy rates as well as the highest per-class accuracy rates. Moreover, by selecting 70% of the top-ranked features after applying a feature selection strategy, slightly higher accuracy rates were achieved. A detailed discussion of the quantitative and qualitative evaluation outcomes further elaborates on the performance of all considered classes and highlights different aspects of their spectral behaviour and separability. © 2018 by the authors." "56784712800;23091749600;","Assessing the Performance of a Machine Learning Algorithm in Identifying Bubbles in Dust Emission",2017,"10.3847/1538-4357/aa9a42","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85039700362&doi=10.3847%2f1538-4357%2faa9a42&partnerID=40&md5=f32d60525b2d64ae118a8e734fcf4b5b","Stellar feedback created by radiation and winds from massive stars plays a significant role in both physical and chemical evolution of molecular clouds. This energy and momentum leaves an identifiable signature (""bubbles"") that affects the dynamics and structure of the cloud. Most bubble searches are performed ""by eye,"" which is usually time-consuming, subjective, and difficult to calibrate. Automatic classifications based on machine learning make it possible to perform systematic, quantifiable, and repeatable searches for bubbles. We employ a previously developed machine learning algorithm, Brut, and quantitatively evaluate its performance in identifying bubbles using synthetic dust observations. We adopt magnetohydrodynamics simulations, which model stellar winds launching within turbulent molecular clouds, as an input to generate synthetic images. We use a publicly available three-dimensional dust continuum Monte Carlo radiative transfer code, hyperion, to generate synthetic images of bubbles in three Spitzer bands (4.5, 8, and 24 μm). We designate half of our synthetic bubbles as a training set, which we use to train Brut along with citizen-science data from the Milky Way Project (MWP). We then assess Brut's accuracy using the remaining synthetic observations. We find that Brut's performance after retraining increases significantly, and it is able to identify yellow bubbles, which are likely associated with B-type stars. Brut continues to perform well on previously identified high-score bubbles, and over 10% of the MWP bubbles are reclassified as high-confidence bubbles, which were previously marginal or ambiguous detections in the MWP data. We also investigate the influence of the size of the training set, dust model, evolutionary stage, and background noise on bubble identification. © 2017. The American Astronomical Society. All rights reserved." "35293201200;","Towards a digital ecosystem for predictive healthcare analytics",2014,"10.1145/2668260.2668286","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84910020381&doi=10.1145%2f2668260.2668286&partnerID=40&md5=1210bd30a71227bf31713d95a90ecde3","Healthcare applications depend on voluminous data containing rich and important insights that can support predictive analysis by discovering these patterns using machine learning algorithms. Unfortunately this data is not always available for multiple reasons including lack of unified and standardized architecture for distribution / dissipation of such data. In this paper, we propose a framework for cloud based architecture of a digital ecosystem for storing, sharing and predictive analysis of healthcare data. We address the requirement for such a system and provide the design and architecture of the framework comprising of various interconnecting components. The proposed framework provides a cost effective digital ecosystem for collaboration, sharing and integration of Electronic Health Record systems by leveraging the benefits of cloud computing technologies. The framework provides a platform for various health care institutes, research organizations and government agencies to log data, develop predictive analytics models and collaborate in future research. A proof-of concept of the framework is implemented. Copyright © 2014 ACM." "23398420700;56118777300;","An expert fitness diagnosis system based on elastic cloud computing",2014,"10.1155/2014/981207","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84898792083&doi=10.1155%2f2014%2f981207&partnerID=40&md5=889ff5c61660972a909c3a0f17db0075","This paper presents an expert diagnosis system based on cloud computing. It classifies a user's fitness level based on supervised machine learning techniques. This system is able to learn and make customized diagnoses according to the user's physiological data, such as age, gender, and body mass index (BMI). In addition, an elastic algorithm based on Poisson distribution is presented to allocate computation resources dynamically. It predicts the required resources in the future according to the exponential moving average of past observations. The experimental results show that Naïve Bayes is the best classifier with the highest accuracy (90.8%) and that the elastic algorithm is able to capture tightly the trend of requests generated from the Internet and thus assign corresponding computation resources to ensure the quality of service. © 2014 Kevin C. Tseng and Chia-Chuan Wu." "57202622081;55839248600;36814094500;","A machine learning approach to Cepheid variable star classification using data alignment and maximum likelihood",2013,"10.1016/j.ascom.2013.07.002","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84883138747&doi=10.1016%2fj.ascom.2013.07.002&partnerID=40&md5=bdddc5956a7ab7e969bf9ece70031cde","Our study centers on the classification of two subtypes of Cepheid variable stars. Such a classification is relatively easy to obtain for nearby galaxies, but as we incorporate new galaxies, the cost of labeling stars calls for some form of model adaptation. Adapting a predictive model to differentiate Cepheids across galaxies is difficult because of the sample bias problem in star distribution (due to the limitation of telescopes in observing faint stars as we try to reach distant galaxies). In addition, estimating the luminosity of a star as we reach distant galaxies carries some inevitable shift in the data distribution. We propose an approach to predict the class of Cepheid stars on a target domain, by first building a model on an ""anchor"" source domain. Our methodology then shifts the target data until it is well aligned with the source data by maximizing two different likelihood functions. Experimental results with two galaxy datasets (Large Magellanic Cloud as the source domain, and M33 as the target domain), show the efficacy of the proposed method. © 2013 Elsevier B.V." "15020631200;57203279842;55135808700;56539588700;57192838041;7004251030;57211924569;57211925753;6602451418;","Agricultural cropland extent and areas of South Asia derived using Landsat satellite 30-m time-series big-data using random forest machine learning algorithms on the Google Earth Engine cloud",2020,"10.1080/15481603.2019.1690780","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075351890&doi=10.1080%2f15481603.2019.1690780&partnerID=40&md5=5a6f0da4198b137ce691961811a6612a","The South Asia (India, Pakistan, Bangladesh, Nepal, Sri Lanka and Bhutan) has a staggering 900 million people (~43% of the population) who face food insecurity or severe food insecurity as per United Nations, Food and Agriculture Organization’s (FAO) the Food Insecurity Experience Scale (FIES). The existing coarse-resolution (≥250-m) cropland maps lack precision in geo-location of individual farms and have low map accuracies. This also results in uncertainties in cropland areas calculated from such products. Thereby, the overarching goal of this study was to develop a high spatial resolution (30-m or better) baseline cropland extent product of South Asia for the year 2015 using Landsat satellite time-series big-data and machine learning algorithms (MLAs) on the Google Earth Engine (GEE) cloud computing platform. To eliminate the impact of clouds, 10 time-composited Landsat bands (blue, green, red, NIR, SWIR1, SWIR2, Thermal, EVI, NDVI, NDWI) were derived for each of the three time-periods over 12 months (monsoon: Days of the Year (DOY) 151–300; winter: DOY 301–365 plus 1–60; and summer: DOY 61–150), taking the every 8-day data from Landsat-8 and 7 for the years 2013–2015, for a total of 30-bands plus global digital elevation model (GDEM) derived slope band. This 31-band mega-file big data-cube was composed for each of the five agro-ecological zones (AEZ’s) of South Asia and formed a baseline data for image classification and analysis. Knowledge-base for the Random Forest (RF) MLAs were developed using spatially well spread-out reference training data (N = 2179) in five AEZs. The classification was performed on GEE for each of the five AEZs using well-established knowledge-base and RF MLAs on the cloud. Map accuracies were measured using independent validation data (N = 1185). The survey showed that the South Asia cropland product had a producer’s accuracy of 89.9% (errors of omissions of 10.1%), user’s accuracy of 95.3% (errors of commission of 4.7%) and an overall accuracy of 88.7%. The National and sub-national (districts) areas computed from this cropland extent product explained 80-96% variability when compared with the National statistics of the South Asian Countries. The full-resolution imagery can be viewed at full-resolution, by zooming-in to any location in South Asia or the world, at www.croplands.org and the cropland products of South Asia downloaded from The Land Processes Distributed Active Archive Center (LP DAAC) of National Aeronautics and Space Administration (NASA) and the United States Geological Survey (USGS): https://lpdaac.usgs.gov/products/gfsad30saafgircev001/. © 2019, © 2019 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group." "55548541000;6507214354;8053177000;57201541167;","Correcting image refraction: Towards accurate aerial image-based bathymetry mapping in shallow waters",2020,"10.3390/rs12020322","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081083628&doi=10.3390%2frs12020322&partnerID=40&md5=1a90b1446ae79a399e28714ec9e913ba","Although aerial image-based bathymetric mapping can provide, unlike acoustic or LiDAR (Light Detection and Ranging) sensors, both water depth and visual information, water refraction poses significant challenges for accurate depth estimation. In order to tackle this challenge, we propose an image correction methodology, which first exploits recent machine learning procedures that recover depth from image-based dense point clouds and then corrects refraction on the original imaging dataset. This way, the structure from motion (SfM) and multi-view stereo (MVS) processing pipelines are executed on a refraction-free set of aerial datasets, resulting in highly accurate bathymetric maps. Performed experiments and validation were based on datasets acquired during optimal sea state conditions and derived from four different test-sites characterized by excellent sea bottom visibility and textured seabed. Results demonstrated the high potential of our approach, both in terms of bathymetric accuracy, as well as texture and orthoimage quality. © 2020 by the authors." "56814894700;7006418829;57203552230;8718417000;57219203953;","Crop NDVI monitoring based on sentinel 1",2019,"10.3390/rs11121441","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068108763&doi=10.3390%2frs11121441&partnerID=40&md5=abf6d59b0381937d455f438fb4dd0b94","Monitoring agricultural crops is necessary for decision-making in the field. However, it is known that in some regions and periods, cloud cover makes this activity difficult to carry out in a systematic way throughout the phenological cycle of crops. This circumstance opens up opportunities for techniques involving radar sensors, resulting in images that are free of cloud effects. In this context, the objective of this work was to obtain a normalized different vegetation index (NDVI) cloudless product (NDVInc) by modeling Sentinel 2 NDVI using different regression techniques and the Sentinel 1 radar backscatter as input. To do this, we used four pairs of Sentinel 2 and Sentinel 1 images on coincident days, aiming to achieve the greatest range of NDVI values for agricultural crops (soybean and maize). These coincident pairs were the only ones in which the percentage of clouds was not equal to 100% for 33 central pivot areas in western Bahia, Brazil. The dataset used for NDVInc modeling was divided into two subsets: training and validation. The training and validation datasets were from the period from 24 June 2017 to 19 July 2018 (four pairs of images). The best performing model was used in a temporal analysis from 02 October 2017 to 08 August 2018, totaling 55 Sentinel 2 images and 25 Sentinel 1 images. The selection of the best regression algorithm was based on two validation methodologies: K-fold cross-validation (k = 10) and holdout. We tested four modeling approaches with eight regression algorithms. The random forest was the algorithm that presented the best statistical metrics, regardless of the validation methodology and the approach used. Therefore, this model was applied to a time series of Sentinel 1 images in order to demonstrate the robustness and applicability of the model created. We observed that the data derived from Sentinel 1 allowed us to model, with great reliability, the NDVI of agricultural crops throughout the phenological cycle, making the methodology developed in this work a relevant solution for the monitoring of various regions, regardless of cloud cover. © 2019 by the authors. All right reserved." "57204567584;7801421591;6603347989;","Delineation of cocoa agroforests using multiseason sentinel-1 SAR images: A low grey level range reduces uncertainties in GLCM texture-based mapping",2019,"10.3390/ijgi8040179","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066321784&doi=10.3390%2fijgi8040179&partnerID=40&md5=c04c9451d80096511ec25d2252694d99","Delineating the cropping area of cocoa agroforests is a major challenge in quantifying the contribution of land use expansion to tropical deforestation. Discriminating cocoa agroforests from tropical transition forests using multispectral optical images is difficult due to the similarity of the spectral characteristics of their canopies. Moreover, the frequent cloud cover in the tropics greatly impedes optical sensors. This study evaluated the potential of multiseason Sentinel-1 C-band synthetic aperture radar (SAR) imagery to discriminate cocoa agroforests from transition forests in a heterogeneous landscape in central Cameroon. We used an ensemble classifier, Random Forest (RF), to average the SAR image texture features of a grey level co-occurrence matrix (GLCM) across seasons. We then compared the classification performance with results from RapidEye optical data. Moreover, we assessed the performance of GLCM texture feature extraction at four different grey levels of quantization: 32 bits, 8 bits, 6 bits, and 4 bits. The classification’s overall accuracy (OA) from texture-based maps outperformed that from an optical image. The highest OA (88.8%) was recorded at the 6 bits grey level. This quantization level, in comparison to the initial 32 bits in the SAR images, reduced the class prediction error by 2.9%. The texture-based classification achieved an acceptable accuracy and revealed that cocoa agroforests have considerably fragmented the remnant transition forest patches. The Shannon entropy (H) or uncertainty provided a reliable validation of the class predictions and enabled inferences about discriminating inherently heterogeneous vegetation categories. © 2019 by the authors." "57191576880;55887116200;16022263500;36350847100;56942309200;15838180100;","Fully scalable forward model grid of exoplanet transmission spectra",2019,"10.1093/mnras/sty3001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85061101491&doi=10.1093%2fmnras%2fsty3001&partnerID=40&md5=1143be9b3350a4c4a9c82d1289b364ab","Simulated exoplanet transmission spectra are critical for planning and interpretation of observations and to explore the sensitivity of spectral features to atmospheric thermochemical processes. We present a publicly available generic model grid of planetary transmission spectra, scalable to a wide range of H2/He dominated atmospheres. The grid is computed using the 1D/2D atmosphere model ATMO for two different chemical scenarios, first considering local condensation only, secondly considering global condensation and removal of species from the atmospheric column (rainout). The entire grid consists of 56 320 model simulations across 22 equilibrium temperatures (400–2600 K), four planetary gravities (5–50 ms−2), five atmospheric metallicities (1x–200x), four C/O ratios (0.35–1.0), four scattering haze parameters, four uniform cloud parameters, and two chemical scenarios. We derive scaling equations which can be used with this grid, for a wide range of planet–star combinations. We validate this grid by comparing it with other model transmission spectra available in the literature. We highlight some of the important findings, such as the rise of SO2 features at 100x solar metallicity, differences in spectral features at high C/O ratios between two condensation approaches, the importance of VO features without TiO to constrain the limb temperature and features of TiO/VO both, to constrain the condensation processes. Finally, this generic grid can be used to plan future observations using the HST, VLT, JWST, and various other telescopes. The fine variation of parameters in the grid also allows it to be incorporated in a retrieval framework, with various machine learning techniques. © 2018 The Author(s)." "57201333629;7201562710;7102530986;","An automated algorithm for mapping building impervious areas from airborne LiDAR point-cloud data for flood hydrology",2018,"10.1080/15481603.2018.1452588","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85044347773&doi=10.1080%2f15481603.2018.1452588&partnerID=40&md5=e46dbfca033a075a54bc9fa79c7f5d5d","Buildings, as impervious surfaces, are an important component of total impervious surface areas that drive urban stormwater response to intense rainfall events. Most stormwater models that use percent impervious area (PIA) are spatially lumped models and do not require precise locations of building roofs, as in other applications of building maps, but do require accurate estimates of total impervious areas within the geographic units of observation (e.g. city blocks or sub-watershed units). Two-dimensional mapping of buildings from aerial imagery requires laborious efforts from image analysts or elaborate image analysis techniques using high spatial resolution imagery. Moreover, large uncertainties exist where tall, dense vegetation obscures the structures. Analyzing LiDAR point-cloud data, however, can distinguish buildings from vegetation canopy and facilitate the mapping of buildings. This paper presents a new building extraction approach that is based on and optimized for estimating building impervious areas (BIA) for hydrologic purposes and can be used with standard GIS software to identify building roofs under tall, thick canopy. Accuracy assessment methods are presented that can optimize model performance for modeling BIA within the geographic units of observation for hydrologic applications. The Building Extraction from LiDAR Last Returns (BELLR) model, a 2.5D rule-based GIS model, uses a non-spatial, local vertical difference filter (VDF) on LiDAR point-cloud data to automatically identify and map building footprints. The model includes an absolute difference in elevation (AdE) parameter in the VDF that compares the difference between mean and modal elevations of last-returns in each cell. The BELLR model is calibrated for an extensive inner-city, highly urbanized small watershed in Columbia, South Carolina, USA that is covered by tall, thick vegetation canopy that obscures many buildings. The calibration of BELLR used a set of building locations compiled by photo-analysts, and validation used independent building reference data. The model is applied to two residential neighborhoods, one of which is a residential area within the primary watershed and the other is a younger suburban neighborhood with a less-well developed tree canopy used as a validation site. Performance results indicate that the BELLR model is highly sensitive to concavity in the lasboundary tool of LAStools® and those settings are highly site specific. The model is also sensitive to cell size and the AdE threshold values. However, properly calibrated the BIA for the two residential sites could be estimated within 1% error for optimized experiments. To examine results in a hydrologic application, the BELLR estimated BIAs were tested using two different types of hydrologic models to compare BELLR results with results using the National Land Cover Database (NLCD) 2011 Percent Developed Imperviousness data. The BELLR BIA values provide more accurate results than the use of the 2011 NLCD PIA data in both models. The VDF developed in this study to map buildings could be applied to LiDAR point-cloud filtering algorithms for feature extraction in machine learning or mapping other planar surfaces in more broad-based land-cover classifications. © 2018, © 2018 Informa UK Limited, trading as Taylor & Francis Group." "57194573861;7006508549;7101619588;12753248500;","Detection of transverse cirrus bands in satellite imagery using deep learning",2018,"10.1016/j.cageo.2018.05.012","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048465491&doi=10.1016%2fj.cageo.2018.05.012&partnerID=40&md5=19b0828ce2471dd4ba8b4d59655edad0","We demonstrate the viability of using a convolutional neural network (CNN) for facial recognition of meteorological phenomena in satellite imagery. Transfer learning was used to fine tune the widely used VGG-16 network architecture and allow the network to successfully detect (94% accuracy) the presence of transverse cirrus bands (TCBs) in NASA MODIS and VIIRS satellite browse imagery. The CNN exhibited better performance compared to a random forest classifier (84% accuracy) and was further validated by applying it to NASA satellite browse imagery in order to create a multi-year (2013–2015) global heat map of TCB occurrence. The annual heat map shows spatial patterns that are consistent with known mechanisms for the generation of TCBs, providing confidence in the CNN classifications. Our study shows that CNNs are well suited for meteorological phenomena detection due to their generalization capabilities and strong performance. An immediate application of our work intends to enable phenomena-based search of big satellite imagery databases. With additional modifications, the CNN could be utilized for other applications such as providing situational awareness to operational forecasters or developing phenomena specific climatologies. © 2018 Elsevier Ltd" "57199999083;24460392200;56100447800;15835468800;57190384098;6603043158;","A neural network approach to estimating a posteriori distributions of Bayesian retrieval problems",2018,"10.5194/amt-11-4627-2018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051359687&doi=10.5194%2famt-11-4627-2018&partnerID=40&md5=ebf06c07909a78fbb27f98113f07ac90","A neural-network-based method, quantile regression neural networks (QRNNs), is proposed as a novel approach to estimating the a posteriori distribution of Bayesian remote sensing retrievals. The advantage of QRNNs over conventional neural network retrievals is that they learn to predict not only a single retrieval value but also the associated, case-specific uncertainties. In this study, the retrieval performance of QRNNs is characterized and compared to that of other state-of-the-art retrieval methods. A synthetic retrieval scenario is presented and used as a validation case for the application of QRNNs to Bayesian retrieval problems. The QRNN retrieval performance is evaluated against Markov chain Monte Carlo simulation and another Bayesian method based on Monte Carlo integration over a retrieval database. The scenario is also used to investigate how different hyperparameter configurations and training set sizes affect the retrieval performance. In the second part of the study, QRNNs are applied to the retrieval of cloud top pressure from observations by the Moderate Resolution Imaging Spectroradiometer (MODIS). It is shown that QRNNs are not only capable of achieving similar accuracy to standard neural network retrievals but also provide statistically consistent uncertainty estimates for non-Gaussian retrieval errors. The results presented in this work show that QRNNs are able to combine the flexibility and computational efficiency of the machine learning approach with the theoretically sound handling of uncertainties of the Bayesian framework. Together with this article, a Python implementation of QRNNs is released through a public repository to make the method available to the scientific community. © 2018 Author(s)." "54396725800;57200648324;8924809400;56087837500;56274863900;35956222500;57193613007;52364041600;6505626516;8727294800;7004510450;15832380200;16070455300;55181653000;35321650700;","Clustering the Orion B giant molecular cloud based on its molecular emission",2018,"10.1051/0004-6361/201731833","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85042095524&doi=10.1051%2f0004-6361%2f201731833&partnerID=40&md5=93b946156ecaa1de77c3d43f1b21e2bd","Context. Previous attempts at segmenting molecular line maps of molecular clouds have focused on using position-position-velocity data cubes of a single molecular line to separate the spatial components of the cloud. In contrast, wide field spectral imaging over a large spectral bandwidth in the (sub)mm domain now allows one to combine multiple molecular tracers to understand the different physical and chemical phases that constitute giant molecular clouds (GMCs). Aims. We aim at using multiple tracers (sensitive to different physical processes and conditions) to segment a molecular cloud into physically/chemically similar regions (rather than spatially connected components), thus disentangling the different physical/chemical phases present in the cloud. Methods. We use a machine learning clustering method, namely the Meanshift algorithm, to cluster pixels with similar molecular emission, ignoring spatial information. Clusters are defined around each maximum of the multidimensional probability density function (PDF) of the line integrated intensities. Simple radiative transfer models were used to interpret the astrophysical information uncovered by the clustering analysis. Results. A clustering analysis based only on the J = 1-0 lines of three isotopologues of CO proves sufficient to reveal distinct density/column density regimes (nH ∼ 100 cm-3, ∼500 cm-3, and >1000 cm-3), closely related to the usual definitions of diffuse, translucent and high-column-density regions. Adding two UV-sensitive tracers, the J = 1-0 line of HCO+ and the N = 1-0 line of CN, allows us to distinguish two clearly distinct chemical regimes, characteristic of UV-illuminated and UV-shielded gas. The UV-illuminated regime shows overbright HCO+ and CN emission, which we relate to a photochemical enrichment effect. We also find a tail of high CN/HCO+ intensity ratio in UV-illuminated regions. Finer distinctions in density classes (nH ∼ 7 × 103 cm-3, ∼4 × 104 cm-3) for the densest regions are also identified, likely related to the higher critical density of the CN and HCO+ (1-0) lines. These distinctions are only possible because the high-density regions are spatially resolved. Conclusions. Molecules are versatile tracers of GMCs because their line intensities bear the signature of the physics and chemistry at play in the gas. The association of simultaneous multi-line, wide-field mapping and powerful machine learning methods such as the Meanshift clustering algorithm reveals how to decode the complex information available in these molecular tracers. © ESO, 2018." "56458013900;16317584800;","An interactive visual analytic tool for semantic classification of 3D urban LiDAR point cloud",2015,"10.1145/2820783.2820863","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84961214773&doi=10.1145%2f2820783.2820863&partnerID=40&md5=53f59986a38be20b6e36f5fd74f52492","We propose a novel unsupervised machine learning approach for effective semantic labeling by combining two different multi-class classifications, structural and contextual classification, of points in airborne LiDAR point cloud of urban environment. Structural classification labels a point in the cloud as a point-, line-, or surface-type feature. An additional outcome of this classification is the geometrypreserving downsampling of the point cloud. The contextual classification, on the other hand, labels the points in four classes, namely, buildings, vegetation, natural ground, and asphalt ground, by using data derived from the raw input, which includes the structural classification. Preserving these two classifications in the labeling of the points gives a geometry-aware contextual semantic labeling. We propose: (a) an augmented semantic classification which preserves both structural and contextual classification, (b) an interactive hierarchical clustering method for contextual classification, and (c) an interactive visual analytic framework to aid both the structural and contextual classifications. © 2015 ACM." "56022096000;","Dynamic protection for critical health care systems using cisco CWS: Unleashing the power of big data analytics",2014,"10.1109/COM.Geo.2014.28","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84908587537&doi=10.1109%2fCOM.Geo.2014.28&partnerID=40&md5=0e0bdbb412d39edae1393e5f32e3cdd9","Critical Care IT systems such as life support devices, vitals monitoring systems, information systems that provide point of care guidance to care teams are a key component of a lifesaving effort in Healthcare. The mega trends of mobility, social, cloud combined with wide spread increase and sophistication of malware, has created new challenges and the point in time detection methods at the hospitals are no longer effective and pose a big threat to the critical care systems. To maintain the availability and integrity of these critical care systems, new adaptive, learning security defense systems are required that not only learns from the traffic entering the hospital, but also proactively learns from the traffic worldwide. Cisco's Cloud web security (CWS) provides industry-leading security and control for the distributed enterprise by protecting users everywhere, anytime through Cisco worldwide threat intelligence, advanced threat defense capabilities, and roaming user protection. It leverages the big data to perform behavioral analysis, anomaly detection, evasion resistance, rapid Detection services using flow based, signature based, behavior based and full packet capture models to identify threats. This tech talk looks at how big Data Analytics is used in combination with other security capabilities to proactively identify threats and prevent wide spread damage to healthcare critical assets. © 2014 IEEE." "35316037700;57209845617;24777858900;55698016900;","Evaluating combinations of sentinel-2 data and machine-learning algorithms for mangrove mapping in West Africa",2019,"10.3390/rs11242928","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077862543&doi=10.3390%2frs11242928&partnerID=40&md5=d416d3c8cfe1411510e8a6f4ac1189ed","Creating a national baseline for natural resources, such as mangrove forests, and monitoring them regularly often requires a consistent and robust methodology. With freely available satellite data archives and cloud computing resources, it is now more accessible to conduct such large-scale monitoring and assessment. Yet, few studies examine the reproducibility of such mangrove monitoring frameworks, especially in terms of generating consistent spatial extent. Our objective was to evaluate a combination of image processing approaches to classify mangrove forests along the coast of Senegal and The Gambia. We used freely available global satellite data (Sentinel-2), and cloud computing platform (Google Earth Engine) to run two machine learning algorithms, random forest (RF), and classification and regression trees (CART). We calibrated and validated the algorithms using 800 reference points collected using high-resolution images. We further re-ran 10 iterations for each algorithm, utilizing unique subsets of the initial training data. While all iterations resulted in thematic mangrove maps with over 90% accuracy, the mangrove extent ranges between 827-2807 km2 for Senegal and 245-1271 km2 for The Gambia with one outlier for each country. We further report ""Places of Agreement"" (PoA) to identify areas where all iterations for both methods agree (506.6 km2 and 129.6 km2 for Senegal and The Gambia, respectively), thus have a high confidence in predicting mangrove extent. While we acknowledge the time- and cost-effectiveness of such methods for the landscape managers, we recommend utilizing them with utmost caution, as well as post-classification on-the-ground checks, especially for decision making. © 2019 by the authors." "57195135675;6603146282;7004196511;8445905500;","Addressing overfitting on point cloud classification using Atrous XCRF",2019,"10.1016/j.isprsjprs.2019.07.002","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068556361&doi=10.1016%2fj.isprsjprs.2019.07.002&partnerID=40&md5=06762fba0e2ee5207b871cc8d7f41831","Advances in techniques for automated classification of point cloud data introduce great opportunities for many new and existing applications. However, with a limited number of labelled points, automated classification by a machine learning model is prone to overfitting and poor generalization. The present paper addresses this problem by inducing controlled noise (on a trained model) generated by invoking conditional random field similarity penalties using nearby features. The method is called Atrous XCRF and works by forcing a trained model to respect the similarity penalties provided by unlabeled data. In a benchmark study carried out using the ISPRS 3D labeling dataset, our technique achieves 85.0% in term of overall accuracy, and 71.1% in term of F1 score. The result is on par with the current best model for the benchmark dataset and has the highest value in term of F1 score. Additionally, transfer learning using the Bergen 2018 dataset, without model retraining, was also performed. Even though our proposal provides a consistent 3% improvement in term of accuracy, more work still needs to be done to alleviate the generalization problem on the domain adaptation and the transfer learning field. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57189340738;55947165200;57209231097;57209233506;7003424893;","Estimating the daily pollen concentration in the atmosphere using machine learning and NEXRAD weather radar data",2019,"10.1007/s10661-019-7542-9","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066939934&doi=10.1007%2fs10661-019-7542-9&partnerID=40&md5=a3f164cbd862aed741914b642c18ac1f","Millions of people have an allergic reaction to pollen. The impact of pollen allergies is on the rise due to increased pollen levels caused by global warming and the spread of highly invasive weeds. The production, release, and dispersal of pollen depend on the ambient weather conditions. The temperature, rainfall, humidity, cloud cover, and wind are known to affect the amount of pollen in the atmosphere. In the past, various regression techniques have been applied to estimate and forecast the daily pollen concentration in the atmosphere based on the weather conditions. In this research, machine learning methods were applied to the Next Generation Weather Radar (NEXRAD) data to estimate the daily Ambrosia pollen over a 300 km × 300 km region centered on a NEXRAD weather radar. The Neural Network and Random Forest machine learning methods have been employed to develop separate models to estimate Ambrosia pollen over the region. A feasible way of estimating the daily pollen concentration using only the NEXRAD radar data and machine learning methods would lay the foundation to forecast daily pollen at a fine spatial resolution nationally. © 2019, Springer Nature Switzerland AG." "57202643803;6507811592;57203178537;","Machine learning approach to classify rain type based on thies disdrometers and cloud observations",2019,"10.3390/atmos10050251","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074069139&doi=10.3390%2fatmos10050251&partnerID=40&md5=ce95e1eefc2e2a67a2dac83153ab90d0","Rain microstructure parameters assessed by disdrometers are commonly used to classify rain into convective and stratiform. However, different types of disdrometer result in different values for these parameters. This in turn potentially deteriorates the quality of rain type classifications. Thies disdrometer measurements at two sites in Bavaria in southern Germany were combined with cloud observations to construct a set of clear convective and stratiform intervals. This reference dataset was used to study the performance of classification methods from the literature based on the rain microstructure. We also explored the possibility of improving the performance of these methods by tuning the decision boundary. We further identified highly discriminant rain microstructure parameters and used these parameters in five machine-learning classification models. Our results confirm the potential of achieving high classification performance by applying the concepts of machine learning compared to already available methods. Machine-learning classification methods provide a concrete and flexible procedure that is applicable regardless of the geographical location or the device. The suggested procedure for classifying rain types is recommended prior to studying rain microstructure variability or any attempts at improving radar estimations of rain intensity. © 2019 by the authors." "55548541000;57201541167;8053177000;6507214354;","SHALLOW WATER BATHYMETRY MAPPING from UAV IMAGERY BASED on MACHINE LEARNING",2019,"10.5194/isprs-archives-XLII-2-W10-9-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065647354&doi=10.5194%2fisprs-archives-XLII-2-W10-9-2019&partnerID=40&md5=d3c5ad9b14e262036005965e3aec2df8","The determination of accurate bathymetric information is a key element for near offshore activities, hydrological studies such as coastal engineering applications, sedimentary processes, hydrographic surveying as well as archaeological mapping and biological research. UAV imagery processed with Structure from Motion (SfM) and Multi View Stereo (MVS) techniques can provide a low-cost alternative to established shallow seabed mapping techniques offering as well the important visual information. Nevertheless, water refraction poses significant challenges on depth determination. Till now, this problem has been addressed through customized image-based refraction correction algorithms or by modifying the collinearity equation. In this paper, in order to overcome the water refraction errors, we employ machine learning tools that are able to learn the systematic underestimation of the estimated depths. In the proposed approach, based on known depth observations from bathymetric LiDAR surveys, an SVR model was developed able to estimate more accurately the real depths of point clouds derived from SfM-MVS procedures. Experimental results over two test sites along with the performed quantitative validation indicated the high potential of the developed approach. © 2019 Copernicus GmbH. All righhts reserved." "57201655246;48260990900;55608476300;15034793900;56622117600;","A Comparative Land-Cover Classification Feature Study of Learning Algorithms: DBM, PCA, and RF Using Multispectral LiDAR Data",2019,"10.1109/JSTARS.2019.2899033","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064718834&doi=10.1109%2fJSTARS.2019.2899033&partnerID=40&md5=29d77713d1740f5c5d79c8bcbb8b0d42","Multispectral LiDAR, characterization of completeness, and consistency of spectrum and spatial geometric data provide a new data source for land cover classification. However, how to choose the optimal features for a given set of land covers is an open problem for effective land cover classification. To address this problem, we propose a comparative scheme, which investigates a popular deep learning (deep Boltzmann machine, DBM) model for high-level feature representation and widely used machine learning methods for low-level feature extraction and selection [principal component analysis (PCA) and random forest (RF)] in land cover classification. The comparative study was conducted on the multispectral LiDAR point clouds, acquired by a Teledyne Optech's Titan airborne system. The deep learning-based high-level feature representation experimental results showed that, on an ordinary personal computer or workstation, this method required larger training samples and more computational complexity than the machine learning-based low-level feature extraction and selection methods. However, our comparative experiments demonstrated that the classification accuracies of the DBM-based method were higher than those of the RF-based and PCA-based methods using multispectral LiDAR data. © 2008-2012 IEEE." "57200602927;15034793900;56579557400;57205883470;57205888652;57203541311;","Learning high-level features by fusing multi-view representation of MLS point clouds for 3D object recognition in road environments",2019,"10.1016/j.isprsjprs.2019.01.024","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85061529859&doi=10.1016%2fj.isprsjprs.2019.01.024&partnerID=40&md5=69f88d058bb213bd01e7be0ce9b2f329","Most existing 3D object recognition methods still suffer from low descriptiveness and weak robustness although remarkable progress has made in 3D computer vision. The major challenge lies in effectively mining high-level 3D shape features. This paper presents a high-level feature learning framework for 3D object recognition through fusing multiple 2D representations of point clouds. The framework has two key components: (1) three discriminative low-level 3D shape descriptors for obtaining multi-view 2D representation of 3D point clouds. These descriptors preserve both local and global spatial relationships of points from different perspectives and build a bridge between 3D point clouds and 2D Convolutional Neural Networks (CNN). (2) A two-stage fusion network, which consists of a deep feature learning module and two fusion modules, for extracting and fusing high-level features. The proposed method was tested on three datasets, one of which is Sydney Urban Objects dataset and the other two were acquired by a mobile laser scanning (MLS) system along urban roads. The results obtained from comprehensive experiments demonstrated that our method is superior to the state-of-the-art methods in descriptiveness, robustness and efficiency. Our method achieves high recognition rates of 94.6%, 93.1% and 74.9% on the above three datasets, respectively. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "7006970928;7004150925;55454728300;55568526801;7003466102;","Alternating decision trees for cloud masking in MODIS and VIIRS NASA sea surface temperature products",2019,"10.1175/JTECH-D-18-0103.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063689115&doi=10.1175%2fJTECH-D-18-0103.1&partnerID=40&md5=02943fed4a139a40bd665c1ba3a776d0","Identification and exclusion of clouds from satellite-based infrared fields is critical to achieve accurate retrievals of sea surface temperature (SST). Historically, identification of clouds has been driven primarily by a few uniformity tests involving a small number of pixels, brightness temperature range tests, and comparisons to low-resolution gap-free reference fields. Collectively these tests are adequate at identifying large, upper-level, very cold cumulus clouds, and uniformity tests identify moderately sized patchy cumulus clouds. But the efficacy of cloud identification often decreases at cloud edges, for small or thin cirrus clouds, and for the lower, more uniform stratus clouds, for which cloud-top temperature can be comparable to that of the sea surface, particularly at high latitudes. The heavy reliance on stringent uniformity thresholds often also has the unintended consequence of eliminating strong SST frontal regions from the pool of best-quality retrievals. This paper presents results for an ensemble cloud classifier based on a machine-learning approach, boosted alternating decision trees (ADtrees), applied to NASA MODIS and VIIRS SST imagery. The ADtree algorithm relies on the use of a majority vote from a collection of both ""weak"" and ""strong"" classifiers. This approach offers the potential to identify more cloud types and improve the retention of SST gradients in best-quality SST retrievals and also provides a per pixel confidence estimate in the classification. © 2019 American Meteorological Society." "57195469687;54400787600;6507482782;57210948650;6602228395;","Studying the impact on urban health over the greater delta region in Egypt due to aerosol variability using optical characteristics from satellite observations and ground-based AERONET measurements",2019,"10.3390/rs11171998","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071984228&doi=10.3390%2frs11171998&partnerID=40&md5=4d25f8fa9439bd25a17a23dac884192c","This research addresses the aerosol characteristics and variability over Cairo and the Greater Delta region over the last 20 years using an integrative multi-sensor approach of remotely sensed and PM10 ground data. The accuracy of these satellite aerosol products is also evaluated and compared through cross-validation against ground observations from the AErosol RObotic NETwork (AERONET) project measured at local stations. The results show the validity of using Multi-angle Imaging Spectroradiometer (MISR) and Moderate Resolution Imaging Spectroradiometer (MODIS) sensors on the Terra and Aqua platforms for quantitative aerosol optical depth (AOD) assessment as compared to Ozone Monitoring Instrument (OMI), Sea-viewingWide Field-of-view Sensor (SeaWiFS), and POLarization and Directionality of the Earth's Reflectances (POLDER). In addition, extracted MISR-based aerosol products have been proven to be quite effective in investigating the characteristics of mixed aerosols. Daily AERONET AOD observations were collected and classified using K-means unsupervised machine learning algorithms, showing five typical patterns of aerosols in the region under investigation. Four seasonal aerosol emerging episodes are identified and analyzed using multiple indicators, including aerosol optical depth (AOD), size distribution, single scattering albedo (SSA), and Ångström exponent (AE). The movements and detailed aerosol composition of the aforementioned episodes are demonstrated using NASA's Goddard Space Flight Center (GSFC) back trajectories model in collaboration with aerosol subtype products from the Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observation (CALIPSO) mission. These episodes indicate that during the spring, fall, and summer, most of the severe aerosol events are caused by dust or mixed related scenarios, whereas during winter, aerosols of finer size lead to severe heavy conditions. It also demonstrates the impacts of different aerosol sources on urban human health, which are presented by the variations of multiple parameters, including solar radiation, air temperature, humidity, and UV exposure. Scarce ground PM10 data were collected and compared against satellite products, yet owed to their discrete nature of availability, our approach made use of the Random Decision Forest (RDF) model to convert satellite-based AOD and other meteorological parameters to predict PM10. The RDF model with inputs from the Modern-Era Retrospective analysis for Research and Applications version 2 (MERRA-2) and Global Land Data Assimilation System (GLDAS) datasets improves the performance of using AOD products to estimate PM10 values. The connection between climate variability and aerosol intensity, as well as their impact on health-related PM2.5 over Egypt is also demonstrated. © 2019 by the authors." "56241073100;56335113800;25222759800;55314770300;57217367195;7405410341;","Integration of machine learning and open access geospatial data for land cover mapping",2019,"10.3390/rs11161907","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071561906&doi=10.3390%2frs11161907&partnerID=40&md5=fb8528f3221654fff8e9d3730a42d978","In-time and accurate monitoring of land cover and land use are essential tools for countries to achieve sustainable food production. However, many developing countries are struggling to efficiently monitor land resources due to the lack of financial support and limited access to adequate technology. This study aims at offering a solution to fill in such a gap in developing countries, by developing a land cover solution that is free of costs. A fully automated framework for land cover mapping was developed using 10-m resolution open access satellite images and machine learning (ML) techniques for the African country of Lesotho. Sentinel-2 satellite images were accessed through Google Earth Engine (GEE) for initial processing and feature extraction at a national level. Also, Food and Agriculture Organization's land cover of Lesotho (FAO LCL) data were used to train a support vector machine (SVM) and bagged trees (BT) classifiers. SVM successfully classified urban and agricultural lands with 62 and 67% accuracy, respectively. Also, BT could classify the two categories with 81 and 65% accuracy, correspondingly. The trained models could provide precise LC maps in minutes or hours. they can also be utilized as a viable solution for developing countries as an alternative to traditional geographic information system (GIS) methods, which are often labor intensive, require acquisition of very high-resolution commercial satellite imagery, time consuming and call for high budgets. © 2019 by the authors." "57208121325;57204578151;54684821900;57205350690;57210823248;7401925341;","Global sensitivity analysis of leaf-canopy-atmosphere RTMs: Implications for biophysical variables retrieval from top-of-atmosphere radiance data",2019,"10.3390/rs11161923","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071518620&doi=10.3390%2frs11161923&partnerID=40&md5=0d430f466ffe1eb2c47b7e7d8b708907","Knowledge of key variables driving the top of the atmosphere (TOA) radiance over a vegetated surface is an important step to derive biophysical variables from TOA radiance data, e.g., as observed by an optical satellite. Coupled leaf-canopy-atmosphere Radiative Transfer Models (RTMs) allow linking vegetation variables directly to the at-sensor TOA radiance measured. Global Sensitivity Analysis (GSA) of RTMs enables the computation of the total contribution of each input variable to the output variance. We determined the impacts of the leaf-canopy-atmosphere variables into TOA radiance using the GSA to gain insights into retrievable variables. The leaf and canopy RTM PROSAIL was coupled with the atmospheric RTM MODTRAN5. Because of MODTRAN's computational burden and GSA's demand for many simulations, we first developed a surrogate statistical learning model, i.e., an emulator, that allows approximating RTM outputs through a machine learning algorithm with low computation time. A Gaussian process regression (GPR) emulator was used to reproduce lookup tables of TOA radiance as a function of 12 input variables with relative errors of 2.4%. GSA total sensitivity results quantified the driving variables of emulated TOA radiance along the 400-2500 nm spectral range at 15 cm-1 (between 0.3-9 nm); overall, the vegetation variables play a more dominant role than atmospheric variables. This suggests the possibility to retrieve biophysical variables directly from at-sensor TOA radiance data. Particularly promising are leaf chlorophyll content, leaf water thickness and leaf area index, as these variables are the most important drivers in governing TOA radiance outside the water absorption regions. A software framework was developed to facilitate the development of retrieval models from at-sensor TOA radiance data. As a proof of concept, maps of these biophysical variables have been generated for both TOA (L1C) and bottom-of-atmosphere (L2A) Sentinel-2 data by means of a hybrid retrieval scheme, i.e., training GPR retrieval algorithms using the RTM simulations. Obtained maps from L1C vs L2A data are consistent, suggesting that vegetation properties can be directly retrieved from TOA radiance data given a cloud-free sky, thus without the need of an atmospheric correction. © 2019 by the authors." "57202430907;57208411135;7201554889;57204546393;54417881500;","Land-use and land-cover classification using Sentinel-2 data and machine-learning algorithms: Operational method and its implementation for a mountainous area of Nepal",2019,"10.1117/1.JRS.13.014530","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064161526&doi=10.1117%2f1.JRS.13.014530&partnerID=40&md5=a946e1e95d5e991015288fb1f2560dd6","In the context of land-use and land-cover (LULC) classification, there is a lack of leverage of the recent increase of the ease of access to satellite imagery data, cloud computing platforms, and classification techniques. We present both the development of an operational method for LULC classification that considers these progresses and the implementation of this operational method for a mountainous area of Nepal. The operational method allows the comparison of three LULC maps, each derived with a different classification technique [classification and regression tree (CART), max entropy (MaxEnt), and random forest (RF)] applied to Sentinel-2 data on the Google Earth Engine platform. The results derived with the RF technique have the highest overall accuracy coefficient (92%). The probabilities that the RF technique produces a more accurate LULC map than the MaxEnt (95%) and CART (61%) techniques are based on Kappa statistics. Results of general linear models suggest that some LULC types have higher producer's and user's accuracies at a statistically significant level. The operational method can help the producers of LULC maps conduct future work on areas in developing countries, as such contributing to addressing various issues that involve land use. © 2019 Society of Photo-Optical Instrumentation Engineers (SPIE)." "57194590031;56216874200;","Machine learning-based retrieval of benthic reflectance and Posidonia oceanica seagrass extent using a semi-analytical inversion of Sentinel-2 satellite data",2018,"10.1080/01431161.2018.1519289","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054865532&doi=10.1080%2f01431161.2018.1519289&partnerID=40&md5=b71cfaeb6ca51c7419348c086305aedc","In the epoch of the human-induced climate change, seagrasses can mitigate the resulting negative impacts due to their carbon sequestration ability. The endemic and dominant in the Mediterranean Posidonia oceanica seagrass contains the largest stocks of organic carbon among all seagrass species, yet it undergoes a significant regression in its extent. Therefore, suitable quantitative assessment of its extent and optically shallow environment are required to allow good conservation and management practices. Here, we parameterise a semi-analytical inversion model which employs above-surface remote sensing reflectance of Sentinel-2A to derive water column and bottom properties in the Thermaikos Gulf, NW Aegean Sea, Greece (eastern Mediterranean). In the model, the diffuse attenuation coefficients are expressed as functions of absorption and backscattering coefficients. We apply a comprehensive pre-processing workflow which includes atmospheric correction using C2RCC (Case 2 Regional CoastColour) neural network, resampling of the lower spatial resolution Sentinel-2A bands to 10m/pixel, as well as empirical derivation of water bathymetry and machine learning-based classification of the resulting bottom properties using the Support Vector Machines. SVM-based classification of benthic reflectance reveals ~300 ha of P. oceanica seagrass between 2 and 16 m of depth, and yields very high producer and user accuracies of 95.3% and 99.5%, respectively. Sources of errors and uncertainties are discussed. All in all, recent advances in Earth Observation in terms of optical satellite technology, cloud computing and machine learning algorithms have created the perfect storm which could aid high spatio-temporal, large-scale seagrass habitat mapping and monitoring, allowing for its integration to the Analysis Ready Data era and ultimately enabling more efficient management and conservation in the epoch of climate change. © 2018, © 2018 Informa UK Limited, trading as Taylor & Francis Group." "57210320800;57210317468;57209059605;55062577500;","Linear Regression Assisted Prediction Based Load Balancer for Cloud Computing",2018,"10.1109/PUNECON.2018.8745409","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070314973&doi=10.1109%2fPUNECON.2018.8745409&partnerID=40&md5=15617b63618e17cb5cd6adaa45323edd","Cloud Computing is one of the most ubiquitous technologies today. Cloud service providers are required to provide services to its users productively and efficiently. Resources must be selected and allocated properly based on the task's attributes. The objective of a load balancing algorithm is to ensure efficient and fair distribution of load among all the computing resources. The challenge today, however, is to maintain performance standards in spite of the rapidly increasing data and performance needs of the users. This motivates an automated approach towards load balancing. In this work, we proposed a load balancing algorithm based on the usage statistics which predicts the queue of virtual machines for the optimal usage of cloud resources. Experimental results shows the improvement in response time with compared to the conventional approach. © 2018 IEEE." "57202198444;7102586329;56245476200;","Evaluation of regression and neural network models for solar forecasting over different short-term horizons",2018,"10.1080/23744731.2018.1464348","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047432203&doi=10.1080%2f23744731.2018.1464348&partnerID=40&md5=7bf7ec082a7f2f33d3d7b587d6fc9c58","Forecasting solar irradiation has acquired immense importance in view of the exponential increase in the number of solar photovoltaic (PV) system installations. In this article, analyses results involving statistical and machine-learning techniques to predict solar irradiation for different forecasting horizons are reported. Yearlong typical meteorological year 3 (TMY3) datasets from three cities in the United States with different climatic conditions have been used in this analysis. A simple forecast approach that assumes consecutive days to be identical serves as a baseline model to compare forecasting alternatives. To account for seasonal variability and to capture short-term fluctuations, different variants of the lagged moving average (LMX) model with cloud cover as the input variable are evaluated. Finally, the proposed LMX model is evaluated against an artificial neural network (ANN) model. How the one-hour and 24-hour models can be used in conjunction to predict different short-term rolling horizons is discussed, and this joint application is illustrated for a four-hour rolling horizon forecast scheme. Finally, the effect of using predicted cloud cover values, instead of measured ones, on the accuracy of the models is assessed. Results show that LMX models do not degrade in forecast accuracy if models are trained with the forecast cloud cover data. © 2018, Copyright © 2018 ASHRAE." "37023334200;57198861721;7006011851;","Modelling the structure of star clusters with fractional Brownian motion",2018,"10.1093/mnras/sty1788","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85052585957&doi=10.1093%2fmnras%2fsty1788&partnerID=40&md5=35bcba2cea48afcb31f1fa6ecf22ac9a","The degree of fractal substructure in molecular clouds can be quantified by comparing them with fractional Brownian motion (FBM) surfaces or volumes. These fields are self-similar over all length-scales and characterized by a drift exponent H, which describes the structural roughness. Given that the structure of molecular clouds and the initial structure of star clusters are almost certainly linked, it would be advantageous to also apply this analysis to clusters. Currently, the structure of star clusters is often quantified by applying Q analysis. Q values from observed targets are interpreted by comparing them with those from artificial clusters. These are typically generated using a box-fractal (BF) or radial density profile (RDP) model. We present a single cluster model, based on FBM, as an alternative to these models. Here, the structure is parametrized by H and the standard deviation of the log-surface/volume density σ. The FBM model is able to reproduce both centrally concentrated and substructured clusters, and is able to provide a much better match to observations than the BF model. We show that Q analysis is unable to estimate FBM parameters. Therefore, we develop and train a machine learning algorithm that can estimate values of H and σ, with uncertainties. This provides us with a powerful method for quantifying the structure of star clusters in terms that relate to the structure of molecular clouds. We use the algorithm to estimate the H and σ for several young star clusters, some of which have no measurable BF or RDP analogue. © 2018 The Author(s). Published by Oxford University Press on behalf of the Royal Astronomical Society." "56641880000;8977349800;7006213017;","Prioritized multi-view stereo depth map generation using confidence prediction",2018,"10.1016/j.isprsjprs.2018.03.022","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85044727874&doi=10.1016%2fj.isprsjprs.2018.03.022&partnerID=40&md5=2c52db6dfd9fa8736ea7b3b9d35fdcb7","In this work, we propose a novel approach to prioritize the depth map computation of multi-view stereo (MVS) to obtain compact 3D point clouds of high quality and completeness at low computational cost. Our prioritization approach operates before the MVS algorithm is executed and consists of two steps. In the first step, we aim to find a good set of matching partners for each view. In the second step, we rank the resulting view clusters (i.e. key views with matching partners) according to their impact on the fulfillment of desired quality parameters such as completeness, ground resolution and accuracy. Additional to geometric analysis, we use a novel machine learning technique for training a confidence predictor. The purpose of this confidence predictor is to estimate the chances of a successful depth reconstruction for each pixel in each image for one specific MVS algorithm based on the RGB images and the image constellation. The underlying machine learning technique does not require any ground truth or manually labeled data for training, but instead adapts ideas from depth map fusion for providing a supervision signal. The trained confidence predictor allows us to evaluate the quality of image constellations and their potential impact to the resulting 3D reconstruction and thus builds a solid foundation for our prioritization approach. In our experiments, we are thus able to reach more than 70% of the maximal reachable quality fulfillment using only 5% of the available images as key views. For evaluating our approach within and across different domains, we use two completely different scenarios, i.e. cultural heritage preservation and reconstruction of single family houses. © 2018 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "7102112270;6701529090;41360958400;54963768800;56291196300;56291092500;25822604000;7101975791;22635005800;","A Replicated Network Approach to ‘Big Data’ in Ecology",2018,"10.1016/bs.aecr.2018.04.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048721028&doi=10.1016%2fbs.aecr.2018.04.001&partnerID=40&md5=e168c1937826fc66f6eac3f354c005a8","Global environmental change is a pressing issue as evidenced by the rise of extreme weather conditions in many parts of the world, threatening the survival of vulnerable species and habitats. Effective monitoring of climatic and anthropogenic impacts is therefore critical to safeguarding ecosystems, and it would allow us to better understand their response to stressors and predict long-term impacts. Ecological networks provide a biomonitoring framework for examining the system-level response and functioning of an ecosystem, but have been, until recently, constrained by limited empirical data due to the laborious nature of their construction. Hence, most experimental designs have been confined to a single network or a small number of replicate networks, resulting in statistical uncertainty, low resolution, limited spatiotemporal scale and oversimplified assumptions. Advances in data sampling and curation methodologies, such as next-generation sequencing (NGS) and the Internet ‘Cloud’, have facilitated the emergence of the ‘Big Data’ phenomenon in Ecology, enabling the construction of ecological networks to be carried out effectively and efficiently. This provides to ecologists an excellent opportunity to expand the way they study ecological networks. In particular, highly replicated networks are now within our grasp if new NGS technologies are combined with machine learning to develop network building methods. A replicated network approach will allow temporal and spatial variations embedded in the data to be taken into consideration, overcoming the limitations in the current 'single network’ approach. We are still at the embryonic stage in exploring replicated networks, and with these new opportunities we also face new challenges. In this chapter, we discuss some of these challenges and highlight potential approaches that will help us build and analyse replicated networks to better understand how complex ecosystems operate, and the services and functioning they provide, paving the way for deciphering ecological big data reliably in the future. © 2018 Elsevier Ltd" "57191670651;56161261300;57193995429;57201649886;57201649041;57201648338;57201644801;","Development of a Real-Time Stroke Detection System for Elderly Drivers Using Quad-Chamber Air Cushion and IoT Devices",2018,"10.4271/2018-01-0046","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045515678&doi=10.4271%2f2018-01-0046&partnerID=40&md5=928a289e8fae35df329ff443bf101f8a","IoT (Internet of things) is considered most innovative technology in smart healthcare monitoring system which is able to show real-time physiological parameters feed data to web cloud, analysis using machine learning, artificial intelligence and big data. Stroke is most deadly diseases and real-time monitoring is desired to detect stroke onset during regular activities. The aim of our study is to develop a Real-time health monitoring system for elderly drivers using air cushion seat and IoT devices in order to detect stroke onset during driving. We have also made a prototype of brain stroke detection system using Quad-chamber air cushion system and IoT devices. This system can measure ECG, EEG, heart rate, seat pressure balance data, face/eye tracking etc. using IoT sensors, compare real-time data with reference data, predict abnormality, generate alarm and send message to relatives and emergency services if any stroke onset happens in order to provide emergency assistance to driver. © 2018 SAE International. All Rights Reserved." "6603354695;57192947904;8213128500;6603888005;","Cloud detection machine learning algorithms for PROBA-V",2017,"10.1109/IGARSS.2017.8127437","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041830531&doi=10.1109%2fIGARSS.2017.8127437&partnerID=40&md5=83b76ce567cbc0caee960a494aaea3bc","This paper presents the development and implementation of a cloud detection algorithm for Proba-V. Accurate and automatic detection of clouds in satellite scenes is a key issue for a wide range of remote sensing applications. With no accurate cloud masking, undetected clouds are one of the most significant sources of error in both sea and land cover biophysical parameter retrieval. The objective of the algorithms presented in this paper is to detect clouds accurately providing a cloud flag per pixel. For this purpose, the method exploits the information of Proba-V using statistical machine learning techniques to identify the clouds present in Proba-V products. The effectiveness of the proposed method is successfully illustrated using a large number of real Proba-V images. © 2017 IEEE." "54796852600;56274915100;57201106974;","Estimating extinction using unsupervised machine learning",2017,"10.1051/0004-6361/201630032","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85019704685&doi=10.1051%2f0004-6361%2f201630032&partnerID=40&md5=bd5b770c99267ea062b73f2f0f68757f","Dust extinction is the most robust tracer of the gas distribution in the interstellar medium, but measuring extinction is limited by the systematic uncertainties involved in estimating the intrinsic colors to background stars. In this paper we present a new technique, Pnicer, that estimates intrinsic colors and extinction for individual stars using unsupervised machine learning algorithms. This new method aims to be free from any priors with respect to the column density and intrinsic color distribution. It is applicable to any combination of parameters and works in arbitrary numbers of dimensions. Furthermore, it is not restricted to color space. Extinction toward single sources is determined by fitting Gaussian mixture models along the extinction vector to (extinction-free) control field observations. In this way it becomes possible to describe the extinction for observed sources with probability densities, rather than a single value. Pnicer effectively eliminates known biases found in similar methods and outperforms them in cases of deep observational data where the number of background galaxies is significant, or when a large number of parameters is used to break degeneracies in the intrinsic color distributions. This new method remains computationally competitive, making it possible to correctly de-redden millions of sources within a matter of seconds. With the ever-increasing number of large-scale high-sensitivity imaging surveys, Pnicer offers a fast and reliable way to efficiently calculate extinction for arbitrary parameter combinations without prior information on source characteristics. The Pnicer software package also offers access to the well-established Nicer technique in a simple unified interface and is capable of building extinction maps including the Nicest correction for cloud substructure. Pnicer is offered to the community as an open-source software solution and is entirely written in Python. © 2017 ESO." "57193607019;9132950200;57212198580;57022224600;","A secure data enclave and analytics platform for social scientists",2017,"10.1109/eScience.2016.7870917","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85016799019&doi=10.1109%2feScience.2016.7870917&partnerID=40&md5=985913b185f18aa7a792f4e6a88a0851","Data-driven research is increasingly ubiquitous and data itself is a defining asset for researchers, particularly in the computational social sciences and humanities. Entire careers and research communities are built around valuable, proprietary or sensitive datasets. However, many existing computation resources fail to support secure and cost-effective storage of data while also enabling secure and flexible analysis of the data. To address these needs we present CLOUD KOTTA, a cloud-based architecture for the secure management and analysis of social science data. CLOUD KOTTA leverages reliable, secure, and scalable cloud resources to deliver capabilities to users, and removes the need for users to manage complicated infrastructure. CLOUD KOTTA implements automated, cost-aware models for efficiently provisioning tiered storage and automatically scaled compute resources. CLOUD KOTTA has been used in production for several months and currently manages approximately 10TB of data and has been used to process more than 5TB of data with over 75,000 CPU hours. It has been used for a broad variety of text analysis workflows, matrix factorization, and various machine learning algorithms, and more broadly, it supports fast, secure and cost-effective research. © 2016 IEEE." "55576827900;57197143203;6701500641;55603109000;56220554800;","An approach for automated lithological classification of point clouds",2016,"10.1130/GES01326.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85006511966&doi=10.1130%2fGES01326.1&partnerID=40&md5=de3c88babbbe48b7ab1b3fef512b1104","Terrestrial light detection and ranging (LiDAR) data can be acquired from either static or mobile platforms. The latter presents some challenges in terms of resolution and accuracy, but the opportunity to cover a larger region and repeat surveys often prevails in practice. This paper presents a machine learning algorithm (MLA) for automated lithological classification of individual points within LiDAR point clouds based on intensity and geometry information. Two example data sets were collected by static and mobile platforms in an oil sands pit mine and the MLA was trained to distinguish sandstone and mudstone laminations. The type of approach presented here has the potential to be developed and applied for geological mapping applications such as reservoir characterization or underground excavation face mapping. © 2016 Geological Society of America." "57156634800;12645007600;7102192954;55192437200;57188723177;22952101400;","Evaluating the potential of proba-v satellite image time series for improving lc classification in semi-arid african landscapes",2016,"10.3390/rs8120987","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85002838340&doi=10.3390%2frs8120987&partnerID=40&md5=b62e1843857436be13adabbf53b9cb1b","Satellite based land cover classification for Africa's semi-arid ecosystems is hampered commonly by heterogeneous landscapes with mixed vegetation and small scale land use. Higher spatial resolution remote sensing time series data can improve classification results under these difficult conditions. While most large scale land cover mapping attempts rely on moderate resolution data, PROBA-V provides five-daily time series at 100mspatial resolution. This improves spatial detail and resilience against high cloud cover, but increases the data load. Cloud-based processing platforms can leverage large scale land cover monitoring based on such finer time series. We demonstrate this with PROBA-V 100 m time series data from 2014-2015, using temporal metrics and cloud filtering in combination with in-situ training data and machine learning, implemented on the ESA (European Space Agency) Cloud Toolbox infrastructure. We apply our approach to two use cases for a large study area over West Africa: land-and forest cover classification. Our land cover classification reaches a 7% to 21% higher overall accuracy when compared to four global land cover maps (i.e., Globcover-2009, Cover-CCI-2010, MODIS-2010, and Globeland30). Our forest cover classification shows 89% correspondence with the Tropical Ecosystem Environment Observation System (TREES)-3 forest cover data which is based on spatially finer Landsat data. This paper illustrates a proof of concept for cloud-based ""big-data"" driven land cover monitoring. Furthermore, we show that a wide range of temporal metrics can be extracted from detailed PROBA-V 100 m time series data to continuously optimize land cover monitoring. © 2016 by the authors; licensee MDPI, Basel, Switzerland." "23486298200;35502407200;6507070902;36718621300;55009105200;6507474628;55050209700;57188728777;","Morphological evidences and computer science techniques in order to evaluate tsunami inundation limit [Evidenze morfologiche ed applicazioni informatiche al fine della valutazione del limite di inondazione da tsunami]",2010,"10.5721/itjrs201042210","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84857754070&doi=10.5721%2fitjrs201042210&partnerID=40&md5=66287bc579ff6c18d76fa67cbb081386","Terrestrial Laser Scanner surveys performed in coastal area have generated 3D cloud points used to obtain digital elevation model and standard deviation of the micro-topography of coastal surfaces. Starting from data collected, roughness coefficients have been estimated for each surface typology characterizing the coastal area (sand, calcarenite, vegetation, etc). Applying Machine Learning techniques on digital images, the extension and the surface typology of these areas have been obtained. All data collected have been elaborated by means of software implemented starting from known hydrodynamic formula to evaluate the inland penetration of a hypothesized tsunami." "57204481038;56898969100;56684747900;57190089926;16070064500;55178361600;","Improved land cover map of Iran using Sentinel imagery within Google Earth Engine and a novel automatic workflow for land cover classification using migrated training samples",2020,"10.1016/j.isprsjprs.2020.07.013","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089363199&doi=10.1016%2fj.isprsjprs.2020.07.013&partnerID=40&md5=2514a5cac51d09ff890a69a9624cbaa3","Accurate information about the location, extent, and type of Land Cover (LC) is essential for various applications. The only recent available country-wide LC map of Iran was generated in 2016 by the Iranian Space Agency (ISA) using Moderate Resolution Imaging Spectroradiometer (MODIS) images with a considerably low accuracy. Therefore, the production of an up-to-date and accurate Iran-wide LC map using the most recent remote sensing, machine learning, and big data processing algorithms is required. Moreover, it is important to develop an efficient method for automatic LC generation for various time periods without the need to collect additional ground truth data from this immense country. Therefore, this study was conducted to fulfill two objectives. First, an improved Iranian LC map with 13 LC classes and a spatial resolution of 10 m was produced using multi-temporal synergy of Sentinel-1 and Sentinel-2 satellite datasets applied to an object-based Random forest (RF) algorithm. For this purpose, 2,869 Sentinel-1 and 11,994 Sentinel-2 scenes acquired in 2017 were processed and classified within the Google Earth Engine (GEE) cloud computing platform allowing big geospatial data analysis. The Overall Accuracy (OA) and Kappa Coefficient (KC) of the final Iran-wide LC map for 2017 was 95.6% and 0.95, respectively, indicating the considerable potential of the proposed big data processing method. Second, an efficient automatic method was developed based on Sentinel-2 images to migrate ground truth samples from a reference year to automatically generate an LC map for any target year. The OA and KC for the LC map produced for the target year 2019 were 91.35% and 0.91, respectively, demonstrating the efficiency of the proposed method for automatic LC mapping. Based on the obtained accuracies, this method can potentially be applied to other regions of interest for LC mapping without the need for ground truth data from the target year. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57194694881;6507122674;","Machine learning generalisation across different 3D architectural heritage",2020,"10.3390/ijgi9060379","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086237519&doi=10.3390%2fijgi9060379&partnerID=40&md5=93a39656d49651816dd13739b376dfc7","The use of machine learning techniques for point cloud classification has been investigated extensively in the last decade in the geospatial community, while in the cultural heritage field it has only recently started to be explored. The high complexity and heterogeneity of 3D heritage data, the diversity of the possible scenarios, and the different classification purposes that each case study might present, makes it difficult to realise a large training dataset for learning purposes. An important practical issue that has not been explored yet, is the application of a single machine learning model across large and different architectural datasets. This paper tackles this issue presenting a methodology able to successfully generalise to unseen scenarios a random forest model trained on a specific dataset. This is achieved looking for the best features suitable to identify the classes of interest (e.g., wall, windows, roof and columns). © 2020 by the authors." "57203766164;17346547400;14825916200;","Smart production planning and control: Concept, use-cases and sustainability implications",2020,"10.3390/su12093791","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084836845&doi=10.3390%2fsu12093791&partnerID=40&md5=877536aeb1b6932e61c98a0d73e1d493","Many companies are struggling to manage their production systems due to increasing market uncertainty. While emerging 'smart' technologies such as the internet of things, machine learning, and cloud computing have been touted as having the potential to transform production management, the realities of their adoption and use have been much more challenging than anticipated. In this paper, we explore these challenges and present a conceptual model, a use-case matrix and a product-process framework for a smart production planning and control (smart PPC) system and illustrate the use of these artefacts through four case companies. The presented model adopts an incremental approach that companies with limited resources could employ in improving their PPC process in the context of industry 4.0 and sustainability. The results reveal that while make-to-order companies are more likely to derive greater benefits from a smart product strategy, make-to-stock companies are more likely to derive the most benefit from pursuing a smart process strategy, and consequently a smart PPC solution. © 2020 by the authors." "57209238608;55574436900;56890893700;57192690229;","Improved Supervised Learning-Based Approach for Leaf and Wood Classification from LiDAR Point Clouds of Forests",2020,"10.1109/TGRS.2019.2947198","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084151595&doi=10.1109%2fTGRS.2019.2947198&partnerID=40&md5=d009971641b97a63f058acac140d58a3","Accurately classifying 3-D point clouds into woody and leafy components has been an interest for applications in forestry and ecology including the better understanding of radiation transfer between canopy and atmosphere. The past decade has seen an increase in the methods attempting to classify leaves and wood in point clouds based on radiometric or geometric features. However, classification purely based on radiometric features is sensor-specific, and the method by which the local neighborhood of a point is defined affects the accuracy of classification based on geometric features. Here, we present a leaf-wood classification method combining geometrical features defined by radially bounded nearest neighbors at multiple spatial scales in a machine learning model. We compared the performance of three different machine learning models generated by the random forest (RF), XGBoost, and lightGBM algorithms. Using multiple spatial scales eliminates the need for an optimal neighborhood size selection and defining the local neighborhood by radially bounded nearest neighbors makes the method broadly applicable for point clouds of varying quality. We assessed the model performance at the individual tree-and plot-level on field data from tropical and deciduous forests, as well as on simulated point clouds. The method has an overall average accuracy of 94.2% on our data sets. For other data sets, the presented method outperformed the methods in literature in most cases without the need for additional postprocessing steps that are needed in most of the existing methods. We provide the entire framework as an open-source python package. © 1980-2012 IEEE." "57189347119;7005981926;36505022100;57207876086;57200270007;6602453684;24075297800;56727588300;15848875300;57192677861;7003997306;","Comparative performance of convolutional neural network, weighted and conventional support vector machine and random forest for classifying tree species using hyperspectral and photogrammetric data",2020,"10.1080/15481603.2020.1712102","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078586550&doi=10.1080%2f15481603.2020.1712102&partnerID=40&md5=dff44c1a0a971000d51bca599040160a","The classification of tree species can significantly benefit from high spatial and spectral information acquired by unmanned aerial vehicles (UAVs) associated with advanced classification methods. This study investigated the following topics concerning the classification of 16 tree species in two subtropical forest fragments of Southern Brazil: i) the potential integration of UAV-borne hyperspectral images with 3D information derived from their photogrammetric point cloud (PPC); ii) the performance of two machine learning methods (support vector machine–SVM and random forest–RF) when employing different datasets at a pixel and individual tree crown (ITC) levels; iii) the potential of two methods for dealing with the imbalanced sample set problem: a new weighted SVM (wSVM) approach, which attributes different weights to each sample and class, and a deep learning classifier (convolutional neural network–CNN), associated with a previous step to balance the sample set; and finally, iv) the potential of this last classifier for tree species classification as compared to the above mentioned machine learning methods. Results showed that the inclusion of the PPC features to the hyperspectral data provided a great accuracy increase in tree species classification results when conventional machine learning methods were applied, between 13 and 17% depending on the classifier and the study area characteristics. When using the PPC features and the canopy height model (CHM), associated with the majority vote (MV) rule, the SVM, wSVM and RF classifiers reached accuracies similar to the CNN, which outperformed these classifiers for both areas when considering the pixel-based classifications (overall accuracy of 84.4% in Area 1, and 74.95% in Area 2). The CNN was between 22% and 26% more accurate than the SVM and RF when only the hyperspectral bands were employed. The wSVM provided a slight increase in accuracy not only for some lesser represented classes, but also some major classes in Area 2. While conventional machine learning methods are faster, they demonstrated to be less stable to changes in datasets, depending on prior segmentation and hand-engineered features to reach similar accuracies to those attained by the CNN. To date, CNNs have been barely explored for the classification of tree species, and CNN-based classifications in the literature have not dealt with hyperspectral data specifically focusing on tropical environments. This paper thus presents innovative strategies for classifying tree species in subtropical forest areas at a refined legend level, integrating UAV-borne 2D hyperspectral and 3D photogrammetric data and relying on both deep and conventional machine learning approaches. © 2020, © 2020 Informa UK Limited, trading as Taylor & Francis Group." "55923557900;57217766925;55924332100;","Digital twin and CyberGIS for improving connectivity and measuring the impact of infrastructure construction planning in smart cities",2020,"10.3390/ijgi9040240","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083321923&doi=10.3390%2fijgi9040240&partnerID=40&md5=ef3d9bdc2455373d422e38881a5d13b4","Smart technologies are advancing, and smart cities can be made smarter by increasing the connectivity and interactions of humans, the environment, and smart devices. This paper discusses selective technologies that can potentially contribute to developing an intelligent environment and smarter cities. While the connectivity and efficiency of smart cities is important, the analysis of the impact of construction development and large projects in the city is crucial to decision and policy makers, before the project is approved. This raises the question of assessing the impact of a new infrastructure project on the community prior to its commencement-what type of technologies can potentially be used for creating a virtual representation of the city? How can a smart city be improved by utilizing these technologies? There are a wide range of technologies and applications available but understanding their function, interoperability, and compatibility with the community requires more discussion around system designs and architecture. These questions can be the basis of developing an agenda for further investigations. In particular, the need for advanced tools such as mobile scanners, Geospatial Artificial Intelligence, Unmanned Aerial Vehicles, Geospatial Augmented Reality apps, Light Detection, and Ranging in smart cities is discussed. In line with smart city technology development, this Special Issue includes eight accepted articles covering trending topics, which are briefly reviewed. © 2020 by the authors." "57206424059;55720588700;55574865800;57207451960;7102063963;","Retrieval of cloud top properties from advanced geostationary satellite imager measurements based on machine learning algorithms",2020,"10.1016/j.rse.2019.111616","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076917542&doi=10.1016%2fj.rse.2019.111616&partnerID=40&md5=a4b836175e27ac72d2a0f3408df46d44","The cloud-top height (CTH) product derived from passive satellite instrument measurements is often used to make climate data records (CDR). CALIPSO (Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observations) provides CTH parameters with high accuracy, but with limited temporal-spatial resolution. Recently, the Advanced Himawari Imager (AHI) onboard Japanese Himawari-8/-9, provides high temporal (every 10 min) and high spatial (2 km at nadir) resolution measurements with 16 spectral bands. This paper reports on a study to derive the CTH from combined AHI and CALIPSO using advanced machine learning (ML) algorithms with better accuracy than that from the traditional physical (TRA) algorithms. We find significant CTH improvements (1.54–2.72 km for mean absolute error, MAE) from four different machine learning algorithms (original MAE from TRA method is about 3.24 km based on CALIPSO data validation), particularly in high and optically thin clouds. In addition, we also develop a joint algorithm to combine optimal machine learning and traditional physical (TRA) algorithms of CTH to further reduce MAE to 1.53 km and enhance the layered accuracy (CTH < 18 km). While the ML-based algorithm improves CTH retrieval over the TRA algorithm, the lower or higher clouds still exhibit relatively large uncertainty. Combining both methods provides the better CTH than either alone. The combined approach could be used to process data from advanced geostationary imagers for climate and weather applications. © 2019" "57200918295;24072139200;57203833103;56545781200;57204877883;8632957700;","Oil palm mapping over Peninsular Malaysia using Google Earth Engine and machine learning algorithms",2020,"10.1016/j.rsase.2020.100287","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078101296&doi=10.1016%2fj.rsase.2020.100287&partnerID=40&md5=0f2cbd69ee8dccde76c778f2f802d28e","Oil palm plays a pivotal role in the ecosystem, environment, economy and without proper monitoring, uncontrolled oil palm activities could contribute to deforestation that can cause high negative impacts on the environment and therefore, proper management and monitoring of the oil palm industry are necessary. Mapping the distribution of oil palm is crucial in order to manage and plan the sustainable operations of oil palm plantations. Remote sensing provides a means to detect and map oil palm from space effectively. Recent advances in cloud computing and big data allow rapid mapping to be performed over large a geographical scale. In this study, 30 m Landsat 8 data were processed using a cloud computing platform of Google Earth Engine (GEE) in order to classify oil palm land cover using non-parametric machine learning algorithms such as Support Vector Machine (SVM), Classification and Regression Tree (CART) and Random Forest (RF) for the first time over Peninsular Malaysia. The hyperparameters were tuned, and the overall accuracy produced by the SVM, CART and RF were 93.16%, 80.08% and 86.50% respectively. Overall, the SVM classified the 7 classes (water, built-up, bare soil, forest, oil palm, other vegetation and paddy) the best. However, RF extracted oil palm information better than the SVM. The algorithms were compared and the McNemar's test showed significant values for comparisons between SVM and CART and RF and CART. On the other hand, the performance of SVM and RF are considered equally effective. Despite the challenges in implementing machine learning optimisation using GEE over a large area, this paper shows the efficiency of GEE as a cloud-based free platform to perform bioresource distributions mapping such as oil palm over a large area in Peninsular Malaysia. © 2020 Elsevier B.V." "57208602700;6602377076;57204354106;","Leveraging Google Earth Engine (GEE) and machine learning algorithms to incorporate in situ measurement from different times for rangelands monitoring",2020,"10.1016/j.rse.2019.111521","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074748720&doi=10.1016%2fj.rse.2019.111521&partnerID=40&md5=13cc2c1d8ae1cd7ef815f2e83a5afc3e","Mapping and monitoring of indicators of soil cover, vegetation structure, and various native and non-native species is a critical aspect of rangeland management. With the advancement in satellite imagery as well as cloud storage and computing, the capability now exists to conduct planetary-scale analysis, including mapping of rangeland indicators. Combined with recent investments in the collection of large amounts of in situ data in the western U.S., new approaches using machine learning can enable prediction of surface conditions at times and places when no in situ data are available. However, little analysis has yet been done on how the temporal relevancy of training data influences model performance. Here, we have leveraged the Google Earth Engine (GEE) platform and a machine learning algorithm (Random Forest, after comparison with other candidates) to identify the potential impact of different sampling times (across months and years) on estimation of rangeland indicators from the Bureau of Land Management's (BLM) Assessment, Inventory, and Monitoring (AIM) and Landscape Monitoring Framework (LMF) programs. Our results indicate that temporally relevant training data improves predictions, though the training data need not be from the exact same month and year for a prediction to be temporally relevant. Moreover, inclusion of training data from the time when predictions are desired leads to lower prediction error but the addition of training data from other times does not contribute to overall model error. Using all of the available training data can lead to biases, toward the mean, for times when indicator values are especially high or low. However, for mapping purposes, limiting training data to just the time when predictions are desired can lead to poor predictions of values outside the spatial range of the training data for that period. We conclude that the best Random Forest prediction maps will use training data from all possible times with the understanding that estimates at the extremes will be biased. © 2019 Elsevier Inc." "57203745186;38062095100;57211925496;35104961900;55972822400;35189680700;57191270129;57193401553;","Estimating maize above-ground biomass using 3D point clouds of multi-source unmanned aerial vehicle data at multi-spatial scales",2019,"10.3390/rs11222678","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075339132&doi=10.3390%2frs11222678&partnerID=40&md5=b43134f50b855660edcb0e17c300e784","Crop above-ground biomass (AGB) is a key parameter used for monitoring crop growth and predicting yield in precision agriculture. Estimating the crop AGB at a field scale through the use of unmanned aerial vehicles (UAVs) is promising for agronomic application, but the robustness of the methods used for estimation needs to be balanced with practical application. In this study, three UAV remote sensing flight missions (using a multiSPEC-4C multispectral camera, a Micasense RedEdge-M multispectral camera, and an Alpha Series AL3-32 Light Detection and Ranging (LiDAR) sensor onboard three different UAV platforms) were conducted above three long-term experimental plots with different tillage treatments in 2018. We investigated the performances of the multi-source UAV-based 3D point clouds at multi-spatial scales using the traditional multivariable linear regression model (OLS), random forest (RF), backpropagation neural network (BP), and support vector machine (SVM) methods for accurate AGB estimation. Results showed that crop height (CH) was a robust proxy for AGB estimation, and that high spatial resolution in CH datasets helps to improve maize AGB estimation. Furthermore, the OLS, RF, BP, and SVM methods all maintained an acceptable accuracy for AGB estimation; however, the SVM and RF methods performed slightly more robustly. This study is expected to optimize UAV systems and algorithms for specific agronomic applications. © 2019 by the authors." "57207264404;56437083400;55368886300;55363375100;36698938800;57210846238;","Mapping sugarcane in complex landscapes by integrating multi-temporal Sentinel-2 images and machine learning algorithms",2019,"10.1016/j.landusepol.2019.104190","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071667146&doi=10.1016%2fj.landusepol.2019.104190&partnerID=40&md5=36508d64dbb68778b86d83897fd56c78","Sugarcane is an important type of cash crop and plays a crucial role in global sugar production. Clarifying the magnitude of sugarcane planting will likely provide very evident supports for local land use management and policy-making. However, sugarcane growth environment in complex landscapes with frequent rainy weather conditions poses many challenges for its rapid mapping. This study thus tried and used 10-m Sentinel-2 images as well as crop phenology information to map sugarcane in Longzhou county of China in 2018. To minimize the influences of cloudy and rainy conditions, this study firstly fused all available images in each phenology stage to obtain cloud-free remote sensing images of three phenology stage (seedling, elongation and harvest) with the help of Google Earth Engine platform. Then, the study used the fused images to compute the normalized difference vegetation index (NDVI) of each stage. A three-band NDVI dataset along with 4000 training samples and 2000 random validation samples was finally used for sugarcane mapping. To assess the robustness of the three-band NDVI dataset with phenological characteristics for sugarcane mapping, this study employed five classifiers based on machine learning algorithms, including two support vector machine classifiers (Polynomial-SVM and RBF-SVM), a random forest classifier (RF), an artificial neural network classifier (ANN) and a decision tree classifier (CART-DT). Results showed that except for ANN classifier, Polynomial-SVM, RBF-SVM, RF and CART-DT classifiers displayed high accuracy sugarcane resultant maps with producer's and user's accuracies of greater than 91%. The ANN classifier tended to overestimate area of sugarcane and underestimate area of forests. Overall performances of five classifiers suggest Polynomial-SVM has the best potential to improve sugarcane mapping at the regional scale. Also, this study observed that most sugarcane (more than 75% of entire study area) tends to grow in flat regions with slope of less than 10°. This study emphasizes the importance of considering phenology in rapid sugarcane mapping, and suggests the potential of fine-resolution Sentinel-2 images and machine learning approaches in high-accuracy land use management and decision-making. © 2019 Elsevier Ltd" "56019121300;7403557805;8977584600;8528600400;","Identifying Subsurface Drainage using Satellite Big Data and Machine Learning via Google Earth Engine",2019,"10.1029/2019WR024892","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074301469&doi=10.1029%2f2019WR024892&partnerID=40&md5=36ec45796397c2e71b00e3365abaa96a","Human-induced landscape changes affect hydrologic responses (e.g., floods) that can be detected from a suite of satellite and model data sets. Tapping these vast data sets using machine learning algorithms can produce critically important and accurate insights. In the Red River of the North Basin in the United States, agricultural subsurface drainage (SD; so-called tile drainage) systems have greatly increased since the late 1990s. Over this period, river flow in the Red River has markedly increased and 6 of 13 major floods during the past century have occurred in the past two decades. The impact of SD systems on river flow is elusive because there are surprisingly few SD records in the United States. In this study, Random Forest machine learning (RFML) classification method running on Google Earth Engine's cloud computing platform was able to capture SD within a field (30 m) and its expansion over time for a large watershed (>100,000 km2). The resulting RFML classifier drew from operational multiple satellites and model data sets (total 14 variables with 36 layers including vegetation, land cover, soil properties, and climate variables). The classifier identified soil properties and land surface temperature to be the strongest predictors of SD. The maps agreed well with SD permit records (overall accuracies of 76.9–87.0%) and corresponded with subwatershed-level statistics (r = 0.77–0.96). It is expected that the maps produced with this data-intensive machine learning approach will help water resource managers to assess the hydrological impact from SD expansion and improve flood predictions in SD-dominated regions. ©2019. American Geophysical Union. All Rights Reserved." "55548541000;57201541167;8053177000;6507214354;","DepthLearn: Learning to correct the refraction on point clouds derived from aerial imagery for accurate dense shallow water bathymetry based on SVMs-fusion with LiDAR point clouds",2019,"10.3390/rs11192225","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073429435&doi=10.3390%2frs11192225&partnerID=40&md5=491ff32f4d8640f2acd7e29013b527a9","The determination of accurate bathymetric information is a key element for near offshore activities; hydrological studies, such as coastal engineering applications, sedimentary processes, hydrographic surveying, archaeological mapping and biological research. Through structure from motion (SfM) and multi-view-stereo (MVS) techniques, aerial imagery can provide a low-cost alternative compared to bathymetric LiDAR (Light Detection and Ranging) surveys, as it offers additional important visual information and higher spatial resolution. Nevertheless, water refraction poses significant challenges on depth determination. Till now, this problem has been addressed through customized image-based refraction correction algorithms or by modifying the collinearity equation. In this article, in order to overcome the water refraction errors in a massive and accurate way, we employ machine learning tools, which are able to learn the systematic underestimation of the estimated depths. In particular, an SVR (support vector regression) model was developed, based on known depth observations from bathymetric LiDAR surveys, which is able to accurately recover bathymetry from point clouds derived from SfM-MVS procedures. Experimental results and validation were based on datasets derived from different test-sites, and demonstrated the high potential of our approach. Moreover, we exploited the fusion of LiDAR and image-based point clouds towards addressing challenges of both modalities in problematic areas. © 2019 by the authors." "57207244189;15847387700;7004126410;57217546608;57207256541;","Comparison of machine learning algorithms for classification of LiDAR points for characterization of canola canopy structure",2019,"10.1080/01431161.2019.1584929","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85062337072&doi=10.1080%2f01431161.2019.1584929&partnerID=40&md5=c7b835e35ad78cac8619aef96a118fd6","Recent changes to plant architectural traits that influence the canopy have produced high yielding cultivars in rice, wheat and maize. In breeding programs, rapid assessments of the crop canopy and other structural traits are needed to facilitate the development of advanced cultivars in other crops such as Canola. LiDAR has the potential to provide insights into plant structural traits such as canopy height, aboveground biomass, and light penetration. These parameters all rely heavily on classifying LiDAR returns as ground or vegetation as they rely on the number of ground returns and the number of vegetation returns. The aim of this study is to propose a point classification method for canola using machine learning approach. The training and testing datasets were clusters sampled from field plots for flower, plant and ground. The supervised learning algorithms chosen are Decision Tree, Random Forest, Support Vector Machine, and Naïve Bayes. K-means Clustering was also used as an unsupervised learning algorithm. The results show that Random Forest models (error rate = 0.006%) are the most accurate to use for canola point classification, followed by Support Vector Machine (0.028%) and Decision Tree (0.169%). Naïve Bayes (2.079%) and K-means Clustering (48.806%) are not suitable for this purpose. This method provides the true ground and canopy in point clouds rather than determining ground points via a fixed height rely on the accuracy of the point clouds, subsequently gives more representative measurements of the crop canopy. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57190445030;6602848822;7006222187;","Active-Passive Surface Water Classification: A New Method for High-Resolution Monitoring of Surface Water Dynamics",2019,"10.1029/2019GL082562","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065332475&doi=10.1029%2f2019GL082562&partnerID=40&md5=1c7d1e56368ebc291777c93ddf6985a7","This study develops a new, highly efficient method to produce accurate, high-resolution surface water maps. The “active-passive surface water classification” method leverages cloud-based computing resources and machine learning techniques to merge Sentinel 1 synthetic aperture radar and Landsat observations and generate monthly 10-m-resolution water body maps. The skill of the active-passive surface water classification method is demonstrated by mapping surface water change over the Awash River basin in Ethiopia during the 2015 East African regional drought and 2016 localized flood events. Errors of omission (water incorrectly classified as nonwater) and commission (nonwater incorrectly classified as water) in the case study area are 7.16% and 1.91%, respectively. The case study demonstrates the method's ability to generate accurate, high-resolution water body maps depicting surface water dynamics in data-sparse regions. The developed technique will facilitate better monitoring and understanding of the impact of environmental change and climate extremes on global freshwater ecosystems. ©2019. American Geophysical Union. All Rights Reserved." "57192180109;35782476600;56250119900;7004587644;57211565887;55796882100;7103016965;24168416900;","Cluster-Based Evaluation of Model Compensating Errors: A Case Study of Cloud Radiative Effect in the Southern Ocean",2019,"10.1029/2018GL081686","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063565470&doi=10.1029%2f2018GL081686&partnerID=40&md5=480a5ba6877efecf090e9658dd1789b9","Model evaluation is difficult and generally relies on analysis that can mask compensating errors. This paper defines new metrics, using clusters generated from a machine learning algorithm, to estimate mean and compensating errors in different model runs. As a test case, we investigate the Southern Ocean shortwave radiative bias using clusters derived by applying self-organizing maps to satellite data. In particular, the effects of changing cloud phase parameterizations in the MetOffice Unified Model are examined. Differences in cluster properties show that the regional radiative biases are substantially different than the global bias, with two distinct regions identified within the Southern Ocean, each with a different signed bias. Changing cloud phase parameterizations can reduce errors at higher latitudes but increase errors at lower latitudes of the Southern Ocean. Ranking the parameterizations often shows a contrast in mean and compensating errors, notably in all cases large compensating errors remain. ©2019. American Geophysical Union. All Rights Reserved." "57193628652;57205738039;55588066400;57205741483;57195134622;9132950200;35572232000;","Skluma: An extensible metadata extraction pipeline for disorganized data",2018,"10.1109/eScience.2018.00040","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85061386667&doi=10.1109%2feScience.2018.00040&partnerID=40&md5=6bc987313b669fddcb7d53fed6e8c8c1","To mitigate the effects of high-velocity data expansion and to automate the organization of filesystems and data repositories, we have developed Skluma-a system that automatically processes a target filesystem or repository, extracts content-and context-based metadata, and organizes extracted metadata for subsequent use. Skluma is able to extract diverse metadata, including aggregate values derived from embedded structured data; named entities and latent topics buried within free-text documents; and content encoded in images. Skluma implements an overarching probabilistic pipeline to extract increasingly specific metadata from files. It applies machine learning methods to determine file types, dynamically prioritizes and then executes a suite of metadata extractors, and explores contextual metadata based on relationships among files. The derived metadata, represented in JSON, describes probabilistic knowledge of each file that may be subsequently used for discovery or organization. Skluma's architecture enables it to be deployed both locally and used as an on-demand, cloud-hosted service to create and execute dynamic extraction workflows on massive numbers of files. It is modular and extensible-allowing users to contribute their own specialized metadata extractors. Thus far we have tested Skluma on local filesystems, remote FTP-accessible servers, and publicly-accessible Globus endpoints. We have demonstrated its efficacy by applying it to a scientific environmental data repository of more than 500,000 files. We show that we can extract metadata from those files with modest cloud costs in a few hours. © 2018 IEEE." "57195480875;55211613900;8265216600;16642111900;","Identifying tree-related microhabitats in TLS point clouds using machine learning",2018,"10.3390/rs10111735","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057071462&doi=10.3390%2frs10111735&partnerID=40&md5=7da76c5fc4ccb2545d5bb68c87a7a84f","Tree-related microhabitats (TreMs) play an important role in maintaining forest biodiversity and have recently received more attention in ecosystem conservation, forest management and research. However, TreMs have until now only been assessed by experts during field surveys, which are time-consuming and difficult to reproduce. In this study, we evaluate the potential of close-range terrestrial laser scanning (TLS) for semi-automated identification of different TreMs (bark, bark pockets, cavities, fungi, ivy and mosses) in dense TLS point clouds using machine learning algorithms, including deep learning. To classify the TreMs, we applied: (1) the Random Forest (RF) classifier, incorporating frequently used local geometric features and two additional self-developed orientation features, and (2) a deep Convolutional Neural Network (CNN) trained using rasterized multiview orthographic projections (MVOPs) containing top view, front view and side view of the point's local 3D neighborhood. The results confirmed that using local geometric features is beneficial for identifying the six groups of TreMs in dense tree-stem point clouds, but the rasterized MVOPs are even more suitable. Whereas the overall accuracy of the RF was 70%, that of the deep CNN was substantially higher (83%). This study reveals that close-range TLS is promising for the semi-automated identification of TreMs for forest monitoring purposes, in particular when applying deep learning techniques. © 2018 by the authors." "57192947904;6603354695;","Convolutional neural networks for cloud screening: Transfer learning from landsat-8 to ProBA-V",2018,"10.1109/IGARSS.2018.8517975","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063151514&doi=10.1109%2fIGARSS.2018.8517975&partnerID=40&md5=55ad8b05e03cb123890e2fd4f8f44197","Cloud detection is a key issue for exploiting the information from Earth observation satellites multispectral sensors. For Proba-V, cloud detection is challenging due to the limited number of spectral bands. Advanced machine learning methods, such as convolutional neural networks (CNN), have shown to work well on this problem provided enough labeled data. However, simultaneous collocated information about the presence of clouds is usually not available or requires a great amount of manual labor. In this work, we propose to learn from the available Landsat-8 cloud masks datasets and transfer this learning to solve the Proba-V cloud detection problem. CNN are trained with Landsat images adapted to resemble Proba-V characteristics and tested on a large set of real Proba-V scenes. Developed models outperform current operational Proba-V cloud detection without being trained with any real Proba-V data. Moreover, cloud detection accuracy can be further increased if the CNN are fine-tuned using a limited amount of Proba-V data. © 2018 IEEE" "57207877170;6507657461;57207881223;","Water Across Synthetic Aperture Radar Data (WASARD): SAR water body classification for the open data cube",2018,"10.1109/IGARSS.2018.8517447","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063145692&doi=10.1109%2fIGARSS.2018.8517447&partnerID=40&md5=1836e517da4381973712508fedd8d09a","The detection of inland water bodies from Synthetic Aperture Radar (SAR) data provides a great advantage over water detection with optical data, since SAR imaging is not impeded by cloud cover. Traditional methods of detecting water from SAR data involves using thresholding methods that can be labor intensive and imprecise. This paper describes Water Across Synthetic Aperture Radar Data (WASARD): a method of water detection from SAR data which automates and simplifies the thresholding process using machine learning on training data created from Geoscience Australia's WOFS algorithm. Of the machine learning models tested, the Linear Support Vector Machine was determined to be optimal, with the option of training using solely the VH polarization or a combination of the VH and VV polarizations. WASARD was able to identify water in the target area with a correlation of 97% with WOFS. © 2018 IEEE" "57192986457;57197822561;57201670483;","Templet Web: The use of volunteer computing approach in PaaS-style cloud",2018,"10.1515/eng-2018-0007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045682229&doi=10.1515%2feng-2018-0007&partnerID=40&md5=83aed994fa9177ee3ec2e2c92632b4e8","This article presents the Templet Web cloud service. The service is designed for high-performance scientific computing automation. The use of high-performance technology is specifically required by new fields of computational science such as data mining, artificial intelligence, machine learning, and others. Cloud technologies provide a significant cost reduction for high-performance scientific applications. The main objectives to achieve this cost reduction in the Templet Web service design are: (a) the implementation of ""on-demand"" access; (b) source code deployment management; (c) high-performance computing programs development automation. The distinctive feature of the service is the approach mainly used in the field of volunteer computing, when a person who has access to a computer system delegates his access rights to the requesting user. We developed an access procedure, algorithms, and software for utilization of free computational resources of the academic cluster system in line with the methods of volunteer computing. The Templet Web service has been in operation for five years. It has been successfully used for conducting laboratory workshops and solving research problems, some of which are considered in this article. The article also provides an overview of research directions related to service development. © 2018 Sergei Vostokin et al." "14028131100;57191840076;55363497200;57205428127;57189308892;55369740900;56096540000;","Augmented AI solutions for heavy oil reservoirs: Innovative workflows that build from smart analytics, machine learning and expert-based systems",2018,"10.2118/193650-MS","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059990586&doi=10.2118%2f193650-MS&partnerID=40&md5=66af2aa5514b0135a26777f8db8ac69c","Many heavy oil fields have recently seen exponentially higher volumes of data being made available from omnipresent connectivity. Existing data platforms have traditionally focused on solving the problem of data storage and access. The more complex problem of true knowledge discovery and systematic value creation from all this massive amount of data is less frequently addressed. There is a new, novel workflow that tackles the problem of building intelligent data analytics in heavy oil fields. This multidisciplinary solution builds from Augmented AI, i.e., a combination of artificial intelligence/machine learning and domain knowledge or expertise. Our solution has been proven effective in a Latin American heavy oil field and is presented here. We built an integrated view of the reservoir through a series of smart metrics and KPIs. This was accomplished by integrating expert-based knowledge with the analysis of geological data, reservoir behavior and production and operational performance. Our analytics-based solutions were designed from the perspective of reservoir management, and consequently, they could assimilate production and cost/economic analysis with geological information (e.g., well logs and/or existing geological models) and reservoir performance (metrics for pressure, voidage, fractional flow, reservoir contact, etc.). From here, KROs (key recovery obstacles) were identified for this heavy oil field, and robust field development opportunities (i.e., behind-pipe opportunities and/or new well targets) were methodically proposed based on an innovative saturation mapping approach. Our solution exploited existing cloud economics at scale, sustained advances in hardware capabilities (including GPUs running machine learning workloads) and iterative improvements in algorithmic learning techniques. Ultimately, we will demonstrate a smart analytics-based solution that can be systematically implemented in heavy oil reservoirs for robust diagnostics and comprehensive field development opportunity identification. The proposed methodology intelligently combines both field experience with data-driven and machine learning methods. Copyright 2018, Society of Petroleum Engineers." "56068624000;57200294912;56188627800;14625770800;54402367600;","Cross-domain ground-based cloud classification based on transfer of local features and discriminative metric learning",2018,"10.3390/rs10010008","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85040689691&doi=10.3390%2frs10010008&partnerID=40&md5=c3480efa7895bc8d67e3d170e4d027e9","Cross-domain ground-based cloud classification is a challenging issue as the appearance of cloud images from different cloud databases possesses extreme variations. Two fundamental problems which are essential for cross-domain ground-based cloud classification are feature representation and similarity measurement. In this paper, we propose an effective feature representation called transfer of local features (TLF), and measurement method called discriminative metric learning (DML). The TLF is a generalized representation framework that can integrate various kinds of local features, e.g., local binary patterns (LBP), local ternary patterns (LTP) and completed LBP (CLBP). In order to handle domain shift, such as variations of illumination, image resolution, capturing location, occlusion and so on, the TLF mines the maximum response in regions to make a stable representation for domain variations. We also propose to learn a discriminant metric, simultaneously. We make use of sample pairs and the relationship among cloud classes to learn the distancemetric. Furthermore, in order to improve the practicability of the proposedmethod,we replace the original cloud images with the convolutional activation maps which are then applied to TLF and DML. The proposed method has been validated on three cloud databases which are collected in China alone, provided by Chinese Academy ofMeteorological Sciences (CAMS),Meteorological Observation Centre (MOC), and Institute of Atmospheric Physics (IAP). The classification accuracies outperform the state-of-the-art methods." "57201861483;25522330800;","Detecting and tracking mesoscale precipitating objects using machine learning algorithms",2017,"10.1080/01431161.2017.1323280","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046261988&doi=10.1080%2f01431161.2017.1323280&partnerID=40&md5=69b44601b973df0c0d3b17ac6c33cdf6","Accurate identification of precipitating clouds is a challenging task. In the present work, Support Vector Machines (SVMs), Decision Trees (DT), and Random Forests (RD) algorithms were applied to extract and track mesoscale convective precipitating clouds from a series of 22 Geostationary Operational Environmental Satellite-13 meteorological image sub-scenes over the continental territory of Colombia. This study’s aims are twofold: (i) to establish whether the use of five meteorological spectral channels, rather than a single infrared (IR) channel, improves rainfall objects detection and (ii) to evaluate the potential of machine learning algorithms to locate precipitation clouds. Results show that while the SVM algorithm provides more accurate classification of rainfall cloud objects than the traditional IR brightness temperature threshold method, such improvement is not statistically significant. Accuracy assessment was performed using STEP (shape (S), theme (T), edge (E), and position (P)) object-based similarity matrix method, taking as reference precipitation satellite images from the Tropical Rainfall Measuring Mission. Best thematic and geometric accuracies were obtained applying the SVM algorithm. © 2017 Informa UK Limited, trading as Taylor & Francis Group." "56082380600;23035058500;57203999691;","OBJECT CLASSIFICATION VIA PLANAR ABSTRACTION",2016,"10.5194/isprs-annals-III-3-225-2016","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85031042352&doi=10.5194%2fisprs-annals-III-3-225-2016&partnerID=40&md5=64c3decc38c25fc225f84e58aae1fe96","We present a supervised machine learning approach for classification of objects from sampled point data. The main idea consists in first abstracting the input object into planar parts at several scales, then discriminate between the different classes of objects solely through features derived from these planar shapes. Abstracting into planar shapes provides a means to both reduce the computational complexity and improve robustness to defects inherent to the acquisition process. Measuring statistical properties and relationships between planar shapes offers invariance to scale and orientation. A random forest is then used for solving the multiclass classification problem. We demonstrate the potential of our approach on a set of indoor objects from the Princeton shape benchmark and on objects acquired from indoor scenes and compare the performance of our method with other point-based shape descriptors." "46761518900;7004435336;","Autocalibration experiments using machine learning and high performance computing",2013,"10.1016/j.envsoft.2012.10.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84871812202&doi=10.1016%2fj.envsoft.2012.10.007&partnerID=40&md5=28daad949b1dc5bcf1e54f22bd1240e7","Using as example the Soil and Water Assessment Tool (SWAT) model and a Southern Ontario Canada watershed, we conduct a set of experiments on calibration using a manual approach, a parallelized version of the shuffled complex evolution (SCE), Generalized Likelihood Uncertainty Estimation (GLUE), Sequential Uncertainty Fitting (SUFI-2) and compare to a simple parallel search on a finite set of gridded input parameter values invoking the probably approximately correct (PAC) learning hypothesis. We derive an estimation of the error in fitting and a prior estimate of the probability of success, based on the PAC hypothesis. We conclude that from the equivalent effort expended on initial setup for the other named algorithms we can already find directly a good parameter set for calibration. We further note that, in this algorithm, simultaneous co-calibration of flow and chemistry (total nitrogen and total phosphorous) is more likely to produce acceptable results, as compared to flow first, even with a simple weighted multiobjective approach. This approach is especially suited to a parallel, distributed or cloud computational environment. © 2012 Elsevier Ltd." "57200370110;57204395597;56524539600;55558822000;7202446179;57216620868;57216618614;57216616900;","Monitoring inland water quality using remote sensing: potential and limitations of spectral indices, bio-optical simulations, machine learning, and cloud computing",2020,"10.1016/j.earscirev.2020.103187","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084066895&doi=10.1016%2fj.earscirev.2020.103187&partnerID=40&md5=d72af64c4db132aab8efb6d149002b8b","Given the recent advances in remote sensing analytics, cloud computing, and machine learning, it is imperative to evaluate capabilities of remote sensing for water quality monitoring in the context of water resources management and decision-making. The objectives of this review were to analyze recent advances in water quality remote sensing and determine limitations of current systems, estimation methods, and suggest future improvements. To that end, we collected over 200 sets of water quality data including blue-green algae phycocyanin (BGA-PC), chlorophyll-a (Chl-a), dissolved oxygen (DO), specific conductivity (SC), fluorescent dissolved organic matter (fDOM), turbidity, and pollution-sediments from 2016 to 2018. The water quality data, generated from laboratory analysis of grab samples and in-situ real-time monitoring sensors distributed in eight lakes and rivers in Midwestern United States, were paired with synchronous proximal spectra, tripod-mounted hyperspectral imagery, and satellite data. The results showed that both proximal and satellite-based sensors have great potential to provide accurate estimate of optically active parameters, and remote sensing of non-optically active parameters may be indirectly estimated but still remains a challenge. Data-driven empirical approaches, i.e., deep learning outperformed the other competing methods, providing promising possibility for operational use of remote sensing in water quality monitoring and decision-making. As the first-time review of deep neural networks for water quality estimation, the paper concludes that anomaly detection utilizing multi-sensor data fusion and virtual constellation in cloud-computing is the most promising means for predicting impending water pollution outbreaks such as algal blooms. © 2020 Elsevier B.V." "57190677799;57189582972;7003384607;7003978405;6701355494;","Employing a multi-input deep convolutional neural network to derive soil clay content from a synergy of multi-temporal optical and radar imagery data",2020,"10.3390/RS12091389","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085257255&doi=10.3390%2fRS12091389&partnerID=40&md5=96dcc24c44815be46f1ad324136dd55a","Earth observation (EO) has an immense potential as being an enabling tool formapping spatial characteristics of the topsoil layer. Recently, deep learning based algorithms and cloud computing infrastructure have become available with a great potential to revolutionize the processing of EO data. This paper aims to present a novel EO-based soil monitoring approach leveraging open-access Copernicus Sentinel data and Google Earth Engine platform. Building on key results from existing data mining approaches to extract bare soil reflectance values the current study delivers valuable insights on the synergistic use of open access optical and radar images. The proposed framework is driven by the need to eliminate the influence of ambient factors and evaluate the efficiency of a convolutional neural network (CNN) to effectively combine the complimentary information contained in the pool of both optical and radar spectral information and those form auxiliary geographical coordinates mainly for soil. We developed and calibrated our multi-input CNN model based on soil samples (calibration = 80% and validation 20%) of the LUCAS database and then applied this approach to predict soil clay content. A promising prediction performance (R2 = 0.60, ratio of performance to the interquartile range (RPIQ) = 2.02, n = 6136) was achieved by the inclusion of both types (synthetic aperture radar (SAR) and laboratory visible near infrared-short wave infrared (VNIR-SWIR) multispectral) of observations using the CNN model, demonstrating an improvement of more than 5.5% in RMSE using the multi-year median optical composite and current state-of-the-art non linear machine learning methods such as random forest (RF; R2 = 0.55, RPIQ = 1.91, n = 6136) and artificial neural network (ANN; R2 = 0.44, RPIQ = 1.71, n = 6136). Moreover, we examined post-hoc techniques to interpret the CNN model and thus acquire an understanding of the relationships between spectral information and the soil target identified by the model. Looking to the future, the proposed approach can be adopted on the forthcoming hyperspectral orbital sensors to expand the current capabilities of the EO component by estimating more soil attributes with higher predictive performance. © 2020 by the authors." "57200517230;9036557400;57216491769;36440631200;35099662000;37088140000;","Estimation of all-weather 1 km MODIS land surface temperature for humid summer days",2020,"10.3390/RS12091398","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085241151&doi=10.3390%2fRS12091398&partnerID=40&md5=42d48cff7a4514a7e15962642969870e","Land surface temperature (LST) is used as a critical indicator for various environmental issues because it links land surface fluxes with the surface atmosphere. Moderate-resolution imaging spectroradiometers (MODIS) 1 km LSTs have been widely utilized but have the serious limitation of not being provided under cloudy weather conditions. In this study, we propose two schemes to estimate all-weather 1 km Aqua MODIS daytime (1:30 p.m.) and nighttime (1:30 a.m.) LSTs in South Korea for humid summer days. Scheme 1 (S1) is a two-step approach that first estimates 10 km LSTs and then conducts the spatial downscaling of LSTs from 10 km to 1 km. Scheme 2 (S2), a one-step algorithm, directly estimates the 1 km all-weather LSTs. Eight advanced microwave scanning radiometer 2 (AMSR2) brightness temperatures, three MODIS-based annual cycle parameters, and six auxiliary variables were used for the LST estimation based on random forest machine learning. To confirm the effectiveness of each scheme, we have performed different validation experiments using clear-sky MODIS LSTs. Moreover, we have validated all-weather LSTs using bias-corrected LSTs from 10 in situ stations. In clear-sky daytime, the performance of S2 was better than S1. However, in cloudy sky daytime, S1 simulated low LSTs better than S2, with an average root mean squared error (RMSE) of 2.6 °C compared to an average RMSE of 3.8 °C over 10 stations. At nighttime, S1 and S2 demonstrated no significant difference in performance both under clear and cloudy sky conditions. When the two schemes were combined, the proposed all-weather LSTs resulted in an average R2 of 0.82 and 0.74 and with RMSE of 2.5 °C and 1.4 °C for daytime and nighttime, respectively, compared to the in situ data. This paper demonstrates the ability of the two different schemes to produce all-weather dynamic LSTs. The strategy proposed in this study can improve the applicability of LSTs in a variety of research and practical fields, particularly for areas that are very frequently covered with clouds. © 2020 by the authors." "57201243646;57203905850;57214207213;12240696000;57201809767;8359720900;","Thick cloud and cloud shadow removal in multitemporal imagery using progressively spatio-temporal patch group deep learning",2020,"10.1016/j.isprsjprs.2020.02.008","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85080092790&doi=10.1016%2fj.isprsjprs.2020.02.008&partnerID=40&md5=91e34739f4b7969a2973f3c955c31ee1","Thick cloud and its shadow severely reduce the data usability of optical satellite remote sensing data. Although many approaches have been presented for cloud and cloud shadow removal, most of these approaches are still inadequate in terms of dealing with the following three issues: (1) thick cloud cover with large-scale areas, (2) all the temporal images included cloud or shadow, and (3) deficient utilization of only single temporal images. A novel spatio-temporal patch group deep learning framework for gap-filling through multiple temporal cloudy images is proposed to overcome these issues. The global-local loss function is presented to optimize the training model through cloud-covered and free regions, considering both the global consistency and local particularity. In addition, weighted aggregation and progressive iteration are utilized for reconstructing the holistic results. A series of simulated and real experiments are then performed to validate the effectiveness of the proposed method. Especially on Sentinel-2 MSI and Landsat-8 OLI with single/multitemporal images, under small/large scale regions, respectively. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57200555819;57211430607;57126877800;56506111400;7005693588;","Directionally constrained fully convolutional neural network for airborne LiDAR point cloud classification",2020,"10.1016/j.isprsjprs.2020.02.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079552935&doi=10.1016%2fj.isprsjprs.2020.02.004&partnerID=40&md5=9e82c0e57ce7b67b4a9e2ac64403189c","Point cloud classification plays an important role in a wide range of airborne light detection and ranging (LiDAR) applications, such as topographic mapping, forest monitoring, power line detection, and road detection. However, due to the sensor noise, high redundancy, incompleteness, and complexity of airborne LiDAR systems, point cloud classification is challenging. Traditional point cloud classification methods mostly focus on the development of handcrafted point geometry features and employ machine learning-based classification models to conduct point classification. In recent years, the advances of deep learning models have caused researchers to shift their focus towards machine learning-based models, specifically deep neural networks, to classify airborne LiDAR point clouds. These learning-based methods start by transforming the unstructured 3D point sets to regular 2D representations, such as collections of feature images, and then employ a 2D CNN for point classification. Moreover, these methods usually need to calculate additional local geometry features, such as planarity, sphericity and roughness, to make use of the local structural information in the original 3D space. Nonetheless, the 3D to 2D conversion results in information loss. In this paper, we propose a directionally constrained fully convolutional neural network (D-FCN) that can take the original 3D coordinates and LiDAR intensity as input; thus, it can directly apply to unstructured 3D point clouds for semantic labeling. Specifically, we first introduce a novel directionally constrained point convolution (D-Conv) module to extract locally representative features of 3D point sets from the projected 2D receptive fields. To make full use of the orientation information of neighborhood points, the proposed D-Conv module performs convolution in an orientation-aware manner by using a directionally constrained nearest neighborhood search. Then, we design a multiscale fully convolutional neural network with downsampling and upsampling blocks to enable multiscale point feature learning. The proposed D-FCN model can therefore process input point cloud with arbitrary sizes and directly predict the semantic labels for all the input points in an end-to-end manner. Without involving additional geometry features as input, the proposed method demonstrates superior performance on the International Society for Photogrammetry and Remote Sensing (ISPRS) 3D labeling benchmark dataset. The results show that our model achieves a new state-of-the-art performance on powerline, car, and facade categories. Moreover, to demonstrate the generalization abilities of the proposed method, we conduct further experiments on the 2019 Data Fusion Contest Dataset. Our proposed method achieves superior performance than the comparing methods and accomplishes an overall accuracy of 95.6% and an average F1 score of 0.810. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57213143651;57208619728;56540687200;","Inland harmful cyanobacterial bloom prediction in the eutrophic Tri An Reservoir using satellite band ratio and machine learning approaches",2020,"10.1007/s11356-019-07519-3","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077529085&doi=10.1007%2fs11356-019-07519-3&partnerID=40&md5=fb48c3b4729270ad20187f2d581ae682","In recent years, Tri An, a drinking water reservoir for millions of people in southern Vietnam, has been affected by harmful cyanobacterial blooms (HCBs), raising concerns about public health. It is, therefore, crucial to gain insights into the outbreak mechanism of HCBs and understand the spatiotemporal variations of chlorophyll-a (Chl-a) in this highly turbid and productive water. This study aims to evaluate the predictable performance of both approaches using satellite band ratio and machine learning for Chl-a concentration retrieval—a proxy of HCBs. The monthly water quality samples collected from 2016 to 2018 and 23 cloud free Sentinel-2A/B scenes were used to develop Chl-a retrieval models. For the band ratio approach, a strong linear relationship with in situ Chl-a was found for two-band algorithm of Green-NIR. The band ratio-based model accounts for 72% of variation in Chl-a concentration from 2016 to 2018 datasets with an RMSE of 5.95 μg/L. For the machine learning approach, Gaussian process regression (GPR) yielded superior results for Chl-a prediction from water quality parameters with the values of 0.79 (R2) and 3.06 μg/L (RMSE). Among various climatic parameters, a high correlation (R2 = 0.54) between the monthly total precipitation and Chl-a concentration was found. Our analysis also found nitrogen-rich water and TSS in the rainy season as the driving factors of observed HCBs in the eutrophic Tri An Reservoir (TAR), which offer important solutions to the management of HCBs in the future. © 2020, Springer-Verlag GmbH Germany, part of Springer Nature." "56200753200;57202812543;57191994356;57111374800;57200599957;57200790647;56720614400;57190136115;","Satellite data cloud detection using deep learning supported by hyperspectral data",2020,"10.1080/01431161.2019.1667548","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074039119&doi=10.1080%2f01431161.2019.1667548&partnerID=40&md5=e58c7b5fb5002dfe88c746e7a770481a","Deep learning methods can play an important role in satellite data cloud detection. The number and quality of training samples directly affect the accuracy of cloud detection based on deep learning. Therefore, selecting a large number of representative and high-quality training samples is a key step in cloud detection based on deep learning. For different satellite data sources, choosing sufficient and high-quality training samples has become an important factor limiting the application of deep learning in cloud detection. This paper presents a fast method for obtaining high-quality learning samples, which can be used for cloud detection of different satellite data with deep learning methods. AVIRIS (Airborne Visible Infrared Imaging Spectrometer) data, which have 224 continuous bands in the spectral range from 400–2500 nm, are used to provide cloud detection samples for different types of satellite data. Through visual interpretation, a sufficient number of cloud and clear sky pixels are selected from the AVIRIS data to construct a hyperspectral data sample library, which is used to simulate different satellite data (such as data from the Moderate Resolution Imaging Spectroradiometer (MODIS) and Landsat Operational Land Imager (OLI) satellites) as training samples. This approach avoids selecting training samples for different satellite sensors. Based on the Keras deep learning framework platform, a backpropagation (BP) neural network is employed for cloud detection from Landsat 8 OLI, National Polar-orbiting Partnership (NPP) Visible Infrared Imaging Radiometer Suite (VIIRS) and Terra MODIS data. The results are compared with cloud coverage results interpreted via artificial vision. The results demonstrate that the algorithm achieves good cloud detection results for the above data, and the overall accuracy is greater than 90%. © 2019, © 2019 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group." "55671149700;54583393400;7005861597;57215417747;25923235500;57215429017;","Characterizing land surface phenology and exotic annual grasses in dryland ecosystems using landsat and sentinel-2 data in harmony",2020,"10.3390/rs12040725","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85080942751&doi=10.3390%2frs12040725&partnerID=40&md5=e3d307ffface93b629e23b3ab3242d6c","Invasive annual grasses, such as cheatgrass (Bromus tectorum L.), have proliferated in dryland ecosystems of the western United States, promoting increased fire activity and reduced biodiversity that can be detrimental to socio-environmental systems. Monitoring exotic annual grass cover and dynamics over large areas requires the use of remote sensing that can support early detection and rapid response initiatives. However, few studies have leveraged remote sensing technologies and computing frameworks capable of providing rangeland managers with maps of exotic annual grass cover at relatively high spatiotemporal resolutions and near real-time latencies. Here, we developed a system for automated mapping of invasive annual grass (%) cover using in situ observations, harmonized Landsat and Sentinel-2 (HLS) data, maps of biophysical variables, and machine learning techniques. A robust and automated cloud, cloud shadow, water, and snow/ice masking procedure (mean overall accuracy >81%) was implemented using time-series outlier detection and data mining techniques prior to spatiotemporal interpolation of HLS data via regression tree models (r = 0.94; mean absolute error (MAE) = 0.02). Weekly, cloud-free normalized difference vegetation index (NDVI) image composites (2016-2018) were used to construct a suite of spectral and phenological metrics (e.g., start and end of season dates), consistent with information derived from Moderate Resolution Image Spectroradiometer (MODIS) data. These metrics were incorporated into a data mining framework that accurately (r = 0.83; MAE = 11) modeled and mapped exotic annual grass (%) cover throughout dryland ecosystems in the western United States at a native, 30-m spatial resolution. Our results show that inclusion of weekly HLS time-series data and derived indicators improves our ability to map exotic annual grass cover, as compared to distribution models that use MODIS products or monthly, seasonal, or annual HLS composites as primary inputs. This research fills a critical gap in our ability to effectively assess, manage, and monitor drylands by providing a framework that allows for an accurate and timely depiction of land surface phenology and exotic annual grass cover at spatial and temporal resolutions that are meaningful to local resource managers. © 2020 by the author." "57192947904;34869963500;57203813986;6603354695;","Transferring deep learning models for cloud detection between Landsat-8 and Proba-V",2020,"10.1016/j.isprsjprs.2019.11.024","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076098751&doi=10.1016%2fj.isprsjprs.2019.11.024&partnerID=40&md5=13fd31e2fef2127da475b81ef5aa7163","Accurate cloud detection algorithms are mandatory to analyze the large streams of data coming from the different optical Earth observation satellites. Deep learning (DL) based cloud detection schemes provide very accurate cloud detection models. However, training these models for a given sensor requires large datasets of manually labeled samples, which are very costly or even impossible to create when the satellite has not been launched yet. In this work, we present an approach that exploits manually labeled datasets from one satellite to train deep learning models for cloud detection that can be applied (or transferred) to other satellites. We take into account the physical properties of the acquired signals and propose a simple transfer learning approach using Landsat-8 and Proba-V sensors, whose images have different but similar spatial and spectral characteristics. Three types of experiments are conducted to demonstrate that transfer learning can work in both directions: (a) from Landsat-8 to Proba-V, where we show that models trained only with Landsat-8 data produce cloud masks 5 points more accurate than the current operational Proba-V cloud masking method, (b) from Proba-V to Landsat-8, where models that use only Proba-V data for training have an accuracy similar to the operational FMask in the publicly available Biome dataset (87.79–89.77% vs 88.48%), and (c) jointly from Proba-V and Landsat-8 to Proba-V, where we demonstrate that using jointly both data sources the accuracy increases 1–10 points when few Proba-V labeled images are available. These results highlight that, taking advantage of existing publicly available cloud masking labeled datasets, we can create accurate deep learning based cloud detection models for new satellites, but without the burden of collecting and labeling a large dataset of images. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "56684747900;57204481038;57194457319;56898969100;57194760236;57191109163;57196238233;57190089926;57203115077;57200988425;57001818900;7003505161;","Google Earth Engine Cloud Computing Platform for Remote Sensing Big Data Applications: A Comprehensive Review",2020,"10.1109/JSTARS.2020.3021052","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092080246&doi=10.1109%2fJSTARS.2020.3021052&partnerID=40&md5=6454994d71eb94b2f6517da402fa09f1","Remote sensing (RS) systems have been collecting massive volumes of datasets for decades, managing and analyzing of which are not practical using common software packages and desktop computing resources. In this regard, Google has developed a cloud computing platform, called Google Earth Engine (GEE), to effectively address the challenges of big data analysis. In particular, this platform facilitates processing big geo data over large areas and monitoring the environment for long periods of time. Although this platform was launched in 2010 and has proved its high potential for different applications, it has not been fully investigated and utilized for RS applications until recent years. Therefore, this study aims to comprehensively explore different aspects of the GEE platform, including its datasets, functions, advantages/limitations, and various applications. For this purpose, 450 journal articles published in 150 journals between January 2010 and May 2020 were studied. It was observed that Landsat and Sentinel datasets were extensively utilized by GEE users. Moreover, supervised machine learning algorithms, such as Random Forest, were more widely applied to image classification tasks. GEE has also been employed in a broad range of applications, such as Land Cover/land Use classification, hydrology, urban planning, natural disaster, climate analyses, and image processing. It was generally observed that the number of GEE publications have significantly increased during the past few years, and it is expected that GEE will be utilized by more users from different fields to resolve their big data processing challenges. © 2008-2012 IEEE." "47461334700;57203927097;57186194700;24079495100;57203928635;57186534500;57201284843;36188365300;53663404300;57200596300;57203929375;55949126400;20734962900;7004185527;57211507212;54179582200;57115426300;6507582592;6506266992;6602713038;","Hyperspectral outcrop models for palaeoseismic studies",2019,"10.1111/phor.12300","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076888275&doi=10.1111%2fphor.12300&partnerID=40&md5=b11fb1dd15ada12ba98d4f8c7e66db5d","The traditional study of palaeoseismic trenches, involving logging, stratigraphic and structural interpretation, can be time consuming and affected by biases and inaccuracies. To overcome these limitations, a new workflow is presented that integrates infrared hyperspectral and photogrammetric data to support field-based palaeoseismic observations. As a case study, this method is applied on two palaeoseismic trenches excavated across a post-glacial fault scarp in northern Finnish Lapland. The hyperspectral imagery (HSI) is geometrically and radiometrically corrected, processed using established image processing algorithms and machine learning approaches, and co-registered to a structure-from-motion point cloud. HSI-enhanced virtual outcrop models are a useful complement to palaeoseismic field studies as they not only provide an intuitive visualisation of the outcrop and a versatile data archive, but also enable an unbiased assessment of the mineralogical composition of lithologic units and a semi-automatic delineation of contacts and deformational structures in a 3D virtual environment. © 2019 The Authors. The Photogrammetric Record © 2019 The Remote Sensing and Photogrammetry Society and John Wiley & Sons Ltd" "57196198872;55371193400;57212374172;35234980300;7202856872;","Next generation mapping: Combining deep learning, cloud computing, and big remote sensing data",2019,"10.3390/rs11232881","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076550652&doi=10.3390%2frs11232881&partnerID=40&md5=95b7e65d6b45d1317d23fe7e215d2a55","The rapid growth of satellites orbiting the planet is generating massive amounts of data for Earth science applications. Concurrently, state-of-the-art deep-learning-based algorithms and cloud computing infrastructure have become available with a great potential to revolutionize the image processing of satellite remote sensing. Within this context, this study evaluated, based on thousands of PlanetScope images obtained over a 12-month period, the performance of three machine learning approaches (random forest, long short-term memory-LSTM, and U-Net). We applied these approaches to mapped pasturelands in a Central Brazil region. The deep learning algorithms were implemented using TensorFlow, while the random forest utilized the Google Earth Engine platform. The accuracy assessment presented F1 scores for U-Net, LSTM, and random forest of, respectively, 96.94%, 98.83%, and 95.53% in the validation data, and 94.06%, 87.97%, and 82.57% in the test data, indicating a better classification efficiency using the deep learning approaches. Although the use of deep learning algorithms depends on a high investment in calibration samples and the generalization of these methods requires further investigations, our results suggest that the neural network architectures developed in this study can be used to map large geographic regions that consider a wide variety of satellite data (e.g., PlanetScope, Sentinel-2, Landsat-8). © 2019 by the authors." "56156401100;55967027800;34769730500;55901520500;","Regularization parameter selection in minimum volume hyperspectral unmixing",2019,"10.1109/TGRS.2019.2929776","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075667290&doi=10.1109%2fTGRS.2019.2929776&partnerID=40&md5=d41ae01079587d1e7705b4f56371d89d","Linear hyperspectral unmixing (HU) aims at factoring the observation matrix into an endmember matrix and an abundance matrix. Linear HU via variational minimum volume (MV) regularization has recently received considerable attention in the remote sensing and machine learning areas, mainly owing to its robustness against the absence of pure pixels. We put some popular linear HU formulations under a unifying framework, which involves a data-fitting term and an MV-based regularization term, and collectively solve it via a nonconvex optimization. As the former and the latter terms tend, respectively, to expand (reducing the data-fitting errors) and to shrink the simplex enclosing the measured spectra, it is critical to strike a balance between those two terms. To the best of our knowledge, the existing methods find such balance by tuning a regularization parameter manually, which has little value in unsupervised scenarios. In this paper, we aim at selecting the regularization parameter automatically by exploiting the fact that a too large parameter overshrinks the volume of the simplex defined by the endmembers, making many data points be left outside of the simplex and hence inducing a large data-fitting error, while a sufficiently small parameter yields a large simplex making data-fitting error very small. Roughly speaking, the transition point happens when the simplex still encloses the data cloud but there are data points on all its facets. These observations are systematically formulated to find the transition point that, in turn, yields a good parameter. The competitiveness of the proposed selection criterion is illustrated with simulated and real data. © 1980-2012 IEEE." "57202621367;57189296641;55946208600;7103129103;11839001400;35319628300;36018685200;","Daytime rainy cloud detection and convective precipitation delineation based on a deep neural network method using GOES-16 ABI images",2019,"10.3390/rs11212555","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074652871&doi=10.3390%2frs11212555&partnerID=40&md5=a87d5d6056d68cf9edfdb80e10ac6fd2","Precipitation, especially convective precipitation, is highly associated with hydrological disasters (e.g., floods and drought) that have negative impacts on agricultural productivity, society, and the environment. To mitigate these negative impacts, it is crucial to monitor the precipitation status in real time. The new Advanced Baseline Imager (ABI) onboard the GOES-16 satellite provides such a precipitation product in higher spatiotemporal and spectral resolutions, especially during the daytime. This research proposes a deep neural network (DNN) method to classify rainy and non-rainy clouds based on the brightness temperature differences (BTDs) and reflectances (Ref) derived from ABI. Convective and stratiform rain clouds are also separated using similar spectral parameters expressing the characteristics of cloud properties. The precipitation events used for training and validation are obtained from the IMERG V05B data, covering the southeastern coast of the U.S. during the 2018 rainy season. The performance of the proposed method is compared with traditional machine learning methods, including support vector machines (SVMs) and random forest (RF). For rainy area detection, the DNN method outperformed the other methods, with a critical success index (CSI) of 0.71 and a probability of detection (POD) of 0.86. For convective precipitation delineation, the DNN models also show a better performance, with a CSI of 0.58 and POD of 0.72. This automatic cloud classification system could be deployed for extreme rainfall event detection, real-time forecasting, and decision-making support in rainfall-related disasters. © 2019 by the authors." "57200918295;24072139200;57203833103;56545781200;57203501070;8632957700;57208015606;","Mapping the spatial distribution and changes of oil palm land cover using an open source cloud-based mapping platform",2019,"10.1080/01431161.2019.1597311","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063595159&doi=10.1080%2f01431161.2019.1597311&partnerID=40&md5=3f50cc89b902b760f75dc2aa8469394d","Oil palm has become well known for its oil palm yields that can be used to produce food, biodiesel and biogas. The rapid expansion of oil palm plantations over large areas has changed the land use and land cover of surroundings. Changes in land covers can be mapped and later used for further analysis. However, obtaining and classifying large coverages require massive amounts of data and computing resources and the skills and time of analysts. The Remote Ecosystem Monitoring Assessment Pipeline (REMAP) provides a cloud computing platform that hosts an open-source stacked Landsat data that allows land cover classification to be implemented using a built-in random forest supervised machine learning algorithm. Classifications were performed with the aid of predictor layers to discriminate the following land covers in Peninsular Malaysia: oil palm, built-up, bare soil, water, forest, other vegetation and paddy. The classification performed on period 1 (1999–2003) and period 2 (2014–2017) data produced an overall accuracy of 80.34% and 79.53% respectively. The analysis of the changes in oil palm distributions from period 1 to period 2 indicated an increment of 23.59%. Further analysis revealed that oil palm expansion in Peninsular Malaysia only minimally affected forested area and is mostly resulted from the conversion of less productive crops to oil palm. Results prove the land cover mapping and change detection capabilities of REMAP as a cloud computing platform for large areas. Despite its limitations, REMAP has the potential to achieve fast-paced mapping over large areas and monitor land changes in oil palm distributions. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57212501858;55843662000;6603793565;55472628200;","A decision tree approach for spatially interpolating missing land cover data and classifying satellite images",2019,"10.3390/rs11151796","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070438163&doi=10.3390%2frs11151796&partnerID=40&md5=c93f29e52b5039dcc14f2d9fa74c25d2","Sustainable Development Goals (SDGs) are a set of priorities the United Nations and World Bank have set for countries to reach in order to improve quality of life and environment globally by 2030. Free satellite images have been identified as a key resource that can be used to produce official statistics and analysis to measure progress towards SDGs, especially those that are concerned with the physical environment, such as forest, water, and crops. Satellite images can often be unusable due to missing data from cloud cover, particularly in tropical areas where the deforestation rates are high. There are existing methods for filling in image gaps; however, these are often computationally expensive in image classification or not effective at pixel scale. To address this, we use two machine learning methods-gradient boosted machine and random forest algorithms-to classify the observed and simulated 'missing' pixels in satellite images as either grassland or woodland. We also predict a continuous biophysical variable, Foliage Projective Cover (FPC), which was derived from satellite images, and perform accurate binary classification and prediction using only the latitude and longitude of the pixels. We compare the performance of these methods against each other and inverse distance weighted interpolation, which is a well-established spatial interpolation method. We find both of the machine learning methods, particularly random forest, perform fast and accurate classifications of both observed and missing pixels, with up to 0.90 accuracy for the binary classification of pixels as grassland or woodland. The results show that the random forest method is more accurate than inverse distance weighted interpolation and gradient boosted machine for prediction of FPC for observed and missing data. Based on the case study results from a sub-tropical site in Australia, we show that our approach provides an efficient alternative for interpolating images and performing land cover classifications. © 2019 by the authors." "57197529095;57211270228;57211274930;57211276340;35316821200;","Stock price prediction using news sentiment analysis",2019,"10.1109/BigDataService.2019.00035","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073210332&doi=10.1109%2fBigDataService.2019.00035&partnerID=40&md5=6ad1f636f6301adc0320d84e258686ff","Predicting stock market prices has been a topic of interest among both analysts and researchers for a long time. Stock prices are hard to predict because of their high volatile nature which depends on diverse political and economic factors, change of leadership, investor sentiment, and many other factors. Predicting stock prices based on either historical data or textual information alone has proven to be insufficient. Existing studies in sentiment analysis have found that there is a strong correlation between the movement of stock prices and the publication of news articles. Several sentiment analysis studies have been attempted at various levels using algorithms such as support vector machines, naive Bayes regression, and deep learning. The accuracy of deep learning algorithms depends upon the amount of training data provided. However, the amount of textual data collected and analyzed during the past studies has been insufficient and thus has resulted in predictions with low accuracy. In our paper, we improve the accuracy of stock price predictions by gathering a large amount of time series data and analyzing it in relation to related news articles, using deep learning models. The dataset we have gathered includes daily stock prices for S&P500 companies for five years, along with more than 265,000 financial news articles related to these companies. Given the large size of the dataset, we use cloud computing as an invaluable resource for training prediction models and performing inference for a given stock in real time. © 2019 IEEE." "7403968786;57210824690;57209045965;7003917155;36018685200;57210718719;7403924488;","Land surface temperature derivation under all sky conditions through integrating AMSR-E/AMSR-2 and MODIS/GOES observations",2019,"10.3390/rs11141704","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071531519&doi=10.3390%2frs11141704&partnerID=40&md5=c98c7ac2d6cbe920a575e62e372e088e","Land surface temperature (LST) is an important input to the Atmosphere-Land Exchange Inverse (ALEXI) model to derive the Evaporative Stress Index (ESI) for drought monitoring. Currently, LST inputs to the ALEXI model come from the Geostationary Operational Environmental Satellite (GOES) and Moderate Resolution Imaging Spectroradiometer (MODIS) products, but clouds affect them. While passive microwave (e.g., AMSR-E and AMSR-2) sensors can penetrate non-rainy clouds and observe the Earth's surface, but usually with a coarse spatial resolution, how to utilize multiple instruments' advantages is an important methodology in remote sensing. In this study, we developed a new five-channel algorithm to derive LST from the microwave AMSR-E and AMSR-2 measurements and calibrate to the MODIS and GOES LST products. A machine learning method is implemented to further improve its performance. The MODIS and GOES LST products still show better performance than the AMSR-E and AMSR-2 LSTs when evaluated against the ground observations. Therefore, microwave LSTs are only used to fill the gaps due to clouds in the MODIS and GOES LST products. A gap filling method is further applied to fill the remaining gaps in the merged LSTs and downscale to the same spatial resolution as the MODIS and GOES products. With the daily integrated LST at the same spatial resolution as the MODIS and GOES products and available under nearly all sky conditions, the drought index, like the ESI, can be updated on daily basis. The initial implementation results demonstrate that the daily drought map can catch the fast changes of drought conditions and capture the signals of flash drought, and make flash drought monitoring become possible. It is expected that a drought map that is available on daily basis will benefit future drought monitoring. © 2019 by the authors." "57194833104;56482796700;56571063800;24402359000;7004469744;7003591311;","An emulator approach to stratocumulus susceptibility",2019,"10.5194/acp-19-10191-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070613374&doi=10.5194%2facp-19-10191-2019&partnerID=40&md5=b7c5839e25c4562e4ac633cd164b4e4c","The climatic relevance of aerosol-cloud interactions depends on the sensitivity of the radiative effect of clouds to cloud droplet number N, and liquid water path LWP. We derive the dependence of cloud fraction CF, cloud albedo AC, and the relative cloud radiative effect rCRE D CF • AC on N and LWP from 159 large-eddy simulations of nocturnal stratocumulus. These simulations vary in their initial conditions for temperature, moisture, boundary-layer height, and aerosol concentration but share boundary conditions for surface fluxes and subsidence. Our approach is based on Gaussian-process emulation, a statistical technique related to machine learning. We succeed in building emulators that accurately predict simulated values of CF, AC, and rCRE for given values of N and LWP. Emulator-derived susceptibilities @ lnrCRE=@ lnN and @ lnrCRE=@ lnLWP cover the nondrizzling, fully overcast regime as well as the drizzling regime with broken cloud cover. Theoretical results, which are limited to the nondrizzling regime, are reproduced. The susceptibility @ lnrCRE=@ lnN captures the strong sensitivity of the cloud radiative effect to cloud fraction, while the susceptibility @ lnrCRE=@ lnLWP describes the influence of cloud amount on cloud albedo irrespective of cloud fraction. Our emulation-based approach provides a powerful tool for summarizing complex data in a simple framework that captures the sensitivities of cloud-field properties over a wide range of states. © 2019 Author(s)." "52364782100;57188974100;56908481600;","An adaptive offloading method for an IoT-cloud converged virtual machine system using a hybrid deep neural network",2018,"10.3390/su10113955","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85055832565&doi=10.3390%2fsu10113955&partnerID=40&md5=419cebc1b3c4e99a096e75491f4941c1","A virtual machine with a conventional offloading scheme transmits and receives all context information to maintain program consistency during communication between local environments and the cloud server environment. Most overhead costs incurred during offloading are proportional to the size of the context information transmitted over the network. Therefore, the existing context information synchronization structure transmits context information that is not required for job execution when offloading, which increases the overhead costs of transmitting context information in low-performance Internet-of-Things (IoT) devices. In addition, the optimal offloading point should be determined by checking the server's CPU usage and network quality. In this study, we propose a context management method and estimation method for CPU load using a hybrid deep neural network on a cloud-based offloading service that extracts contexts that require synchronization through static profiling and estimation. The proposed adaptive offloading method reduces network communication overheads and determines the optimal offloading time for low-computing-powered IoT devices and variable server performance. Using experiments, we verify that the proposed learning-based prediction method effectively estimates the CPU load model for IoT devices and can adaptively apply offloading according to the load of the server. © 2018 by the authors." "7202257926;7003653764;57193851407;57154391900;8573340700;35612330300;","Variations in snow crystal riming and ZDR: A case analysis",2018,"10.1175/JAMC-D-17-0068.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85043995057&doi=10.1175%2fJAMC-D-17-0068.1&partnerID=40&md5=ab47bf07089a1084cefa286a652e7961","A case study in terms of variations in differential reflectivity ZDR observed at X band and snow crystal riming is presented for a light-snow event that occurred near Greeley, Colorado, on 26-27 November 2015. In the early portion of the event, ZDR values at near-surface levels were low (0-0.25 dB). During a second time period approximately 8 h later, ZDR values became distinctly positive (+2-3 dB). Digital photographs of the snow particles were obtained by a Multi-Angle Snowflake Camera (MASC) installed at a range of 13 km from the radar. Image-processing and machine-learning techniques applied to the MASC data showed that the snow particles were more heavily rimed during the low-ZDR time period. The aerodynamic effects of these rime deposits promoted a wider distribution of hydrometeor canting angles. The shift toward more random particle orientations underlies the observed reduction in ZDR during the period when more heavily rimed particles were observed in the MASC data. © 2018 American Meteorological Society." "50661784500;57195681934;57194448970;16302424800;6507832282;57203104770;7005794259;24385863600;6701762451;","A continuous flow diffusion chamber study of sea salt particles acting as cloud nuclei: deliquescence and ice nucleation",2018,"10.1080/16000889.2018.1463806","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046414211&doi=10.1080%2f16000889.2018.1463806&partnerID=40&md5=55b5307017446ee301835415d53a8b87","Phase changes of sea salt particles alter their physical and chemical properties, which is significant for Earth’s chemistry and energy budget. In this study, continuous flow diffusion chamber is used to investigate deliquescence, homogeneous and heterogeneous ice nucleation between 242 K and 215 K, of four salts: pure NaCl, pure MgCl2, synthetic sea water salt, and salt distilled from sampled sea water. Anhydrous particles, aqueous droplets and ice particles were discriminated using a polarisation-sensitive optical particle counter coupled with a machine learning analysis technique. The measured onset deliquescence relative humidities agree with previous studies, where sea water salts deliquescence at lower humidities than pure NaCl. Deliquesced salt droplets homogenously freeze when the relative humidity reaches a sufficiently high value at temperatures below 233 K. From 224 K and below, deposition nucleation freezing on a fraction of NaCl particles was observed at humidities lower than the deliquescence relative humidity. At these low temperatures, otherwise unactivated salt particles deliquesced at the expected deliquescence point, followed by homogeneous freezing at temperatures as low as 215 K. Thus, the observed sea salt particles exhibit a triad of temperature-dependent behaviours. First, they act as cloud condensation particles (CCNs) > 233 K, second they can be homogeneous freezing nuclei (HFNs) < 233 K and finally they act as ice nucleating particles (INPs) for heterogeneous nucleation <224 K. © 2018 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group." "57201840890;16235589700;","A technology map to facilitate the process of mine modernization throughout the mining cycle",2017,"10.17159/2411-9717/2017/v117n7a5","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85039843503&doi=10.17159%2f2411-9717%2f2017%2fv117n7a5&partnerID=40&md5=0af6c315a63c0cea66d048ddd49b4888","It is vital for organizations and individual operations to have access to a platform with technology-related information to consider for further research and development. This paper presents a technology map that was created with the purpose of facilitating mine modernization through technological advancement throughout the mining lifecycle/cycle. To achieve this, a platform was created to represent the mining lifecycle that incorporates each of the mining phases, i.e. exploration, project evaluation, mine design, operations, closure, and post-closure phases. The constituent value drivers for each phase were then investigated and included. These covered the various focus areas within the mining cycle, such as the applicable sub-phases, processes, systems, activities, or specific challenges, that impact a mine's operation. Technologies, both physical and digital, with the potential to add value to these focus areas were then incorporated into the platform to create a technology map. This potential to add value, if applied or modified for application, was assessed on any combination of five factors, namely the ability to increase production, increase productivity, increase efficiency, improve safety, or reduce the risk of human error. The primary focus was on technologies currently classified as disruptive and/or exponential, e.g. internet of things, cloud computing, advanced robotics, genomics, 3D printing, and artificial Intelligence. Other emerging technologies were also investigated, such as automation, machine learning, renewable energy generation, energy storage, advanced materials, and more. Furthermore, selected innovative technologies adapted for or developed in mining were also investigated, as well as other new technologies in non-mining industries with potential to add value to mining. As such, a technology map was created that covers the entire lifecycle of a mining venture, which highlights technologies with the potential to add value to specific focus areas. This technology map may be applied to facilitate advances in technology for mine modernization. © The Southern African Institute of Mining and Metallurgy, 2017." "57191419079;56702361600;36519976400;","Self-learning method for DDoS detection model in cloud computing",2017,"10.1109/EIConRus.2017.7910612","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85019484776&doi=10.1109%2fEIConRus.2017.7910612&partnerID=40&md5=76a64c72b33fd0057d39aafc4bf77a53","Cloud Computing has many significant benefits like the provision of computing resources and virtual networks on demand. However, there is the problem to assure the security of these networks against Distributed Denial-of-Service (DDoS) attack. Over the past few decades, the development of protection method based on data mining has attracted many researchers because of its effectiveness and practical significance. Most commonly these detection methods use prelearned models or models based on rules. Because of this the proposed DDoS detection methods often failure in dynamically changing cloud virtual networks. In this paper, we purposed self-learning method allows to adapt a detection model to network changes. This is minimized the false detection and reduce the possibility to mark legitimate users as malicious and vice versa. The developed method consists of two steps: collecting data about the network traffic by Netflow protocol and relearning the detection model with the new data. During the data collection we separate the traffic on legitimate and malicious. The separated traffic is labeled and sent to the relearning pool. The detection model is relearned by a data from the pool of current traffic. The experiment results show that proposed method could increase efficiency of DDoS detection systems is using data mining. © 2017 IEEE." "6602728119;","Generating topographic map data from classification results",2017,"10.3390/rs9030224","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85019351640&doi=10.3390%2frs9030224&partnerID=40&md5=439b28f8816960c1e4299280de0bc86c","The use of classification results as topographicmap data requires cartographic enhancement and checking of the geometric accuracy. Urban areas are of special interest. The conversion of the classification result into topographic map data of high thematic and geometric quality is subject of this contribution. After reviewing the existing literature on this topic, a methodology is presented. The extraction of point clouds belonging to line segments is solved by the Hough transform. The mathematics for deriving polygons of orthogonal, parallel and general line segments by least squares adjustment is presented. A unique solution for polylines, where the Hough parameters are optimized, is also given. By means of two data sets land cover maps of six classes were produced and then enhanced by the proposed method. The classification used the decision tree method applying a variety of attributes including object heights derived from imagery. The cartographic enhancement is carried out with two different levels of quality. The user's accuracies for the classes ""impervious surface"" and ""building"" were above 85% in the ""Level 1"" map of Example 1. The geometric accuracy of building corners at the ""Level 2"" maps is assessed bymeans of reference data derived fromortho-images. The obtained rootmean square errors (RMSE) of the generated coordinates (x, y) were RMSEx = 1.2 m and RMSEy = 0.7 m (Example 1) and RMSEx = 0.8 m and RMSEy = 1.0 m (Example 2) using 31 and 62 check points, respectively. All processing for Level 1 (raster data) could be carried out with a high degree of automation. Level 2 maps (vector data) were compiled for the classes ""building"" and ""road and parking lot"". For urban areas with numerous classes and of large size, universal algorithms are necessary to produce vector data fully automatically. The recent progress in sensors and machine learning methods will support the generation of topographic map data of high thematic and geometric accuracy. © 2017 by the author." "57214252363;57191247319;56029052900;57071147200;7701309571;","Severe Thunderstorm Detection by Visual Learning Using Satellite Images",2017,"10.1109/TGRS.2016.2618929","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84995484497&doi=10.1109%2fTGRS.2016.2618929&partnerID=40&md5=2b36701a9e586f6ae521fdbf2c169b95","Computers are widely utilized in today's weather forecasting as a powerful tool to leverage an enormous amount of data. Yet, despite the availability of such data, current techniques often fall short of producing reliable detailed storm forecasts. Each year severe thunderstorms cause significant damage and loss of life, some of which could be avoided if better forecasts were available. We propose a computer algorithm that analyzes satellite images from historical archives to locate visual signatures of severe thunderstorms for short-term predictions. While computers are involved in weather forecasts to solve numerical models based on sensory data, they are less competent in forecasting based on visual patterns from both current and past satellite images. In our system, we extract and summarize important visual storm evidence from satellite image sequences in the way that meteorologists interpret the images. In particular, the algorithm extracts and fits local cloud motion from image sequences to model the storm-related cloud patches. Image data from the year 2008 have been adopted to train the model, and historical severe thunderstorm reports in continental U.S. from 2000 to 2013 have been used as the ground truth and priors in the modeling process. Experiments demonstrate the usefulness and potential of the algorithm for producing more accurate severe thunderstorm forecasts. © 1980-2012 IEEE." "57144622400;57194639489;","An Elliptic Curve Based Schnorr Cloud Security Model in Distributed Environment",2016,"10.1155/2016/4913015","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84959282600&doi=10.1155%2f2016%2f4913015&partnerID=40&md5=d8ebd586429a0db84f86907baa26ee6d","Cloud computing requires the security upgrade in data transmission approaches. In general, key-based encryption/decryption (symmetric and asymmetric) mechanisms ensure the secure data transfer between the devices. The symmetric key mechanisms (pseudorandom function) provide minimum protection level compared to asymmetric key (RSA, AES, and ECC) schemes. The presence of expired content and the irrelevant resources cause unauthorized data access adversely. This paper investigates how the integrity and secure data transfer are improved based on the Elliptic Curve based Schnorr scheme. This paper proposes a virtual machine based cloud model with Hybrid Cloud Security Algorithm (HCSA) to remove the expired content. The HCSA-based auditing improves the malicious activity prediction during the data transfer. The duplication in the cloud server degrades the performance of EC-Schnorr based encryption schemes. This paper utilizes the blooming filter concept to avoid the cloud server duplication. The combination of EC-Schnorr and blooming filter efficiently improves the security performance. The comparative analysis between proposed HCSA and the existing Distributed Hash Table (DHT) regarding execution time, computational overhead, and auditing time with auditing requests and servers confirms the effectiveness of HCSA in the cloud security model creation. Copyright © 2016 Vinothkumar Muthurajan and Balaji Narayanasamy." "7004680232;7004315232;57200066316;36724662900;","Adaptive sky: A Feature correspondence toolbox for a multi-instrument, multi-platform distributed cloud monitoring sensor web",2008,"10.1109/AERO.2008.4526451","https://www.scopus.com/inward/record.uri?eid=2-s2.0-49349092444&doi=10.1109%2fAERO.2008.4526451&partnerID=40&md5=4e0f6999b102fc08b01d08821357e784","The current suite of spaceborne and in-situ assets, including those deployed by NASA, NOAA, and other groups, provides distributed sensing of the Earth's atmosphere, oceans, and land masses. As part of an activity supported through NASA's Earth Science Technology Office (ESTO), we have developed techniques that enable such assets to be dynamically combined to form sensor webs that can respond quickly to short-lived events and provide rich multi-modal observations of objects, such as clouds, that are evolving in space and time. A key focus of this work involves relating the observations made by one instrument to the observations made by another instrument. We have applied approaches derived from data mining, computer vision, and machine learning to automatically establish correspondence between different sets of observations. We will describe a number of Earth science scenarios that were used to direct this development and which have benefited from the approach. ©2008 IEEE." "57207951736;7402291608;57208213994;57204638211;","Data driven battery modeling and management method with aging phenomenon considered",2020,"10.1016/j.apenergy.2020.115340","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086585158&doi=10.1016%2fj.apenergy.2020.115340&partnerID=40&md5=526e6df83c53443e1e0f8c78c0104bba","The battery is one of the most important parts of electric vehicles (EVs), and the establishment of an accurate battery state estimation model is of great significance to improve the management strategy of EVs. However, the battery degrades with the operation of EVs, which brings great difficulties for the battery modeling issue. This paper proposes a novel aging phenomenon considered vehicle battery modeling method by utilizing the cloud battery data. First of all, based on the Rain-flow cycle counting (RCC) algorithm, a battery aging trajectory extraction method is developed to quantify the battery degradation phenomenon and generate the aging index for the cloud battery data. Then, the deep learning algorithm is employed to mine the aging features of the battery, and based on the mined aging features, an aging phenomenon considered battery model is established. The actual operation data of electric buses in Zhengzhou is used to validate the practical performance of the proposed methodologies. The results show that the proposed modeling method can simulate the characteristic of the battery accurately. The terminal voltage and SoC estimation error can be limited within 2.17% and 1.08%, respectively. © 2020" "57212480538;56984291600;57203119863;","Seeing through the Clouds with DeepWaterMap",2020,"10.1109/LGRS.2019.2953261","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076914488&doi=10.1109%2fLGRS.2019.2953261&partnerID=40&md5=4ca204e5842cfee24e52cd873e750787","We present our next-generation surface water mapping model, DeepWaterMapV2, which uses improved model architecture, data set, and a training setup to create surface water maps at lower cost, with higher precision and recall. We designed DeepWaterMapV2 to be memory efficient for large inputs. Unlike earlier models, our new model is able to process a full Landsat scene in one-shot and without dividing the input into tiles. DeepWaterMapV2 is robust against a variety of natural and artificial perturbations in the input, such as noise, different sensor characteristics, and small clouds. Our model can even 'see' through the clouds without relying on any active sensor data, in cases where the clouds do not fully obstruct the scene. Although we trained the model on Landsat-8 images only, it also supports data from a variety of other Earth observing satellites, including Landsat-5, Landsat-7, and Sentinel-2, without any further training or calibration. © 2004-2012 IEEE." "57218566813;7201639219;55419832300;55420166300;8712490600;8419904800;","Machine learning for cloud detection of globally distributed sentinel-2 images",2020,"10.3390/RS12152355","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089604142&doi=10.3390%2fRS12152355&partnerID=40&md5=2d88cb68a2d1852b2f0ddffb0aff6374","In recent years, a number of different procedures have been proposed for segmentation of remote sensing images, basing on spectral information. Model-based and machine learning strategies have been investigated in several studies. This work presents a comprehensive overview and an unbiased comparison of the most adopted segmentation strategies: Support Vector Machines (SVM), Random Forests, Neural networks, Sen2Cor, FMask and MAJA. We used a training set for learning and two different independent sets for testing. The comparison accounted for 135 images acquired from 54 different worldwide sites. We observed that machine learning segmentations are extremely reliable when the training and test are homogeneous. SVM performed slightly better than other methods. In particular, when using heterogeneous test data, SVM remained the most accurate segmentation method while state-of-the-art model-based methods such as MAJA and FMask obtained better sensitivity and precision, respectively. Therefore, even if each method has its specific advantages and drawbacks, SVM resulted in a competitive option for remote sensing applications. © 2020 by the authors." "57191962971;6602413782;57194545411;","Point cloud vs. mesh features for building interior classification",2020,"10.3390/rs12142224","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088640136&doi=10.3390%2frs12142224&partnerID=40&md5=422ab1dd3eab3ba062ea1ec097f2133f","Interpreting 3D point cloud data of the interior and exterior of buildings is essential for automated navigation, interaction and 3D reconstruction. However, the direct exploitation of the geometry is challenging due to inherent obstacles such as noise, occlusions, sparsity or variance in the density. Alternatively, 3D mesh geometries derived from point clouds benefit from preprocessing routines that can surmount these obstacles and potentially result in more refined geometry and topology descriptions. In this article, we provide a rigorous comparison of both geometries for scene interpretation. We present an empirical study on the suitability of both geometries for the feature extraction and classification. More specifically, we study the impact for the retrieval of structural building components in a realistic environment which is a major endeavor in Building Information Modeling (BIM) reconstruction. The study runs on segment-based structuration of both geometries and shows that both achieve recognition rates over 75% F1 score when suitable features are used. © 2020 by the authors." "57211816997;6602927210;","Artificial neural networks for cloud masking of Sentinel-2 ocean images with noise and sunglint",2020,"10.1080/01431161.2020.1714776","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079105064&doi=10.1080%2f01431161.2020.1714776&partnerID=40&md5=3c16c9eb984cf5d4a226a7dfda0ca552","Cloudy regions in optical satellite images prevent the extraction of valuable information by image processing techniques. Several threshold, multi-temporal and machine learning approaches have been developed for the separation of clouds in land and ocean applications, but this task still remains a challenge. Concerning deep water marine applications, the main difficulties are imposed in regions with high noise levels and sunglint. In this study, artificial neural networks (ANNs) with different configurations are evaluated for the detection of clouds in Sentinel-2 images depicting deep water regions with several noise levels. The ANNs are trained on a manual public dataset and on a manual dataset created for the needs of this study, which authors intend to make publicly available. Results are compared with the cloud masks produced by three state-of-the-art algorithms: Fmask, MAJA, and Sen2Cor. It was shown that the ANNs trained on the second dataset perform very favourably, in contrast to the ANNs trained on the first dataset that fails to adequately represent the spectra of the noisy Sentinel-2 images. This study further reinforces the value of the ‘cirrus’ band and indicates the bands that mitigate the influence of noisy spectra, by defining and examining an index that characterizes the importance of the bands according to the weights produced by the ANNs. Finally, the possibility of improving results by making predictions using the feature scaling parameters of the test set instead of those of the training set is also investigated in cases where the test set cannot be adequately represented by the training set. © 2020, © 2020 Informa UK Limited, trading as Taylor & Francis Group." "56609110900;22235255500;57210823248;54684821900;57208222251;57211500277;57208121325;","DATimeS: A machine learning time series GUI toolbox for gap-filling and vegetation phenology trends detection",2020,"10.1016/j.envsoft.2020.104666","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081030418&doi=10.1016%2fj.envsoft.2020.104666&partnerID=40&md5=e2fd8bd1c098b19972faf6a5b1c65a9f","Optical remotely sensed data are typically discontinuous, with missing values due to cloud cover. Consequently, gap-filling solutions are needed for accurate crop phenology characterization. The here presented Decomposition and Analysis of Time Series software (DATimeS) expands established time series interpolation methods with a diversity of advanced machine learning fitting algorithms (e.g., Gaussian Process Regression: GPR) particularly effective for the reconstruction of multiple-seasons vegetation temporal patterns. DATimeS is freely available as a powerful image time series software that generates cloud-free composite maps and captures seasonal vegetation dynamics from regular or irregular satellite time series. This work describes the main features of DATimeS, and provides a demonstration case using Sentinel-2 Leaf Area Index time series data over a Spanish site. GPR resulted as an optimum fitting algorithm with most accurate gap-filling performance and associated uncertainties. DATimeS further quantified LAI fluctuations among multiple crop seasons and provided phenological indicators for specific crop types. © 2020 Elsevier Ltd" "57201291990;56237880400;9036557400;23028717700;56493777900;9233714800;57188725354;57213269104;57213778602;55460555300;57202974918;36779146200;","Estimation of spatially continuous daytime particulate matter concentrations under all sky conditions through the synergistic use of satellite-based AOD and numerical models",2020,"10.1016/j.scitotenv.2020.136516","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077734964&doi=10.1016%2fj.scitotenv.2020.136516&partnerID=40&md5=237f2c2f2289af009ac537b228536772","Satellite-derived aerosol optical depth (AOD) products are one of main predictors to estimate ground-level particulate matter (PM10 and PM2.5) concentrations. Since AOD products, however, are only provided under high-quality conditions, missing values usually exist in areas such as clouds, cloud shadows, and bright surfaces. In this study, spatially continuous AOD and subsequent PM10 and PM2.5 concentrations were estimated over East Asia using satellite- and model-based data and auxiliary data in a Random Forest (RF) approach. Data collected from the Geostationary Ocean Color Imager (GOCI; 8 times per day) in 2016 were used to develop AOD and PM models. Three schemes (i.e. G1, A1, and A2) were proposed for AOD modeling according to target AOD data (GOCI AOD and AERONET AOD) and the existence of satellite-derived AOD. The A2 scheme showed the best performance (validation R2 of 0.74 and prediction R2 of 0.73 when GOCI AOD did not exist) and the resultant AOD was used to estimate spatially continuous PM concentrations. The PM models with location information produced successful estimation results with R2 of 0.88 and 0.90, and rRMSE of 26.9 and 27.2% for PM10 and PM2.5, respectively. The spatial distribution maps of PM well captured the seasonal and spatial characteristics of PM reported in the literature, which implies the proposed approaches can be adopted for an operational estimation of spatially continuous AOD and PMs under all sky conditions. © 2020" "6603023560;57196733964;7801500046;23020325200;","Applying FP-ILM to the retrieval of geometry-dependent effective Lambertian equivalent reflectivity (GE-LER) daily maps from UVN satellite measurements",2020,"10.5194/amt-13-985-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081225999&doi=10.5194%2famt-13-985-2020&partnerID=40&md5=9cfc269b158f4375fdb44cdb14e3ef22","The retrieval of trace gas, cloud, and aerosol measurements from ultraviolet, visible, and near-infrared (UVN) sensors requires precise information on surface properties that are traditionally obtained from Lambertian equivalent reflectivity (LER) climatologies. The main drawbacks of using LER climatologies for new satellite missions are that (a) climatologies are typically based on previous missions with significantly lower spatial resolutions, (b) they usually do not account fully for satellite-viewing geometry dependencies characterized by bidirectional reflectance distribution function (BRDF) effects, and (c) climatologies may differ considerably from the actual surface conditions especially with snow/ice scenarios. In this paper we present a novel algorithm for the retrieval of geometry-dependent effective Lambertian equivalent reflectivity (GE-LER) from UVN sensors; the algorithm is based on the full-physics inverse learning machine (FP-ILM) retrieval. Radiances are simulated using a radiative transfer model that takes into account the satellite-viewing geometry, and the inverse problem is solved using machine learning techniques to obtain the GE-LER from satellite measurements. The GE-LER retrieval is optimized not only for trace gas retrievals employing the DOAS algorithm, but also for the large amount of data from existing and future atmospheric Sentinel satellite missions. The GE-LER can either be deployed directly for the computation of air mass factors (AMFs) using the effective scene approximation or it can be used to create a global gapless geometry-dependent LER (G3-LER) daily map from the GE-LER under clear-sky conditions for the computation of AMFs using the independent pixel approximation. The GE-LER algorithm is applied to measurements of TROPOMI launched in October 2017 on board the EU/ESA Sentinel-5 Precursor (S5P) mission. The TROPOMI GE-LER/G3-LER results are compared with climatological OMI and GOME-2 LER datasets and the advantages of using GE-LER/G3-LER are demonstrated for the retrieval of total ozone from TROPOMI. © Author(s) 2020." "57204696723;7801522301;57202074941;55949540900;57204685304;57214665542;57204524598;35559756000;9744075400;7004246857;","Toward Operational Mapping of Woody Canopy Cover in Tropical Savannas Using Google Earth Engine",2020,"10.3389/fenvs.2020.00004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079477946&doi=10.3389%2ffenvs.2020.00004&partnerID=40&md5=f86206e25bb9003548caa7a95b460304","Savanna woody plants can store significant amounts of carbon while also providing numerous other ecological and socio-economic benefits. However, they are significantly under-represented in widely used tree cover datasets, due to mapping challenges presented by their complex landscapes, and the underestimation of woody plants by methods that exclude short stature trees and shrubs. In this study, we describe a Google Earth Engine (GEE) application and present test case results for mapping percent woody canopy cover (%WCC) over a large savanna area. Relevant predictors of %WCC include information derived from radar backscatter (Sentinel-1) and optical reflectance (Sentinel-2), which are used in conjunction with plot level %WCC measurements to train and evaluate random forest models. We can predict %WCC at 40 m pixel resolution for the full extent of Senegal with a root mean square error of ∼8% (based on independent sample evaluation). Further examination of model results provides insights into method stability and potential generalizability. Annual median radar backscatter intensity is determined to be the most important satellite-based predictor of %WCC in savannas, likely due to its relatively strong response to non-leaf structural components of small woody plants which remain mostly constant across the wet and dry season. However, the best performing model combines radar backscatter metrics with optical reflectance indices that serve as proxies for greenness, dry biomass, burn incidence, plant water content, chlorophyll content, and seasonality. The primary use of GEE in the methodology makes it scalable and replicable by end-users with limited infrastructure for processing large remote sensing data. © Copyright © 2020 Anchang, Prihodko, Ji, Kumar, Ross, Yu, Lind, Sarr, Diouf and Hanan." "57205447750;55466028000;37063799200;","Classification of SAR and PolSAR images using deep learning: a review",2020,"10.1080/19479832.2019.1655489","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071000183&doi=10.1080%2f19479832.2019.1655489&partnerID=40&md5=8e12cb867efe090c8ac95392c42d308c","Advancement in remote sensing technology and microwave sensors explores the applications of remote sensing in different fields. Microwave remote sensing encompasses its benefits of providing cloud-free, all-weather images and images of day and night. Synthetic Aperture Radar (SAR) images own this capability which promoted the use of SAR and PolSAR images in land use/land cover classification and various other applications for different purposes. A review of different polarimetric decomposition techniques for classification of different regions is introduced in the paper. The general objective of the paper is to help researchers in identifying a deep learning technique appropriate for SAR or PolSAR image classification. The architecture of deep networks which ingest new ideas in the given area of research are also analysed in this paper. Benchmark datasets used in microwave remote sensing have been discussed and classification results of those data are analysed. Discussion on experimental results on one of the benchmark datasets is also provided in the paper. The paper discusses challenges, scope and opportunities in research of SAR/PolSAR images which will be helpful to researchers diving into this area. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57211188827;57211190028;57205472969;57215194287;57203164759;57215206697;57202498571;","Real time hydraulic fracturing pressure prediction with machine learning",2020,"10.2118/199699-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085667429&doi=10.2118%2f199699-ms&partnerID=40&md5=8f9e909b1b0c48e00b84ecd78e7351ba","It is very challenging to consider all factors in hydraulic fracturing pre-job design such as reservoir heterogeneity and near wellbore tortuosity. Engineers must monitor the wellhead pressure and adjust the pumping schedule online to avoid screen-out, optimize the proppant and fluid amount, and save cost. In this paper, we use machine learning to predict wellhead pressure in real time during hydraulic fracturing such that the algorithm can assist engineers to monitor and optimize pumping schedule. We explored several neural network models. For each hydraulic fracturing stage, we train a machine learning (ML) model with the data for the first several minutes and predict the wellhead pressure for the next several minutes; we then add the data for the next several minutes, train a second ML model and predict the pressure for the next couple of minutes; and so on. We used several performance metrics to compare different models and select the best model for deployment on Google cloud, where a real time completions platform is developed and hosted. We selected over 100 hydraulic fracturing stages from several wells completed in the Delaware Basin and tested several ML methods on the historical data. The wellhead pressure can be predicted with an acceptable accuracy by a slightly nonlinear machine learning model. We tested the ML model on Google cloud where real time streaming data such as slurry rate and proppant concentration are gathered. The computation is fast enough that a real time wellhead pressure prediction can be realized. © 2020, Society of Petroleum Engineers" "57204572626;57218444105;57204573288;57204571902;35790967600;24386032700;35758658900;55604938200;","Estimation of surface downward shortwave radiation over China from Himawari-8 AHI data based on random forest",2020,"10.3390/RS12010181","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084007149&doi=10.3390%2fRS12010181&partnerID=40&md5=6c79e3409bc4dd7fdcd5f6849376b38f","Downward shortwave radiation (RS) drives many processes related to atmosphere-surface interactions and has great influence on the earth's climate system. However, ground-measured RS is still insufficient to represent the land surface, so it is still critical to generate high accuracy and spatially continuous RS data. This study tries to apply the random forest (RF) method to estimate the RS from the Himawari-8 Advanced Himawari Imager (AHI) data from February to May 2016 with a two-km spatial resolution and a one-day temporal resolution. The ground-measured RS at 86 stations of the Climate Data Center of the Chinese Meteorological Administration (CDC/CMA) are collected to evaluate the estimated RS data from the RF method. The evaluation results indicate that the RF method is capable of estimating the RS well at both the daily and monthly time scales. For the daily time scale, the evaluation results based on validation data show an overall R value of 0.92, a root mean square error (RMSE) value of 35.38 (18.40%) Wm-2, and a mean bias error (MBE) value of 0.01 (0.01%) Wm-2. For the estimated monthly RS, the overall R was 0.99, the RMSE was 7.74 (4.09%) Wm-2, and the MBE was 0.03 (0.02%) Wm-2 at the selected stations. The comparison between the estimated RS data over China and the Clouds and Earth's Radiant Energy System (CERES) Energy Balanced and Filled (EBAF) RS dataset was also conducted in this study. The comparison results indicate that the RS estimates from the RF method have comparable accuracy with the CERES-EBAF RS data over China but provide higher spatial and temporal resolution. © 2020 by the authors." "57216589708;57216592573;57216585190;6701356847;","Preliminary results from a wildfire detection system using deep learning on remote camera images",2020,"10.3390/RS12010166","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083972504&doi=10.3390%2fRS12010166&partnerID=40&md5=a72608139350861e3abde760678b6b95","Pioneering networks of cameras that can search for wildland fire signatures have been in development for some years (High PerformanceWireless Research & Education Network-HPWREN cameras and the ALERTWildfire camera). While these cameras have proven their worth in monitoring fires reported by other means, we have developed a functioning prototype system that can detect smoke from fires usually within 15 min of ignition, while averaging less than one false positive per day per camera. This smoke detection system relies on machine learning-based image recognition software and a cloud-based work-flow capable of scanning hundreds of cameras every minute. The system is operating around the clock in Southern California and has already detected some fires earlier than the current best methods-people calling emergency agencies or satellite detection from the Geostationary Operational Environmental Satellite (GOES) satellites. This system is already better than some commercial systems and there are still many unexplored methods to further improve accuracy. Ground-based cameras are not going to be able to detect every wildfire, and so we are building a system that combines the best of terrestrial camera-based detection with the best approaches to satellite-based detection. © 2020 by the authors." "57212232454;57203541311;57211679548;57213271630;57210911855;57201432287;25423412200;55650978700;15034793900;","Vehicle global 6-DoF pose estimation under traffic surveillance camera",2020,"10.1016/j.isprsjprs.2019.11.005","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075538985&doi=10.1016%2fj.isprsjprs.2019.11.005&partnerID=40&md5=01ccfda6774be277960e2fc3bf966665","Accurately sensing the global position and posture of vehicles in traffic surveillance videos is a challenging but valuable issue for future intelligent transportation systems. Although in recent years, deep learning has brought about major breakthroughs in the six degrees of freedom (6-DoF) pose estimation of objects from monocular images, accurate estimation of the geographic 6-DoF poses of vehicles using images from traffic surveillance cameras remains challenging. We present an architecture that computes continuous global 6-DoF poses throughout joint 2D landmark estimation and 3D pose reconstruction. The architecture infers the 6-DoF pose of a vehicle from the appearance of the image of the vehicle and 3D information. The architecture, which does not rely on intrinsic camera parameters, can be applied to all surveillance cameras by a pre-trained model. Also, with the help of 3D information from the point clouds and the 3D model itself, the architecture can predict landmarks with few and/or blurred textures. Moreover, because of the lack of public training datasets, we release a large-scale dataset, ADFSC, that contains 120 K groups of data with random viewing angles. Regarding both 2D and 3D metrics, our architecture outperforms existing state-of-the-art algorithms in vehicle 6-DoF estimation. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57203220499;55576827900;56674284600;","Classification methods for point clouds in rock slope monitoring: A novel machine learning approach and comparative analysis",2019,"10.1016/j.enggeo.2019.105326","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074562976&doi=10.1016%2fj.enggeo.2019.105326&partnerID=40&md5=5200957cbf9f87a8dadf3f1aaa827501","High-resolution remote monitoring of slopes using terrestrial LiDAR and photogrammetry is a valuable tool for the management of civil and mining geotechnical asset hazards, but accurately classifying regions of interest in the data is sometimes a difficult and time-consuming task. Filtering unwanted areas of a point cloud, such as vegetation and talus, is often a necessary step before rockfall change detection results can be further processed into actionable information. In addition, long-term monitoring through seasonal vegetation changes and snow presents unique challenges to the goal of accurate classification in an automated workflow. This study presents a Random Forest machine learning approach to improve the classification accuracy and efficiency of terrestrial LiDAR monitoring of complex natural slopes. The algorithm classifies points as vegetation, talus, snow, and bedrock using multi-scale neighborhood geometry, slope, change, and intensity features. The classifier was trained on two manually labeled scans from summer and winter, then tested on three other unseen times. We find that F Score generally remains above 0.9 for talus and vegetation, and above 0.95 for bedrock and snow, indicating very high accuracy and an ability to adapt to changing seasonal conditions. In comparing this approach to CANUPO, an existing classification tool, we find our approach to be generally more accurate and flexible, at the expense of increased complexity and computation time. Comparisons with manual masking and a hybrid approach indicate that a machine learning solution is useful primarily in cases of rapidly changing rock slopes or in climates with significant seasonal variability and snow. © 2019 Elsevier B.V." "57211919678;57213338438;57211921046;25422395300;17344649200;8947710900;8908632800;57194778957;","Watson on the farm: Using cloud-based artificial intelligence to identify early indicators of water stress",2019,"10.3390/rs11222645","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075382375&doi=10.3390%2frs11222645&partnerID=40&md5=fe55ec6faf34a28ad6b30689d14e5173","As demand for freshwater increases while supply remains stagnant, the critical need for sustainable water use in agriculture has led the EPA Strategic Plan to call for new technologies that can optimize water allocation in real-time. This work assesses the use of cloud-based artificial intelligence to detect early indicators of water stress across six container-grown ornamental shrub species. Near-infrared images were previously collected with modified Canon and MAPIR Survey II cameras deployed via a small unmanned aircraft system (sUAS) at an altitude of 30 meters. Cropped images of plants in no, low-, and high-water stress conditions were split into four-fold cross-validation sets and used to train models through IBM Watson's Visual Recognition service. Despite constraints such as small sample size (36 plants, 150 images) and low image resolution (150 pixels by 150 pixels per plant),Watson generated models were able to detect indicators of stress after 48 hours of water deprivation with a significant to marginally significant degree of separation in four out of five species tested (p < 0.10). Two models were also able to detect indicators of water stress after only 24 hours, with models trained on images of as few as eight water-stressed Buddleia plants achieving an average area under the curve (AUC) of 0.9884 across four folds. Ease of pre-processing, minimal amount of training data required, and outsourced computation make cloud-based artificial intelligence services such as IBM Watson Visual Recognition an attractive tool for agriculture analytics. Cloud-based artificial intelligence can be combined with technologies such as sUAS and spectral imaging to help crop producers identify deficient irrigation strategies and intervene before crop value is diminished. When brought to scale, frameworks such as these can drive responsive irrigation systems that monitor crop status in real-time and maximize sustainable water use. © 2019 by the authors." "57207370173;8538711400;25639040500;36451312300;23492895800;57210257320;7402462007;7003724199;56209544000;","A labelled ocean SAR imagery dataset of ten geophysical phenomena from Sentinel-1 wave mode",2019,"10.1002/gdj3.73","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070104608&doi=10.1002%2fgdj3.73&partnerID=40&md5=73fb3ac0d2339ab1d9dc345fd1f45b8e","The Sentinel-1 mission is part of the European Copernicus program aiming at providing observations for Land, Marine and Atmosphere Monitoring, Emergency Management, Security and Climate Change. It is a constellation of two (Sentinel-1 A and B) Synthetic Aperture Radar (SAR) satellites. The SAR wave mode (WV) routinely collects high-resolution SAR images of the ocean surface during day and night and through clouds. In this study, a subset of more than 37,000 SAR images is labelled corresponding to ten geophysical phenomena, including both oceanic and meteorologic features. These images cover the entire open ocean and are manually selected from Sentinel-1A WV acquisitions in 2016. For each image, only one prevalent geophysical phenomenon with its prescribed signature and texture is selected for labelling. The SAR images are processed into a quick-look image provided in the formats of PNG and GeoTIFF as well as the associated labels. They are convenient for both visual inspection and machine learning-based methods exploitation. The proposed dataset is the first one involving different oceanic or atmospheric phenomena over the open ocean. It seeks to foster the development of strategies or approaches for massive ocean SAR image analysis. A key objective was to allow exploiting the full potential of Sentinel-1 WV SAR acquisitions, which are about 60,000 images per satellite per month and freely available. Such a dataset may be of value to a wide range of users and communities in deep learning, remote sensing, oceanography and meteorology. © 2019 The Authors. Geoscience Data Journal published by Royal Meteorological Society and John Wiley & Sons Ltd." "57089599900;54790498500;56452028600;57214139779;","Supervised Fine-Grained Cloud Detection and Recognition in Whole-Sky Images",2019,"10.1109/TGRS.2019.2917612","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078294789&doi=10.1109%2fTGRS.2019.2917612&partnerID=40&md5=70d617446112a975128e1774e6669732","The whole-sky imager has been increasingly used for ground-based cloud automatic observation. Many approaches based on image processing have been applied to detect or classify clouds in whole-sky images (WSIs). However, most of the studies only focus on image segmentation for cloud detection or image classification for cloud recognition separately. The cloud detection only does the binary segmentation (sky and cloud) without cloud types, while the cloud recognition only gives the single image-level label without cloud coverage. In this paper, a fine-grained cloud detection and recognition task with a solution is proposed to fill the gap, which can simultaneously detect and classify clouds in a WSI. It can be regarded as a pixel-level fine-grained dense prediction for images. First, a new data set is built with pixel-level annotation of nine different types. Then, a solution based on supervised learning is proposed, in which the pixel-level prediction problem is converted to a superpixel classification problem. Multiview features are extracted, including color, inside texture, neighbor texture, and global relation, to represent the superpixels. Moreover, a class-specific feature space transformation method based on metric learning and subspace alignment is proposed to overcome the challenge brought by the high similarity among cloud types and the feature shifting. Finally, several experiments have verified that our approach is effective to the challenging new task and also outperforms some other methods in the normal tasks of cloud detection and cloud classification, respectively. © 1980-2012 IEEE." "57211313601;25931139100;6701481007;56237449000;57210222492;","Potential of passive microwave around 183 GHz for snowfall detection in the arctic",2019,"10.3390/rs11192200","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073425598&doi=10.3390%2frs11192200&partnerID=40&md5=2d4897156259c2d526d8d2f5a3be359e","This study evaluates the potential use of the Microwave Humidity Sounder (MHS) for snowfall detection in the Arctic. Using two years of colocated MHS and CloudSat observations, we develop an algorithm that is able to detect up to 90% of the most intense snowfall events (snow water path ≥400 g m-2) and 50% of the weak snowfall rate events (snow water path ≤50 g m-2). The brightness temperatures at 190.3 GHz and 183.3 ± 3 GHz, the integrated water vapor, and the temperature at 2mare identified as the most important variables for snowfall detection. The algorithm tends to underestimate the snowfall occurrence over Greenland and mountainous areas (by as much as -30%), likely due to the dryness of these areas, and to overestimate the snowfall occurrence over the northern part of the Atlantic (by up to 30%), likely due to the occurrence of mixed phase precipitation. An interpretation of the selection of the variables and their importance provides a better understanding of the snowfall detection algorithm. This work lays the foundation for the development of a snowfall rate quantification algorithm. © 2019 by the authors." "54963866700;7005397699;7102775296;57202645627;6603135119;57189389987;","Monitoring land-cover and land-use dynamics in Fanjingshan National Nature Reserve",2019,"10.1016/j.apgeog.2019.102077","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071600898&doi=10.1016%2fj.apgeog.2019.102077&partnerID=40&md5=089d45e356459f098561636764b36813","Fanjingshan National Nature Reserve (FNNR) in China is a biodiversity hotspot that is part of a larger, multi-use landscape where tourism, farming, grazing, and other land uses occur. Payment for ecosystem services (PES) programs that encourage afforestation on farmlands may be important drivers of land-cover and land-use change in the region that surrounds FNNR. Our objective is to monitor and examine vegetation and land-use changes, including PES-related afforestation, between 1989 and 2017. We utilize several image processing techniques, such as illumination normalization approaches to suppress terrain effects, and multi-seasonal image compositing to minimize persistent cloud cover. Ancillary data were also incorporated to generate reliable vegetation and land-use change information. A random forest machine learning image classification routine is implemented through the cloud-based Google Earth Engine platform and refined using optimal classifier parameter tuning. Land-use transitions are identified and mapped with the implementation of stable training sites, discrete image classification, and logical land-use transition rules. Accuracy assessment results indicate our change detection workflow provides a reliable methodology to remotely monitor long-term forest cover and land-use changes in this mountainous, forested, and cloud prevalent region. We quantify the area of new built development and afforestation land and found that most of the land transitions took place in reserve buffer and its adjacent environs. For example, less than 2 km2 of new built was identified within the reserve boundary compared to 25 km2 for the entire study area between 1995 and 2016. We also shed light on the strengths and weaknesses of using Google Earth Engine for land-cover and land-use change studies. This efficient and open-access technique is important not only for assessing environmental changes and PES efficacy, but also for evaluating other conservation policies elsewhere. © 2019" "57188722540;57211096161;57203236301;55351273900;","Developing a new machine-learning algorithm for estimating Chlorophyll-a concentration in optically complex waters: A case study for high northern latitude waters by using Sentinel 3 OLCI",2019,"10.3390/rs11182076","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072647161&doi=10.3390%2frs11182076&partnerID=40&md5=4ff1c86de073b83ae89ab256a51a8473","The monitoring of Chlorophyll-a (Chl-a) concentration in high northern latitude waters has been receiving increased focus due to the rapid environmental changes in the sub-Arctic, Arctic. Spaceborne optical instruments allow the continuous monitoring of the occurrence, distribution, and amount of Chl-a. In recent years, the Ocean and Land Color Instruments (OLCI) onboard the Sentinel 3 (S3) A and B satellites were launched, which provide data about various aquatic environments on advantageous spatial, spectral, and temporal resolutions with high SNR. Although S3 OLCI could be favorable to monitor high northern latitude waters, there have been several challenges related to Chl-a concentration retrieval in these waters due to their unique optical properties coupled with challenging environments including high sun zenith angle, presence of sea ice, and frequent cloud covers. In this work, we aim to overcome these difficulties by developing a machine-learning (ML) approach designed to estimate Chl-a concentration from S3 OLCI data in high northern latitude optically complex waters. The ML model is optimized and requires only three S3 OLCI bands, reflecting the physical characteristic of Chl-a as input in the regression process to estimate Chl-a concentration with improved accuracy in terms of the bias (five times improvements.) The ML model was optimized on data from Arctic, coastal, and open waters, and showed promising performance. Finally, we present the performance of the optimized ML approach by computing Chl-a maps and corresponding certainty maps in highly complex sub-Arctic and Arctic waters. We show how these certainty maps can be used as a support to understand possible radiometric calibration issues in the retrieval of Level 2 reflectance over these waters. This can be a useful tool in identifying erroneous Level 2 Remote sensing reflectance due to possible failure of the atmospheric correction algorithm. © 2019 by the authors." "57207775983;23035058500;","Pyramid scene parsing network in 3D: Improving semantic segmentation of point clouds with multi-scale contextual information",2019,"10.1016/j.isprsjprs.2019.06.010","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068136296&doi=10.1016%2fj.isprsjprs.2019.06.010&partnerID=40&md5=3dabd4d669a4f5c06dd7cd919ca55cea","Analyzing and extracting geometric features from 3D data is a fundamental step in 3D scene understanding. Recent works demonstrated that deep learning architectures can operate directly on raw point clouds, i.e. without the use of intermediate grid-like structures. These architectures are however not designed to encode contextual information in-between objects efficiently. Inspired by a global feature aggregation algorithm designed for images (Zhao et al., 2017), we propose a 3D pyramid module to enrich pointwise features with multi-scale contextual information. Our module can be easily coupled with 3D semantic segmentation methods operating on 3D point clouds. We evaluated our method on three large scale datasets with four baseline models. Experimental results show that the use of enriched features brings significant improvements to the semantic segmentation of indoor and outdoor scenes. © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57209238608;55083178100;55574436900;6603637346;57192690229;","Semi-automatic extraction of liana stems from terrestrial LiDAR point clouds of tropical rainforests",2019,"10.1016/j.isprsjprs.2019.05.011","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066956934&doi=10.1016%2fj.isprsjprs.2019.05.011&partnerID=40&md5=9cf26434cb686e9a1fb6dedeb589839e","Lianas are key structural elements of tropical forests having a large impact on the global carbon cycle by reducing tree growth and increasing tree mortality. Despite the reported increasing abundance of lianas across neotropics, very few studies have attempted to quantify the impact of lianas on tree and forest structure. Recent advances in high resolution terrestrial laser scanning (TLS) systems have enabled us to quantify the forest structure, in an unprecedented detail. However, the uptake of TLS technology to study lianas has not kept up with the same pace as it has for trees. The slower technological adoption of TLS to study lianas is due to the lack of methods to study these complex growth forms. In this study, we present a semi-automatic method to extract liana woody components from plot-level TLS data of a tropical rainforest. We tested the method in eight plots from two different tropical rainforest sites (two in Gigante Peninsula, Panama and six in Nouragues, French Guiana) along an increasing gradient of liana infestation (from plots with low liana density to plots with very high liana density). Our method uses a machine learning model based on the Random Forest (RF) algorithm. The RF algorithm is trained on the eigen features extracted from the points in 3D at multiple spatial scales. The RF based liana stem extraction method successfully extracts on average 58% of liana woody points in our dataset with a high precision of 88%. We also present simple post-processing steps that increase the percentage of extracted liana stems from 54% to 90% in Nouragues and 65% to 70% in Gigante Peninsula without compromising on the precision. We provide the entire processing pipeline as an open source python package. Our method will facilitate new research to study lianas as it enables the monitoring of liana abundance, growth and biomass in forest plots. In addition, the method facilitates the easier processing of 3D data to study tree structure from a liana-infested forest. © 2019 The Authors" "57195357570;7006609519;","Evaluation of Machine Learning Classifiers for Predicting Deep Convection",2019,"10.1029/2018MS001561","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067885353&doi=10.1029%2f2018MS001561&partnerID=40&md5=729877b1fba462d714305da5fe448994","The realistic representation of convection in atmospheric models is paramount for skillful predictions of hazardous weather as well as climate, yet climate models especially suffer from large uncertainties in the parameterization of clouds and convection. In this work, we examine the use of machine learning (ML) to predict the occurrence of deep convection from a state-of-the-art atmospheric reanalysis (ERA5). Logistic regression, random forests, gradient-boosted decision trees, and deep neural networks were trained with lightning data to predict thunderstorm occurrence (TO) in Central and Northern Europe (2012–2017) and in Sri Lanka (2016–2017). Up to 40 input variables were used, representing, for example, instability, humidity, and inhibition. Feature importances derived for the various models emphasize the high importance of conditional instability for deep convection in Europe, while in Sri Lanka, TO is more strongly regulated by humidity. The Precision-Recall curve indicates more than a twofold improvement in skill over convective available potential energy for short-term (0–45 min) predictions of TO in Europe by using neural networks or gradient-boosted decision tree and a larger improvement in the tropical domain. The diurnal cycle of deep convection is closely reproduced, suggesting that ML could be used to trigger convection in climate models. Finally, a strong relationship was found between area-mean monthly TO and ML predictions, with correlation coefficients exceeding 0.94 in all domains. Convective available potential energy has a similar level of correlation with monthly thunderstorm activity only in Northern Europe. The results encourage the use of reanalyses and ML to study climate trends in convective storms. ©2019. The Authors." "57191421180;13610913700;57187582800;23390358300;","Deeply integrating Linked Data with Geographic Information Systems",2019,"10.1111/tgis.12538","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067416193&doi=10.1111%2ftgis.12538&partnerID=40&md5=71abf8b4ccf5e2cc34c1bc71e92d343e","The realization that knowledge often forms a densely interconnected graph has fueled the development of graph databases, Web-scale knowledge graphs and query languages for them, novel visualization and query paradigms, as well as new machine learning methods tailored to graphs as data structures. One such example is the densely connected and global Linked Data cloud that contains billions of statements about numerous domains, including life science and geography. While Linked Data has found its way into everyday applications such as search engines and question answering systems, there is a growing disconnect between the classical ways in which Geographic Information Systems (GIS) are still used today and the open-ended, exploratory approaches used to retrieve and consume data from knowledge graphs such as Linked Data. In this work, we conceptualize and prototypically implement a Linked Data connector framework as a set of toolboxes for Esri's ArcGIS to close this gap and enable the retrieval, integration, and analysis of Linked Data from within GIS. We discuss how to connect to Linked Data endpoints, how to use ontologies to probe data and derive appropriate GIS representations on the fly, how to make use of reasoning, how to derive data that are ready for spatial analysis out of RDF triples, and, most importantly, how to utilize the link structure of Linked Data to enable analysis. The proposed Linked Data connector framework can also be regarded as the first step toward a guided geographic question answering system over geographic knowledge graphs. © 2019 John Wiley & Sons Ltd" "57206278550;55695331800;36194585400;56580854000;","LPCCNet: A Lightweight Network for Point Cloud Classification",2019,"10.1109/LGRS.2018.2889472","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066409297&doi=10.1109%2fLGRS.2018.2889472&partnerID=40&md5=8dd55b368382f7a53ee58489e455cb91","Deep learning has achieved much in image and natural language processing, and related research has also expanded into the field of point cloud processing, and deep networks dedicated to point clouds have emerged; however, in many points cloud applications, networks with fewer parameters and lower computational burdens are required to satisfy the requirements of miniaturisation of devices, so research into lightweight point cloud deep networks is justified. In this letter, we described the design of a lightweight point cloud classification deep learning architecture, lightweight point cloud classification network, with the self-designed block structure as a unit, using a novel index fully dense connectivity method that interconnects all convolutional layers of the network, improving the utilisation of features, thereby reducing the parameter size of each layer; at the same time, the network applies grouping convolution and pruning to the convolutional layer and the fully connected layer, further reducing the amount of parameters and calculation burden. The experimental results show the effectiveness of the network that can achieve comparable results with existing large networks, such as PointNet++, with a smaller amount of parameters. The proposed network offers great promise in mobile device deployment and real-time processing. © 2019 IEEE." "57190801454;57196021613;57188582312;23004944100;","FEATURE RELEVANCE ANALYSIS for 3D POINT CLOUD CLASSIFICATION USING DEEP LEARNING",2019,"10.5194/isprs-annals-IV-2-W5-373-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067480190&doi=10.5194%2fisprs-annals-IV-2-W5-373-2019&partnerID=40&md5=e9dc173480ecc83d1aa8eff868e91321","3D point clouds acquired by laser scanning and other techniques are difficult to interpret because of their irregular structure. To make sense of this data and to allow for the derivation of useful information, a segmentation of the points in groups, units, or classes fit for the specific use case is required. In this paper, we present a non-end-to-end deep learning classifier for 3D point clouds using multiple sets of input features and compare it with an implementation of the state-of-the-art deep learning framework PointNet++. We first start by extracting features derived from the local normal vector (normal vectors, eigenvalues, and eigenvectors) from the point cloud, and study the result of classification for different local search radii. We extract additional features related to spatial point distribution and use them together with the normal vector-based features. We find that the classification accuracy improves by up to 33% as we include normal vector features with multiple search radii and features related to spatial point distribution. Our method achieves a mean Intersection over Union (mIoU) of 94% outperforming PointNet++'s Multi Scale Grouping by up to 12%. The study presents the importance of multiple search radii for different point cloud features for classification in an urban 3D point cloud scene acquired by terrestrial laser scanning. © Authors 2019." "57201500562;7005427503;21935206500;55768720600;25642386000;55768214000;","Proof of concept of a novel cloud computing approach for object-based remote sensing data analysis and classification",2019,"10.1080/15481603.2018.1538621","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85061363718&doi=10.1080%2f15481603.2018.1538621&partnerID=40&md5=d8ba99730aa221a105903a42f00790b8","Advances in the development of Earth observation data acquisition systems have led to the continuously growing production of remote sensing datasets, for which timely analysis has become a major challenge. In this context, distributed computing technology can provide support for efficiently handling large amounts of data. Moreover, the use of distributed computing techniques, once restricted by the availability of physical computer clusters, is currently widespread due to the increasing offer of cloud computing infrastructure services. In this work, we introduce a cloud computing approach for object-based image analysis and classification of arbitrarily large remote sensing datasets. The approach is an original combination of different distributed methods which enables exploiting machine learning methods in the creation of classification models, through the use of a web-based notebook system. A prototype of the proposed approach was implemented with the methods available in the InterCloud system integrated with the Apache Zeppelin notebook system, for collaborative data analysis and visualization. In this implementation, the Apache Zeppelin system provided the means for using the scikit-learn Python machine learning library in the design of a classification model. In this work we also evaluated the approach with an object-based image land-cover classification of a GeoEye-1 scene, using resources from a commercial cloud computing infrastructure service provided. The obtained results showed the effectiveness of the approach in efficiently handling a large data volume in a scalable way, in terms of the number of allocated computing resources. © 2018, © 2018 Informa UK Limited, trading as Taylor & Francis Group." "26422033000;54385731300;","VAULT MODELING with NEURAL NETWORKS",2019,"10.5194/isprs-archives-XLII-2-W9-81-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065440524&doi=10.5194%2fisprs-archives-XLII-2-W9-81-2019&partnerID=40&md5=bf6b0d3b12d95814d01431cfa5751286","Nowadays, the digital reconstruction of vaults is carried out using photogrammetric and laser scanning techniques able to capture the visible surface with dense point clouds. Then, different modeling strategies allow the generation of 3D models in various formats, such as meshes that interpolates the acquired point cloud, NURBS-based reconstructions based on manual, semi-automated, or automated procedures, and parametric objects for Building Information Modeling. This paper proposes a novel method that reconstructs the visible surface of a vault using neural networks. It is based on the assumption that vaults are not irregular free-form objects, but they can be reconstructed by mathematical functions calculated from the acquired point clouds. The proposed approach uses the point cloud to train a neural network that approximates vault surface. The achieved solution is not only able to consider the basic geometry of the vault, but also its irregularities that cannot be neglected in the case of accurate and detailed modeling projects of historical vaults. Considerations on the approximation capabilities of neural networks are illustrated and discussed along with the advantages of creating a mathematical representation encapsulated into a function. © Authors 2019." "57208213951;8530680600;57205365890;57142243800;57218527471;","Intelligent oilfield - Cloud based big data service in upstream oil and gas",2019,"10.2523/iptc-19418-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086057515&doi=10.2523%2fiptc-19418-ms&partnerID=40&md5=257be1bd82ef6ca4e02c56e31b08a447","The Oil and Gas (O&G) industry is embracing modern and intelligent digital technologies such as big data analytics, cloud services, machine learning etc. to increase productivity, enhance operations safety, reduce operation cost and mitigate adverse environmental impact. Challenges that come with such an oil field digital transformation include, but are certainly not limited to: information explosion; isolated and incompatible data repositories; logistics for data exchange and communication; obsoleted processes; cost of support; and the lack of data security. In this paper, we introduce an elastically scalable cloud-based platform to provide big data service for the upstream oil and gas industry, with high reliability and high performance on real-time or near real-time services based on industry standards. First, we review the nature of big data within O&G, paying special attention to distributed fiber optic sensing technologies. We highlight the challenges and necessary system requirements to build effective and scalable downhole big data management and analytics. Secondly, we propose a cloud-based platform architecture for data management and analytics services. Finally, we will present multiple case studies and examples with our system as it is applied in the field. We demonstrate that a standardized data communication and security model enables high efficiency for data transmission, storage, management, sharing and processing in a highly secure environment. Using a standard big data framework and tools (e.g., Apache Hadoop, Spark and Kafka) together with machine learning techniques towards autonomous analysis of such data sources, we are able to process extremely large and complex datasets in an efficient way to provide real-time or near real-time data analytical service, including prescriptive and predictive analytics. The proposed integrated service comprises multiple main systems, such as a downhole data acquisition system; data exchange and management system; data processing and analytics system; as well as data visualization, event alerting and reporting system. With emerging fiber optic technologies, this system not only provides services using legacy O&G data such as static reservoir information, fluid characteristics, well log, well completion information, downhole sensing and surface monitoring data, but also incorporates distributed sensing data (DxS) such as distributed temperature sensing (DTS), distributed strain sensing (DSS) and distributed acoustic sensing (DAS) for continuous downhole measurements along the wellbore with very high spatial resolution. It is the addition of the fiber optic distributed sensing technology that has increased exponentially the volume of downhole data needed to be transmitted and securely managed. © 2019, International Petroleum Technology Conference" "57211428966;57211428231;7801639422;","Machine-learning-based petrophysical property modeling",2019,"10.2118/195436-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077088655&doi=10.2118%2f195436-ms&partnerID=40&md5=8c95e52d380ce49e105deb635eb278c3","The current cycle for reservoir management requires several months to years to update static and dynamic models as additional data from the field [logs, production, pressures, core, four-dimensional (4D), etc.] are obtained. These delays in updating the models result in increased risk and contribute to a significant loss of economic value. The ultimate goal for next-generation reservoir management is to reduce the cycle from several months to a few days. The current challenges for developing a proactive/real-time reservoir management solution include but are not limited to the time and manual intervention involved in conditioning and interpreting the logging-while-drilling (LWD) and well log data acquired during and after drilling a well; updating three-dimensional (3D) petrophysical/static models; and the computational cost and time involved in generating reservoir models from static and production data (history matching). However, the current widespread use of machine-learning and cloud-computing capabilities leads to faster and more accurate models, enabling real-time or near-real-time decision making. Using machine learning, one of these challenges—updating the 3D static models—was successfully addressed, namely, updating porosity prediction in a 3D model after new information comes to light, such as logging from a newly drilled well. The conventional geostatistical approach does not always honor geological variations in the subsurface formations, because only one or two seismic attributes can be used for co-simulation, and only with first-order interactions. Additionally, and most important, generating hundreds of realizations on a 3D grid is computationally intense and time consuming. Typically, several weeks are necessary to generate these static models before feeding them into the reservoir model. The proposed solution is a machine-learning-based approach that integrates 3D spatial availability of seismic data with petrophysical properties. One important goal of reservoir management is to understand reservoir uncertainties before they adversely affect field development. This machine learning solution proved to be computationally less costly, more accurate, and much faster than the conventional geostatistical approach. Copyright 2019, Society of Petroleum Engineers." "57209459973;6602513090;15066257900;7102829666;","High temporal rainfall estimations from himawari-8 multiband observations using the random-forest machine-learning method",2019,"10.2151/jmsj.2019-040","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067793825&doi=10.2151%2fjmsj.2019-040&partnerID=40&md5=e72ef45f54abeca1e564e1ab2e84bc96","We introduce a novel rainfall-estimating algorithm with a random-forest machine-learning method only from Infrared (IR) observations. As training data, we use nine-band brightness temperature (BT) observations, obtained from IR radiometers, on the third-generation geostationary meteorological satellite (GEO) Himawari-8 and precipitation radar observations from the Global Precipitation Measurement core observatory. The Himawari-8 Rainfall-estimating Algorithm (HRA) enables us to estimate the rain rate with high spatial and temporal resolution (i.e., 0.04° every 10 min), covering the entire Himawari-8 observation area (i.e., 85°E – 155°W, 60°S – 60°N) based solely on satellite observations. We conducted a case analysis of the Kanto–Tohoku heavy rainfall event to compare HRA rainfall estimates with the near-real-time version of the Global Satellite Mapping of Precipitation (GSMaP_NRT), which combines global rainfall estimation products with microwave and IR BT observations obtained from satellites. In this case, HRA could estimate heavy rainfall from warm-type precipitating clouds. The GSMaP_NRT could not estimate heavy rainfall when microwave satellites were unavailable. Further, a statistical analysis showed that the warm-type heavy rain seen in the Asian monsoon region occurred frequently when there were small BT differences between the 6.9-μm and 7.3-μm of water vapor (WV) bands (ΔT6.9 – 7.3). Himawari-8 is the first GEO to include the 6.9-μm band, which is sensitive to middle-to-upper tropospheric WV. An analysis of the WV multibands’ weighting functions revealed that ΔT6.9 – 7.3 became small when the WV amount in the middle-to-upper troposphere was small and there were optically thick clouds with the cloud top near the middle troposphere. Statistical analyses during boreal summer (August and September 2015 and July 2016) and boreal winter (December 2015 and January and February 2016) indicate that HRA has higher estimation accuracy for heavy rain from warm-type precipitating clouds than a conventional rain estimation method based on only one IR band. © The Author(s) 2019." "15841395400;7102725240;","Comparing the accuracy of MODIS data products for vegetation detection between two environmentally dissimilar ecoregions: the Chocó-Darien of South America and the Great Basin of North America",2019,"10.1080/15481603.2019.1611024","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065199141&doi=10.1080%2f15481603.2019.1611024&partnerID=40&md5=31c9c9234af11281cd73f83cd6613a8f","The daily images produced by the MODIS (Moderate Resolution Imaging Spectroradiometer) sensor aboard the Terra and Aqua satellites have been widely used to monitor global vegetation. Using these data, the Earth Observing System operated by US National Aeronautics and Space Administration (NASA) has developed a variety of MODIS products focused on the monitoring and evaluation of vegetation condition. These products have three possible sources of variation that can affect the sensitivity of vegetation detection: 1) orbital and mechanical differences between MODIS sensors aboard Aqua or Terra, 2) the preprocessing algorithms used to generate multitemporal cloud-free mosaics (MAIAC or original MODIS algorithm), and/or 3) post-processing algorithms applied by users to optimize vegetation index values derived from temporal sequences of imagery. We evaluated these sources of variation by comparing the results of a vegetation classification for two different ecoregions. The accuracies of vegetation classifications utilizing either the Aqua or Terra MODIS sensors, the MAIAC or original MODIS preprocessing algorithms, and two common post-processing techniques (Asymmetric Gaussian or Savitzky and Golay function) were compared to determine which set of techniques or sensors yielded the best results. The ecoregions we chose to use were the Great Basin of North America and Chocó-Darien of South America. We compared four different MODIS data products (MOD13Q1, MYD13Q1, MOD09Q1, and MYD09Q1) as predictor variables using Random Forest as the classification algorithm to generate a land cover map. We found that the accuracy of the vegetation classifications (using Kappa as measure of accuracy) changed significantly depending on the MODIS platform (Terra or Aqua), the preprocessing algorithm (MAIAC or MODIS), and the two postprocessing algorithms for both ecoregions. Our result suggests that comparative analyses are needed to optimize the results when equivalent MODIS products are used in vegetation detection and classification. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "56440802100;55324277300;8516206500;","A weekly, continually updated dataset of the probability of large wildfires across western US forests and woodlands",2018,"10.5194/essd-10-1715-2018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85053902211&doi=10.5194%2fessd-10-1715-2018&partnerID=40&md5=4da8fa568334617fa66dc42c61094b5c","There is broad consensus that wildfire activity is likely to increase in western US forests and woodlands over the next century. Therefore, spatial predictions of the potential for large wildfires have immediate and growing relevance to near-and long-term research, planning, and management objectives. Fuels, climate, weather, and the landscape all exert controls on wildfire occurrence and spread, but the dynamics of these controls vary from daily to decadal timescales. Accurate spatial predictions of large wildfires should therefore strive to integrate across these variables and timescales. Here, we describe a high spatial resolution dataset (250 m pixel) of the probability of large wildfires (> 405 ha) across forests and woodlands in the contiguous western US, from 2005 to the present. The dataset is automatically updated on a weekly basis using Google Earth Engine and a continuous integration/q pipeline. Each image in the dataset is the output of a random forest machine-learning algorithm, trained on random samples of historic small and large wildfires and represents the predicted conditional probability of an individual pixel burning in a large fire, given an ignition or fire spread to that pixel. This novel workflow is able to integrate the near-term dynamics of fuels and weather into weekly predictions while also integrating longer-term dynamics of fuels, the climate, and the landscape. As a continually updated product, the dataset can provide operational fire managers with contemporary, on-the-ground information to closely monitor the changing potential for large wildfire occurrence and spread. It can also serve as a foundational dataset for longer-term planning and research, such as the strategic targeting of fuels management, fire-smart development at the wildland-urban interface, and the analysis of trends in wildfire potential over time. large fire probability GeoTiff products from 2005 to 2017 are archived on the Figshare online digital repository with the DOI https://doi.org/10.6084/m9.figshare.5765967 (available at https://doi.org/10.6084/m9.figshare.5765967.v1). Weekly GeoTiff products and the entire dataset from 2005 onwards are also continually uploaded to a Google Cloud Storage bucket at https://console.cloud.google.com/storage/wffr-preds/V1 (last access: 14 September 2018) and are available free of charge with a Google account. Continually updated products and the long-term archive are also available to registered Google Earth Engine (GEE) users as public GEE assets and can be accessed with the image collection ID ""users/mgray/wffr-preds"" within GEE. © 2018 Author(s)." "57211884590;6602286722;57203254582;7006056290;6602885902;55567000300;7006501368;8950336900;7007170118;7410390738;57192201280;7004215217;","Hubble tarantula treasury project - VI. Identification of pre-main-sequence stars using machine-learning techniques",2018,"10.1093/mnras/sty1317","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076877919&doi=10.1093%2fmnras%2fsty1317&partnerID=40&md5=ea9d1d9b4428593b879b6336cc772dd6","The Hubble Tarantula Treasury Project (HTTP) has provided an unprecedented photometric coverage of the entire starburst region of 30 Doradus down to the half Solar mass limit. We use the deep stellar catalogue of HTTP to identify all the pre-main-sequence (PMS) stars of the region, i.e. stars that have not started their lives on the main-sequence yet. The photometric distinction of these stars from the more evolved populations is not a trivial task due to several factors that alter their colour-magnitude diagram positions. The identification of PMS stars requires, thus, sophisticated statistical methods. We employ machine-learning classification techniques on the HTTP survey of more than 800 000 sources to identify the PMS stellar content of the observed field. Our methodology consists of (1) carefully selecting the most probable low-mass PMS stellar population of the star-forming cluster NGC2070, (2) using this sample to train classification algorithms to build a predictive model for PMS stars, and (3) applying this model in order to identify the most probable PMS content across the entire Tarantula Nebula. We employ decision tree, random forest (RF), and support vector machine (SVM) classifiers to categorize the stars as PMS and non-PMS. The RF and SVM provided the most accurate models, predicting about 20 000 sources with a candidateship probability higher than 50 per cent, and almost 10 000 PMS candidates with a probability higher than 95 per cent. This is the richest and most accurate photometric catalogue of extragalactic PMS candidates across the extent of a whole star-forming complex. © 2018 The Author(s)." "57192176367;7202105464;7103127999;","A Hybrid CNN + Random Forest Approach to Delineate Debris Covered Glaciers Using Deep Features",2018,"10.1007/s12524-018-0750-x","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85048596968&doi=10.1007%2fs12524-018-0750-x&partnerID=40&md5=ce89e8af0a53d1d372afc90def96e3a9","The main aim of this study is to propose a novel hybrid deep learning framework approach for accurate mapping of debris covered glaciers. The framework comprises of integration of several CNNs architecture, in which different combinations of Landsat 8 multispectral bands (including thermal band), topographic and texture parameters are passed as input for feature extraction. The output of an ensemble of these CNNs is hybrid with random forest model for classification. The major pillars of the framework include: (1) technique for implementing topographic and atmospheric corrections (preprocessing), (2) the proposed hybrid of ensemble of CNNs and random forest classifier, and (3) procedures to determine whether a pixel predicted as snow is a cloud edge/shadow (post-processing). The proposed approach was implemented on the multispectral Landsat 8 OLI (operational land imager)/TIRS (thermal infrared sensor) data and Shuttle Radar Topography Mission Digital Elevation Model for the part of the region situated in Alaknanda basin, Uttarakhand, Himalaya. The proposed framework was observed to outperform (accuracy 96.79%) the current state-of-art machine learning algorithms such as artificial neural network, support vector machine, and random forest. Accuracy assessment was performed by means of several statistics measures (precision, accuracy, recall, and specificity). © 2018, Indian Society of Remote Sensing." "57195946585;56067767500;","Downscaling of MODIS land surface temperature to LANDSAT scale using multi-layer perceptron",2017,"10.7848/ksgpc.2017.35.4.313","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85030320620&doi=10.7848%2fksgpc.2017.35.4.313&partnerID=40&md5=d1cc028ae043ad0b64f405f0d964abc0","Land surface temperature is essential for monitoring abnormal climate phenomena such as UHI (Urban Heat Islands), and for modeling weather patterns. However, the quality of surface temperature obtained from the optical space imagery is affected by many factors such as, revisit period of the satellite, instance of capture, spatial resolution, and cloud coverage. Landsat 8 imagery, often used to obtain surface temperatures, has a high resolution of 30 meters (100 meters rearranged to 30 meters) and a revisit frequency of 16 days. On the contrary, MODIS imagery can be acquired daily with a spatial resolution of about 1 kilometer. Many past attempts have been made using both Landsat and MODIS imagery to complement each other to produce an imagery of improved temporal and spatial resolution. This paper applied machine learning methods and performed downscaling which can obtain daily based land surface temperature imagery of 30 meters." "36185813900;6603817132;6602169970;","TRACING DENSE and DIFFUSE NEUTRAL HYDROGEN in the HALO of the MILKY WAY",2017,"10.3847/1538-4357/834/2/155","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85010073105&doi=10.3847%2f1538-4357%2f834%2f2%2f155&partnerID=40&md5=c03ef601e5bf6ef01bac6b79c170ba66","We have combined observations of Galactic high-velocity H i from two surveys: a very sensitive survey from the Green Bank 140 ft Telescope with limited sky coverage, and the less sensitive but complete Galactic All Sky Survey from the 64 m Parkes Radio Telescope. The two surveys preferentially detect different forms of neutral gas due to their sensitivity. We adopt a machine learning approach to divide our data into two populations that separate across a range in column density: (1) a narrow line-width population typical of the majority of bright high velocity cloud components, and (2) a fainter, broad line-width population that aligns well with that of the population found in the Green Bank survey. We refer to these populations as dense and diffuse gas, respectively, and find that diffuse gas is typically located at the edges and in the tails of high velocity clouds, surrounding dense components in the core. A fit to the average spectrum of each type of gas in the Galactic All Sky Survey data reveals the dense population to have a typical line width of ∼20 km s-1 and brightness temperature of ∼0.3 K, while the diffuse population has a typical line width of ∼30 km s-1 and a brightness temperature of ∼0.2 K. Our results confirm that most surveys of high velocity gas in the Milky Way halo are missing the majority of the ubiquitous diffuse gas, and that this gas is likely to contribute at least as much mass as the dense gas. © 2017. The American Astronomical Society. All rights reserved." "49961944400;7003648498;36148030600;","An automated method for time-series human settlement mapping using Landsat data and existing land cover maps",2016,"10.1109/IGARSS.2016.7729458","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85007481059&doi=10.1109%2fIGARSS.2016.7729458&partnerID=40&md5=054dccc9336bbcec28a2335a6cbcf18f","Automation of satellite-based human settlement mapping is highly needed to utilize historical archives of satellite data for urgent issues of urban development in global scale. We developed an automated algorithm to detect human settlement from Landsat satellite data. A machine learning algorithm, Local and Global Consistency (LLGC), was applied with improvements for remote sensing data. The algorithm enables to use existing coarse-resolution land cover maps as a training dataset so that any manual process is not required for preparation of training data. In addition, for better robustness against uncertainty in satellite data, we proposed a method to combine several LLGC results for several dates in a certain period from a target date into a single human settlement map with a pixel-based median composition among the input LLGC results. Combination of the methods enabled to develop timeseries human settlement maps using Landsat data, single ones of which could be affected by cloud contaminations. We applied the algorithm to Landsat data of 838 WRS2 tiles for cities with more than one million people worldwide for 1990, 2000, 2005, and 2010. MCD12Q1, a MODIS-based global land cover map with 500-m resolution, was used as training data. Visual assessment of the results suggested next steps for improvement of the method. © 2016 IEEE." "55793557100;34769569000;","Automated classification of detected surface damage from point clouds with supervised learning",2016,"10.22260/isarc2016/0038","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84994218568&doi=10.22260%2fisarc2016%2f0038&partnerID=40&md5=2033d0787419a759dd815bfd6b7241fa","Recent advances in sensing technologies provide opportunity to utilize advanced data collection equipment such as laser scanners, high-resolution cameras, etc. for both short- and long-time monitoring of structures. Laser scanners, which are capable of collecting up-to two million points per second along with high-resolution images, are especially evolving rapidly and their usage for capturing and documenting the current condition of varying structural types is becoming increasingly feasible. The authors' previous research has focused on developing generalized algorithms for detecting surface damage on structures from captured laser scans and images. These algorithms are capable of automatically detecting surface damage on varying structural types as well as several construction materials since they use underlying surface geometry for performing damage detection. These damage types include concrete cracking, concrete spalling, steel delamination, steel section loss, bent members, and other misalignments. This paper investigates the use of supervised learning methods for determining whether a detected region represents an actual damage location. First, the input feature representations of the learning algorithms for each damage type are determined and the associated training sets, which are representative of real-world use of the algorithms, are prepared. Once the training and validation are completed, the learning algorithms are tested on the detected damage data, which is obtained by running the damage detection algorithms on both laser scans and texture-mapped images from a concrete test setup. Finally, the accuracy of these learning algorithms is investigated." "56566776600;23088477000;55768214000;27567900100;25642386000;6602453684;7006613644;7007165803;","On the architecture of a big data classification tool based on a map reduce approach for hyperspectral image analysis",2015,"10.1109/IGARSS.2015.7326066","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84962592030&doi=10.1109%2fIGARSS.2015.7326066&partnerID=40&md5=3d1bf8f082e3b9ded79e032ae2586fed","Advances in remote sensors are providing exceptional quantities of large-scale data with increasing spatial, spectral and temporal resolutions, raising new challenges in its analysis, e.g. those presents in classification processes. This work presents the architecture of the InterIMAGE Cloud Platform (ICP): Data Mining Package; a tool able to perform supervised classification procedures on huge amounts of data, on a distributed infrastructure. The architecture is implemented on top of the MapReduce framework. The tool has four classification algorithms implemented taken from WEKA's machine learning library, namely: Decision Trees, Naïve Bayes, Random Forest and Support Vector Machines. The SVM classifier was applied on datasets of different sizes (2 GB, 4 GB and 10 GB) for different cluster configurations (5, 10, 20, 50 nodes). The results show the tool as a potential approach to parallelize classification processes on big data. © 2015 IEEE." "55767585900;8709691100;23011351000;55636321858;6505873389;6603890691;57202048106;","Remote Sensing for Assessing Landslides and Associated Hazards",2020,"10.1007/s10712-020-09609-1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090231196&doi=10.1007%2fs10712-020-09609-1&partnerID=40&md5=8ea61dbf11eeb6172389bfe1adbe33bd","Multi-platform remote sensing using space-, airborne and ground-based sensors has become essential tools for landslide assessment and disaster-risk prevention. Over the last 30 years, the multiplicity of Earth Observation satellites mission ensures uninterrupted optical and radar imagery archives. With the popularization of Unmanned Aerial Vehicles, free optical and radar imagery with high revisiting time, ground and aerial possibilities to perform high-resolution 3D point clouds and derived digital elevation models, it can make it difficult to choose the appropriate method for risk assessment. The aim of this paper is to review the mainstream remote-sensing methods commonly employed for landslide assessment, as well as processing. The purpose is to understand how remote-sensing techniques can be useful for landslide hazard detection and monitoring taking into consideration several constraints such as field location or costs of surveys. First we focus on the suitability of terrestrial, aerial and spaceborne systems that have been widely used for landslide assessment to underline their benefits and drawbacks for data acquisition, processing and interpretation. Several examples of application are presented such as Interferometry Synthetic Aperture Radar (InSAR), lasergrammetry, Terrestrial Optical Photogrammetry. Some of these techniques are unsuitable for slow moving landslides, others limited to large areas and others to local investigations. It can be complicated to select the most appropriate system. Today, the key for understanding landslides is the complementarity of methods and the automation of the data processing. All the mentioned approaches can be coupled (from field monitoring to satellite images analysis) to improve risk management, and the real challenge is to improve automatic solution for landslide recognition and monitoring for the implementation of near real-time emergency systems. © 2020, Springer Nature B.V." "57217873531;6602192723;56763995400;23035517400;35618936400;","A reasoned bibliography on SAR interferometry applications and outlook on big interferometric data processing",2020,"10.1016/j.rsase.2020.100358","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087742572&doi=10.1016%2fj.rsase.2020.100358&partnerID=40&md5=740968a7d1315d62ba1cd43d0daa32f8","In the past few decades, Synthetic Aperture Radar Interferometry (InSAR) has proven to be a reliable tool for monitoring land surface deformations occurring naturally (landslides, earthquakes, and volcanoes) or due to some anthropogenic activities, such as extraction of underground materials (, e.g., groundwater, oil, and gas) with acceptable accuracy. The availability of SAR data from various satellites have significantly improved this technology further notably with collecting data from different radar frequencies (X-, C-, and L-band), different spatial resolutions, increased revisit times and diverse imaging geometry including both along ascending and descending orbits. This review provides a description about the InSAR state-of-the-art technology and how it has been effectively used for detecting surface deformations. The techniques of Persistent Scatterer Interferometry, Small Baseline Subset, Stanford Method for Persistent Scatterers, and Offset Tracking are discussed. The paper also discusses the strengths and weaknesses of the different InSAR techniques currently employed in detecting surface deformations, concerning the various types of land cover. It then highlights the optimal methodology and data needs for these different land cover types. This work finally dives into the emergence of new technologies for processing big Earth Observation data and discusses the prospects of using machine/deep learning algorithms powered by advanced cloud computing infrastructure to mine new information hidden within InSAR products and associated land-surface deformations. © 2020 The Authors" "57214798706;57217418237;55696622200;7401931279;","Cloud removal in Sentinel-2 imagery using a deep residual neural network and SAR-optical data fusion",2020,"10.1016/j.isprsjprs.2020.05.013","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087202507&doi=10.1016%2fj.isprsjprs.2020.05.013&partnerID=40&md5=8a72d38ee85ca627e67fed7d7af07681","Optical remote sensing imagery is at the core of many Earth observation activities. The regular, consistent and global-scale nature of the satellite data is exploited in many applications, such as cropland monitoring, climate change assessment, land-cover and land-use classification, and disaster assessment. However, one main problem severely affects the temporal and spatial availability of surface observations, namely cloud cover. The task of removing clouds from optical images has been subject of studies since decades. The advent of the Big Data era in satellite remote sensing opens new possibilities for tackling the problem using powerful data-driven deep learning methods. In this paper, a deep residual neural network architecture is designed to remove clouds from multispectral Sentinel-2 imagery. SAR-optical data fusion is used to exploit the synergistic properties of the two imaging systems to guide the image reconstruction. Additionally, a novel cloud-adaptive loss is proposed to maximize the retainment of original information. The network is trained and tested on a globally sampled dataset comprising real cloudy and cloud-free images. The proposed setup allows to remove even optically thick clouds by reconstructing an optical representation of the underlying land surface structure. © 2020 The Authors" "55835494600;9737355300;7401516165;","A multiclass TrAdaBoost transfer learning algorithm for the classification of mobile lidar data",2020,"10.1016/j.isprsjprs.2020.05.010","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086503239&doi=10.1016%2fj.isprsjprs.2020.05.010&partnerID=40&md5=a0cbc580923ac3089c512acec4a736ff","A major challenge in the application of state-of-the-art deep learning methods to the classification of mobile lidar data is the lack of sufficient training samples for different object categories. The transfer learning technique based on pre-trained networks, which is widely used in deep learning for image classification, is not directly applicable to point clouds, because pre-trained networks trained by a large number of samples from multiple sources are not available. To solve this problem, we design a framework incorporating a state-of-the-art deep learning network, i.e. VoxNet, and propose an extended Multiclass TrAdaBoost algorithm, which can be trained with complementary training samples from other source datasets to improve the classification accuracy in the target domain. In this framework, we first train the VoxNet model with the combined dataset and extract the feature vectors from the fully connected layer, and then use these to train the Multiclass TrAdaBoost. Experimental results show that the proposed method achieves both improvement in the overall accuracy and a more balanced performance in each category. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57216646857;35205101700;35758658900;57206947615;","Estimation of all-sky all-wave daily net radiation at high latitudes from MODIS data",2020,"10.1016/j.rse.2020.111842","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084191728&doi=10.1016%2fj.rse.2020.111842&partnerID=40&md5=f4660fa34160747bb7b84907fb48491b","Surface all-wave net radiation (Rn) plays an important role in various land surface processes, such as agricultural, ecological, hydrological, and biogeochemical processes. Recently, remote sensing of Rn at regional and global scales has attracted considerable attention and has achieved significant advances. However, there are many issues in estimating all-sky daily average Rn at high latitudes, such as posing greater uncertainty by surface and atmosphere satellite products at high latitudes, and unavailability of real-time and accurate cloud base height and temperature parameters. In this study, we developed the LRD (length ratio of daytime) classification model using the genetic algorithm-artificial neural network (GA-ANN) to estimate all-sky daily average Rn at high latitudes. With a very high temporal repeating frequency (~6 to 20 times per day) at high latitudes, data from the Moderate Resolution Imaging Spectroradiometer (MODIS) were used to test the proposed method. Rn measurements at 82 sites and top-of-atmosphere (TOA) data of MODIS from 2000 to 2017 were matched for model training and validation. Two models for estimating daily average Rn were developed: model I based on instantaneous daytime MODIS observation and model II based on instantaneous nighttime MODIS observation. Validation results of model I showed an R2 of 0.85, an RMSE of 23.66 W/m2, and a bias of 0.27 W/m2, whereas these values were 0.51, 15.04 W/m2, and −0.08 W/m2 for model II, respectively. Overall, the proposed machine learning algorithm with the LRD classification can accurately estimate the all-sky daily average Rn at high latitudes. Mapping of Rn over the high latitudes at 1 km spatial resolution showed a similar spatial distribution to Rn estimates from the Clouds and the Earth's Radiant Energy System (CERES) product. This method has the potential for operational monitoring of spatio-temporal change of Rn at high latitudes with a long-term coverage of MODIS observations. © 2020 The Authors" "57204395597;57200370110;7202446179;","Deep learning-based water quality estimation and anomaly detection using Landsat-8/Sentinel-2 virtual constellation and cloud computing",2020,"10.1080/15481603.2020.1738061","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081744968&doi=10.1080%2f15481603.2020.1738061&partnerID=40&md5=402ce17c1f1112c6266a47520904c40d","Monitoring of inland water quality is of significant importance due to the increase in water quality related issues, especially within the Midwestern United States. Traditional monitoring techniques, although highly accurate, are vastly insufficient in terms of spatial and temporal coverage. Using a virtual constellation by harmonizing Landsat-8 and Sentinel-2 data a high temporal frequency dataset can be created at a relatively fine spatial scale. In this study, we apply a novel deep learning method for the estimation of blue-green algae (BGA), chlorophyll-α (Chl), fluorescent dissolved organic matter (fDOM), dissolved oxygen (DO), specific conductance (SC), and turbidity. The developed model is evaluated against previously studied machine learning methods and found to outperform multiple linear regression (MLR), support vector machine regression (SVR), and extreme learning machine regression (ELR) generating R2 of 0.91 for BGA, 0.88, 0.89, 0.93, 0.87, and 0.84 for Chl, DO, SC, and turbidity respectfully. This model is then applied to all available data ranging from 2013–2018 and time series for each variable were generated for four selected waterbodies. We then use the Empirical Data Analytics (EDA) anomaly detection method on the time series to identify abnormal data points. Upon further analysis, the EDA method successfully identifies abnormal events in water quality. Our results also demonstrate strong correlation between non-optically active variables such as SC with Chl and fDOM. The framework developed in this study represents an efficient and accurate empirical method for inland water quality monitoring at the regional scale. © 2020, © 2020 Informa UK Limited, trading as Taylor & Francis Group." "41961412300;55923441000;56926733400;","iSEA: IoT-based smartphone energy assistant for prompting energy-aware behaviors in commercial buildings",2020,"10.1016/j.apenergy.2020.114892","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082611235&doi=10.1016%2fj.apenergy.2020.114892&partnerID=40&md5=418d45effba214b71641be38fb5ba9f9","Providing personalized energy-use information to individual occupants enables the adoption of energy-aware behaviors in commercial buildings. However, the implementation of individualized feedback still remains challenging due to the difficulties in collecting personalized data, tracking personal behaviors, and delivering personalized tailored information to individual occupants. Nowadays, the Internet of Things (IoT) technologies are used in a variety of applications including real-time monitoring, control, and decision-making due to the flexibility of these technologies for fusing different data streams. In this paper, we propose a novel IoT-based smartphone energy assistant (iSEA) framework which prompts energy-aware behaviors in commercial buildings. iSEA tracks individual occupants through tracking their smartphones, uses a deep learning approach to identify their energy usage, and delivers personalized tailored feedback to impact their usage. iSEA particularly uses an energy-use efficiency index (EEI) to understand behaviors and categorize them into efficient and inefficient behaviors. The iSEA architecture includes four layers: physical, cloud, service, and communication. The results of implementing iSEA in a commercial building with ten occupants over a twelve-week duration demonstrate the validity of this approach in enhancing individualized energy-use behaviors. An average of 34% energy savings was measured by tracking occupants’ EEI by the end of the experimental period. In addition, the results demonstrate that commercial building occupants often ignore controlling over lighting systems at their departure events that leads to wasting energy during non-working hours. By utilizing the existing IoT devices in commercial buildings, iSEA significantly contributes to support research efforts into sensing and enhancing energy-aware behaviors at minimal costs. © 2020 Elsevier Ltd" "57191278865;40761033000;8722458900;23473354200;57212254764;55027020000;12789591800;","Ordinal regression algorithms for the analysis of convective situations over Madrid-Barajas airport",2020,"10.1016/j.atmosres.2019.104798","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076281884&doi=10.1016%2fj.atmosres.2019.104798&partnerID=40&md5=3bd9270331fce94749955b25c8e14d14","In this paper we tackle a problem of convective situations analysis at Adolfo-Suarez Madrid-Barajas International Airport (Spain), based on Ordinal Regression algorithms. The diagnosis of convective clouds is key in a large airport like Barajas, since these meteorological events are associated with strong winds and local precipitation, which may affect air and land operations at the airport. In this work, we deal with a 12-h time horizon in the analysis of convective clouds, using as input variables data from a radiosonde station and also from numerical weather models. The information about the objective variable (convective clouds presence at the airport) has been obtained from the Madrid-Barajas METAR and SPECI aeronautical reports. We treat the problem as an ordinal regression task, where there exist a natural order among the classes. Moreover, the classification problem is highly imbalanced, since there are very few convective clouds events compared to clear days. Thus, a process of oversampling is applied to the database in order to obtain a better balance of the samples for this specific problem. An important number of ordinal regression methods are then tested in the experimental part of the work, showing that the best approach for this problem is the SVORIM algorithm, based on the Support Vector Machine strategy, but adapted for ordinal regression problems. The SVORIM algorithm shows a good accuracy in the case of thunderstorms and Cumulonimbus clouds, which represent a real hazard for the airport operations. © 2019 Elsevier B.V." "55330123800;57208765879;7401793588;56722821200;7405367162;","A machine-learning-based cloud detection and thermodynamic-phase classification algorithm using passive spectral observations",2020,"10.5194/amt-13-2257-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085069222&doi=10.5194%2famt-13-2257-2020&partnerID=40&md5=e15f295336a13d290e6bc9ca72521794","
We trained two Random Forest (RF) machine learning models for cloud mask and cloud thermodynamic-phase detection using spectral observations from Visible Infrared Imaging Radiometer Suite (VIIRS) on board Suomi National Polar-orbiting Partnership (SNPP). Observations from Cloud-Aerosol Lidar with Orthogonal Polarization (CALIOP) were carefully selected to provide reference labels. The two RF models were trained for all-day and daytime-only conditions using a 4-year collocated VIIRS and CALIOP dataset from 2013 to 2016. Due to the orbit difference, the collocated CALIOP and SNPP VIIRS training samples cover a broad-viewing zenith angle range, which is a great benefit to overall model performance. The all-day model uses three VIIRS infrared (IR) bands (8.6, 11, and 12 μm), and the daytime model uses five Near-IR (NIR) and Shortwave-IR (SWIR) bands (0.86, 1.24, 1.38, 1.64, and 2.25 μm) together with the three IR bands to detect clear, liquid water, and ice cloud pixels. Up to seven surface types, i.e., ocean water, forest, cropland, grassland, snow and ice, barren desert, and shrubland, were considered separately to enhance performance for both models. Detection of cloudy pixels and thermodynamic phase with the two RF models was compared against collocated CALIOP products from 2017. It is shown that, when using a conservative screening process that excludes the most challenging cloudy pixels for passive remote sensing, the two RF models have high accuracy rates in comparison to the CALIOP reference for both cloud detection and thermodynamic phase. Other existing SNPP VIIRS and Aqua MODIS cloud mask and phase products are also evaluated, with results showing that the two RF models and the MODIS MYD06 optical property phase product are the top three algorithms with respect to lidar observations during the daytime. During the nighttime, the RF all-day model works best for both cloud detection and phase, particularly for pixels over snow and ice surfaces. The present RF models can be extended to other similar passive instruments if training samples can be collected from CALIOP or other lidars. However, the quality of reference labels and potential sampling issues that may impact model performance would need further attention.
. © 2020 Copernicus GmbH. All rights reserved." "57217170189;","A Random Forest Approach to Identifying Young Stellar Object Candidates in the Lupus Star-forming Region",2020,"10.3847/1538-3881/ab72ac","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086570011&doi=10.3847%2f1538-3881%2fab72ac&partnerID=40&md5=1f2b63174bee7e3cee3c539749ea04bb","The identification and characterization of stellar members within a star-forming region are critical to many aspects of star formation, including formalization of the initial mass function, circumstellar disk evolution, and star formation history. Previous surveys of the Lupus star-forming region have identified members through infrared excess and accretion signatures. We use machine learning to identify new candidate members of Lupus based on surveys from two space-based observatories: ESA's Gaia and NASA's Spitzer. Astrometric measurements from Gaia's Data Release 2 and astrometric and photometric data from the Infrared Array Camera on the Spitzer Space Telescope, as well as from other surveys, are compiled into a catalog for the random forest (RF) classifier. The RF classifiers are tested to find the best features, membership list, non-membership identification scheme, imputation method, training set class weighting, and method of dealing with class imbalance within the data. We list 27 candidate members of the Lupus star-forming region for spectroscopic follow-up. Most of the candidates lie in Clouds V and VI, where only one confirmed member of Lupus was previously known. These clouds likely represent a slightly older population of star formation. © 2020. The American Astronomical Society. All rights reserved.." "57190382179;57201377332;6701404111;6701363691;","Land use/land cover mapping using multitemporal sentinel-2 imagery and four classification methods-A case study from Dak Nong, Vietnam",2020,"10.3390/RS12091367","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085246606&doi=10.3390%2fRS12091367&partnerID=40&md5=1ec5552140b0d12a4853ca695175c8f1","Information on land use and land cover (LULC) including forest cover is important for the development of strategies for land planning and management. Satellite remotely sensed data of varying resolutions have been an unmatched source of such information that can be used to produce estimates with a greater degree of confidence than traditional inventory estimates. However, use of these data has always been a challenge in tropical regions owing to the complexity of the biophysical environment, clouds, and haze, and atmospheric moisture content, all of which impede accurate LULC classification. We tested a parametric classifier (logistic regression) and three non-parametric machine learning classifiers (improved k-nearest neighbors, random forests, and support vector machine) for classification of multi-temporal Sentinel 2 satellite imagery into LULC categories in Dak Nong province, Vietnam. A total of 446 images, 235 from the year 2017 and 211 from the year 2018, were pre-processed to gain high quality images for mapping LULC in the 6516 km2 study area. The Sentinel 2 images were tested and classified separately for four temporal periods: (i) dry season, (ii) rainy season, (iii) the entirety of the year 2017, and (iv) the combination of dry and rainy seasons. Eleven different LULC classes were discriminated of which five were forest classes. For each combination of temporal image set and classifier, a confusion matrix was constructed using independent reference data and pixel classifications, and the area on the ground of each class was estimated. For overall temporal periods and classifiers, overall accuracy ranged from 63.9% to 80.3%, and the Kappa coefficient ranged from 0.611 to 0.813. Area estimates for individual classes ranged from 70 km2 (1% of the study area) to 2200 km2 (34% of the study area) with greater uncertainties for smaller classes. © 2020 by the authors." "57215000725;23012792800;6507322779;8839237600;","Near-daily discharge estimation in high latitudes from Sentinel-1 and 2: A case study for the Icelandic Þjórsá river",2020,"10.1016/j.rse.2020.111684","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079616729&doi=10.1016%2fj.rse.2020.111684&partnerID=40&md5=a975d646f6bb71c849a37353935da14f","Climate change is a threat to many high-latitude regions. Changing patterns in precipitation intensity and increasing glacial ablation during spring and summer have major influence on river dynamics and the risk of widespread flooding. To monitor these rapid events, more frequent discharge observations are necessary. Having access to near-daily satellite based discharge observations is therefore highly beneficial. In this context, the recently launched Sentinel-1 and 2 satellites promise unprecedented potential, due to their capacity to obtain radar and optical data at high spatial (10 m) and high temporal (1–3 days) resolutions. Here, we use both missions to provide a novel approach to estimate the discharge of the Þjórsá (Thjórsá) river, Iceland, on a near-daily basis. Iceland, and many other high-latitude regions, are affected by frequent cloud-cover, limiting the availability of cloud-free optical Sentinel-2 data. We trained a Random Forest supervised machine learning classifier with a set of Sentinel-1 backscatter metrics to classify water in the individual Sentinel-1 images. A Sentinel-2 based classification mask was created to improve the classification results. Second, we derived the river surface area and converted it to the effective width, which we used to estimate the discharge using an at-a-station hydraulic geometry (AHG) rating curve. We trained the rating curve for a six-month training period using in situ discharge observations and assessed the effect of training area selection. We used the trained rating curve to estimate discharge for a one-year monitoring period between 2017/10 and 2018/10. Results showed a Kling-Gupta Efficiency (KGE) of 0.831, indicating the usefulness of dense Sentinel-1 and 2 observations for accurate discharge estimations of a medium-sized (200 m width) high-latitude river on a near-daily basis (1.56 days on average). We demonstrated that satellite based discharge products can be a valuable addition to in situ discharge observations, also during ice-jam events. © 2020 Elsevier Inc." "57191430389;56210720700;57204436627;57211576350;57203553986;26026749200;7003361863;7403247998;7101846027;16308514000;","Stratocumulus cloud clearings: Statistics from satellites, reanalysis models, and airborne measurements",2020,"10.5194/acp-20-4637-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083902538&doi=10.5194%2facp-20-4637-2020&partnerID=40&md5=1c943dc4c0d1800f2336d7e1389e7519","This study provides a detailed characterization of stratocumulus clearings off the US West Coast using remote sensing, reanalysis, and airborne in situ data. Ten years (2009 2018) of Geostationary Operational Environmental Satellite (GOES) imagery data are used to quantify the monthly frequency, growth rate of total area (GRArea), and dimensional characteristics of 306 total clearings. While there is interannual variability, the summer (winter) months experienced the most (least) clearing events, with the lowest cloud fractions being in close proximity to coastal topographical features along the central to northern coast of California, including especially just south of Cape Mendocino and Cape Blanco. From 09:00 to 18:00 (PST), the median length, width, and area of clearings increased from 680 to 1231, 193 to 443, and ~ 67000 to ~ 250000 km2, respectively. Machine learning was applied to identify the most influential factors governing the GRArea of clearings between 09:00 and 12:00 PST, which is the time frame of most rapid clearing expansion. The results from gradient-boosted regression tree (GBRT) modeling revealed that air temperature at 850 hPa (T850), specific humidity at 950 hPa (q950), sea surface temperature (SST), and anomaly in mean sea level pressure (MSLPanom) were probably most impactful in enhancing GRArea using two scoring schemes. Clearings have distinguishing features such as an enhanced Pacific high shifted more towards northern California, offshore air that is warm and dry, stronger coastal surface winds, enhanced lower-tropospheric static stability, and increased subsidence. Although clearings are associated obviously with reduced cloud fraction where they reside, the domain-averaged cloud albedo was actually slightly higher on clearing days as compared to non-clearing days. To validate speculated processes linking environmental parameters to clearing growth rates based on satellite and reanalysis data, airborne data from three case flights were examined. Measurements were compared on both sides of the clear cloudy border of clearings at multiple altitudes in the boundary layer and free troposphere, with results helping to support links suggested by this study s model simulations. More specifically, airborne data revealed the influence of the coastal low-level jet and extensive horizontal shear at cloud-relevant altitudes that promoted mixing between clear and cloudy air. Vertical profile data provide support for warm and dry air in the free troposphere, additionally promoting expansion of clearings. Airborne data revealed greater evidence of sea salt in clouds on clearing days, pointing to a possible role for, or simply the presence of, this aerosol type in clearing areas coincident with stronger coastal winds. © 2020 Copernicus GmbH. All rights reserved." "41762272700;57195920156;56707193700;57211466726;","Introduction of digital technologies in the enterprise",2020,"10.1051/e3sconf/202015904004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084044657&doi=10.1051%2fe3sconf%2f202015904004&partnerID=40&md5=64d5622265e329e6ef88a0c8f25c3b10","Digitalization of an enterprise is necessary to increase the efficiency and sustainability of its functioning through dramatic changes in the quality of management, both technological processes and decision-making processes at all levels of management, based on modern methods of production and further use of information about the state and prediction of possible changes in managed elements and subsystems. For the purposes of designing a digital transformation of an enterprise, it is necessary to develop a classification of digital technologies according to the criterion of accessibility and expediency of their implementation in the enterprise. Thus, key digital technologies are grouped into three groups: basic technologies are technologies without which digital transformation of an enterprise is impossible (cloud technologies, wireless technologies, paperless technologies, etc.); critical technologies are technologies that provide a complete digital transformation of the enterprise (big data, cloud computing, unmanned technologies, etc.); breakthrough technologies - technologies that realize the transition from analog to a digital enterprise (artificial intelligence, neural networks, distributed data registry, machine learning, etc.) © The Authors, published by EDP Sciences, 2020." "55850401500;56422261200;16038426300;6602762273;","Riparian trees genera identification based on leaf-on/leaf-off airborne laser scanner data and machine learning classifiers in northern France",2020,"10.1080/01431161.2019.1674457","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074007989&doi=10.1080%2f01431161.2019.1674457&partnerID=40&md5=a21f3aaf8a3ec2c764e7021d5b6bc1b3","Riparian forests are valuable environments delivering multiples ecological services. Because they face both natural and anthropogenic constraints, riparian forests need to be accurately mapped in terms of genera/species diversity. Previous studies have shown that the Airborne Laser Scanner (ALS) data have the potential to classify trees in different contexts. However, an assessment of important features and classification results for broadleaved deciduous riparian forests mapping using ALS remains to be achieved. The objective of this study was to estimate which features derived from ALS data were important for describing trees genera from a riparian deciduous forest, and provide results of classifications using two Machine Learning algorithms. The procedure was applied to 191 trees distributed in eight genera located along the Sélune river in Normandy, northern France. ALS data from two surveys, in the summer and winter, were used. From these data, trees crowns were extracted and global morphology and internal structure features were computed from the 3D points clouds. Five datasets were established, containing for each one an increasing number of genera. This was implemented in order to assess the level of discrimination between trees genera. The most discriminant features were selected using a stepwise Quadratic Discriminant Analysis (sQDA) and Random Forest, allowing the number of features to be reduced from 144 to 3–9, depending on the datasets. The sQDA-selected features highlighted the fact that, with an increasing number of genera in the datasets, internal structure became more discriminant. The selected features were used as variables for classification using Support Vector Machine (SVM) and Random Forest (RF) algorithms. Additionally, Random Forest classifications were conducted using all features computed, without selection. The best classification performances showed that using the sQDA-selected features with SVM produced accuracy ranging from 83.15% when using three genera (Oak, Alder and Poplar). A similar result was obtained using RF and all features available for classification. The latter also achieved the best classification performances when using seven and eight genera. The results highlight that ML algorithms are suitable methods to map riparian trees. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57218168269;24460797900;7005403704;25225898300;57188835107;8602636100;7003877068;36675777000;57217677002;35185338100;55790076000;57218170381;","Physical properties of the star-forming clusters in NGC 6334",2020,"10.1051/0004-6361/201935699","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086125242&doi=10.1051%2f0004-6361%2f201935699&partnerID=40&md5=d1e1f98aa3e388d9479cbe3a4564e724","Aims. We aim to characterise certain physical properties of high-mass star-forming sites in the NGC 6334 molecular cloud, such as the core mass function (CMF), spatial distribution of cores, and mass segregation. Methods. We used the Atacama Large Millimeter/sub-millimeter Array (ALMA) to image the embedded clusters NGC 6334-I and NGC 6334-I(N) in the continuum emission at 87.6 GHz. We achieved a spatial resolution of 1300 au, enough to resolve different compact cores and fragments, and to study the properties of the clusters. Results. We detected 142 compact sources distributed over the whole surveyed area. The ALMA compact sources are clustered in different regions. We used different machine-learning algorithms to identify four main clusters: NGC 6334-I, NGC 6334-I(N), NGC 6334-I(NW), and NGC 6334-E. The typical separations between cluster members range from 4000 au to 12 000 au. These separations, together with the core masses (0.1-100 M· ), are in agreement with the fragmentation being controlled by turbulence at scales of 0.1 pc. We find that the CMFs show an apparent excess of high-mass cores compared to the stellar initial mass function. We evaluated the effects of temperature and unresolved multiplicity on the derived slope of the CMF. Based on this, we conclude that the excess of high-mass cores might be spurious and due to inaccurate temperature determinations and/or resolution limitations. We searched for evidence of mass segregation in the clusters and we find that clusters NGC 6334-I and NGC 6334-I(N) show hints of segregation with the most massive cores located in the centre of the clusters. Conclusions. We searched for correlations between the physical properties of the four embedded clusters and their evolutionary stage (based on the presence of H II regions and infrared sources). NGC 6334-E appears as the most evolved cluster, already harbouring a well-developed H II region. NGC 6334-I is the second-most evolved cluster with an ultra-compact H II region. NGC 6334-I(N) contains the largest population of dust cores distributed in two filamentary structures and no dominant H II region. Finally, NGC 6334-I(NW) is a cluster of mainly low-mass dust cores with no clear signs of massive cores or H II regions. We find a larger separation between cluster members in the more evolved clusters favouring the role of gas expulsion and stellar ejection with evolution. The mass segregation, seen in the NGC 6334-I and NGC 6334-I(N) clusters, suggests a primordial origin for NGC 6334-I(N). In contrast, the segregation in NGC 6334-I might be due to dynamical effects. Finally, the lack of massive cores in the most evolved cluster suggests that the gas reservoir is already exhausted, while the less evolved clusters still have a large gas reservoir along with the presence of massive cores. In general, the fragmentation process of NGC 6334 at large scales (from filament to clump, i.e. at about 1 pc) is likely governed by turbulent pressure, while at smaller scales (scale of cores and sub-fragments, i.e. a few hundred au) thermal pressure starts to be more significant. © 2020 ESO." "56422909100;57205183542;57201261818;57211978863;57211983696;57218174561;57210827853;8632797000;","An updated moving window algorithm for hourly-scale satellite precipitation downscaling: A case study in the Southeast Coast of China",2020,"10.1016/j.jhydrol.2019.124378","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075561162&doi=10.1016%2fj.jhydrol.2019.124378&partnerID=40&md5=0a86827ffb33c11cd0c3d7c561c56b16","Accurate gridded precipitation products with both finer tempo-spatial resolutions are critical for various scientific communities (e.g., hydrology, meteorology, climatology, and agriculture). Downscaling on coarse satellite-based rainfall estimates is an optimal approach to obtain such datasets. The Integrated Multi-satellitE Retrivals for Global Precipitation Measurement (GPM) (IMERG) data provides the “best” satellite-based precipitation estimates at half-hourly/0.1° scales, while its spatial resolution is still coarse for certain hydrometeorology research. To acquire hourly downscaled precipitation estimates based on IMERG, there are two great challenges: (1) limited rainfall-related environmental variables (0.01°×0.01°, hourly) used to downscale the IMERG data; and (2) far few rainfall pixels used for regressing traditional relationships between precipitation and environmental variables. In this case, most traditional or commonly-used regression/empirical models and the state-of-art machine learning algorithms are not suitable to cater these requirements. Therefore, we proposed a new strategy to obtain hourly downscaled precipitation estimates based on IMERG called Geographically Moving Window Weight Disaggregation Analysis (GMWWDA). Additionally, we explored multiple cloud properties, including cloud effective radius (CER), cloud top height (CTH), cloud top temperature (CTT) and cloud optical thickness (COT), as covariates to downscale IMERG using GMWWDA method, and concluded as follows: (1) the downscaled results (0.01° × 0.01°, hourly) based on the above mentioned cloud properties outperformed the original IMERG data; (2) the downscaled results based on CER performed better than those based on CTH, CTT and COT, respectively; (3) the accuracy of satellite-based precipitation products pose significant effects on those of the downscaled results. This study provides a great potential solution for generating satellite-based precipitation dataset with finer spatio-temporal resolutions. © 2019" "57199457111;57209342660;57202461555;57203961945;56421194700;55455270400;55574709900;57190952440;57216593098;56159151300;36462695500;7402170368;","Estimating forest stock volume in Hunan Province, China, by integrating in situ plot data, Sentinel-2 images, and linear and machine learning regression models",2020,"10.3390/RS12010186","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083997153&doi=10.3390%2fRS12010186&partnerID=40&md5=14d9db74b5883060d3c9a5e1e5121a0a","The forest stock volume (FSV) is one of the key indicators in forestry resource assessments on local, regional, and national scales. To date, scaling up in situ plot-scale measurements across landscapes is still a great challenge in the estimation of FSVs. In this study, Sentinel-2 imagery, the Google Earth Engine (GEE) cloud computing platform, three base station joint differential positioning technology (TBSJDPT), and three algorithms were used to build an FSV model for forests located in Hunan Province, southern China. The GEE cloud computing platform was used to extract the imagery variables from the Sentinel-2 imagery pixels. The TBSJDPT was put forward and used to provide high-precision positions of the sample plot data. The random forests (RF), support vector regression (SVR), and multiple linear regression (MLR) algorithms were used to estimate the FSV. For each pixel, 24 variables were extracted from the Sentinel-2 images taken in 2017 and 2018. The RF model performed the best in both the training phase (i.e., R2 = 0.91, RMSE = 35.13 m3 ha-1, n = 321) and in the test phase (i.e., R2 = 0.58, RMSE = 65.03 m3 ha-1, and n = 138). This model was followed by the SVR model (R2 = 0.54, RMSE = 65.60 m3 ha-1, n = 321 in training; R2 = 0.54, RMSE = 66.00 m3 ha-1, n = 138 in testing), which was slightly better than the MLR model (R2 = 0.38, RMSE = 75.74 m3 ha-1, and n = 321 in training; R2 = 0.49, RMSE = 70.22 m3 ha-1, and n = 138 in testing) in both the training phase and test phase. The best predictive band was Red-Edge 1 (B5), which performed well both in the machine learning methods and in the MLR method. The Blue band (B2), Green band (B3), Red band (B4), SWIR2 band (B12), and vegetation indices (TCW, NDVI_B5, and TCB) were used in the machine learning models, and only one vegetation index (MSI) was used in the MLR model. We mapped the FSV distribution in Hunan Province (3.50 x 108 m3) based on the RF model; it reached a total accuracy of 63.87% compared with the ocial forest report in 2017 (5.48 x 108 m3). The results from this study will help develop and improve satellite-based methods to estimate FSVs on local, regional and national scales. © 2020 by the authors." "55353441700;33768158800;6505760587;","Fusion of sentinel-1 with official topographic and cadastral geodata for crop-type enriched LULC mapping using FOSS and open data",2020,"10.3390/ijgi9020120","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081254350&doi=10.3390%2fijgi9020120&partnerID=40&md5=3f2f8ca6de52c7c1b29a62203c5d9a3e","Accurate crop-type maps are urgently needed as input data for various applications, leading to improved planning and more sustainable use of resources. Satellite remote sensing is the optimal tool to provide such data. Images from Synthetic Aperture Radar (SAR) satellite sensors are preferably used as they work regardless of cloud coverage during image acquisition. However, processing of SAR is more complicated and the sensors have development potential. Dealing with such a complexity, current studies should aim to be reproducible, open, and built upon free and open-source software (FOSS). Thereby, the data can be reused to develop and validate new algorithms or improve the ones already in use. This paper presents a case study of crop classification from microwave remote sensing, relying on open data and open software only. We used 70 multitemporal microwave remote sensing images from the Sentinel-1 satellite. A high-resolution, high-precision digital elevation model (DEM) assisted the preprocessing. The multi-data approach (MDA) was used as a framework enabling to demonstrate the benefits of including external cadastral data. It was used to identify the agricultural area prior to the classification and to create land use/land cover (LULC) maps which also include the annually changing crop types that are usually missing in official geodata. All the software used in this study is open-source, such as the Sentinel Application Toolbox (SNAP), Orfeo Toolbox, R, and QGIS. The produced geodata, all input data, and several intermediate data are openly shared in a research database. Validation using an independent validation dataset showed a high overall accuracy of 96.7% with differentiation into 11 different crop-classes. © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/)." "55939190800;","A robust deep learning approach for spatiotemporal estimation of Satellite AOD and PM2.5",2020,"10.3390/rs12020264","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081079391&doi=10.3390%2frs12020264&partnerID=40&md5=313d1c0e312093219c85c580204b81a4","Accurate estimation of fine particulate matter with diameter =2.5 μm (PM2.5) at a high spatiotemporal resolution is crucial for the evaluation of its health effects. Previous studies face multiple challenges including limited ground measurements and availability of spatiotemporal covariates. Although the multiangle implementation of atmospheric correction (MAIAC) retrieves satellite aerosol optical depth (AOD) at a high spatiotemporal resolution, massive non-random missingness considerably limits its application in PM2.5 estimation. Here, a deep learning approach, i.e., bootstrap aggregating (bagging) of autoencoder-based residual deep networks, was developed to make robust imputation of MAIAC AOD and further estimate PM2.5 at a high spatial (1 km) and temporal (daily) resolution. The base model consisted of autoencoder-based residual networks where residual connections were introduced to improve learning performance. Bagging of residual networks was used to generate ensemble predictions for better accuracy and uncertainty estimates. As a case study, the proposed approach was applied to impute daily satellite AOD and subsequently estimate daily PM2.5 in the Jing-Jin-Ji metropolitan region of China in 2015. The presented approach achieved competitive performance in AOD imputation (mean test R2: 0.96; mean test RMSE: 0.06) and PM2.5 estimation (test R2: 0.90; test RMSE: 22.3 μg/m3). In the additional independent tests using ground AERONET AOD and PM2.5 measurements at the monitoring station of the U.S. Embassy in Beijing, this approach achieved high R2 (0.82-0.97). Compared with the state-of-the-art machine learning method, XGBoost, the proposed approach generated more reasonable spatial variation for predicted PM2.5 surfaces. Publically available covariates used included meteorology, MERRA2 PBLH and AOD, coordinates, and elevation. Other covariates such as cloud fractions or land-use were not used due to unavailability. The results of validation and independent testing demonstrate the usefulness of the proposed approach in exposure assessment of PM2.5 using satellite AOD having massive missing values. © 2020 by the authors." "55578822100;15844957700;56337405600;","Next-Generation Artificial Intelligence Techniques for Satellite Data Processing",2020,"10.1007/978-3-030-24178-0_11","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075504732&doi=10.1007%2f978-3-030-24178-0_11&partnerID=40&md5=257bee2f35e72fcc76f0da6861d4b5df","In this chapter, we have tried to cover majority of the artificial intelligence (AI) techniques that has contributed to the remote sensing community in the form of satellite data processing, right from the basics to advanced level. A wide variety of applications and enormous amount of satellite data growing exponentially has critical demands in speedup, cost cutting, and automation in its processing while maintaining the accuracy. We have started with the need of AI techniques and evolution made for revolutionary changes in remote sensing and other areas. Subsequently, the traditional ML techniques and its limitations, advancements, and need of introducing DL in various applications are reviewed with what is the present requisites and expectation from AI community to overcome the issues and meet the upraised demands by emerging applications. We concluded that ML and DL technology should integrate with big data technologies and cloud computing to meet the future needs. © 2020, Springer Nature Switzerland AG." "35247685500;","Comparison of multi-seasonal Landsat 8, Sentinel-2 and hyperspectral images for mapping forest alliances in Northern California",2020,"10.1016/j.isprsjprs.2019.11.007","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074895939&doi=10.1016%2fj.isprsjprs.2019.11.007&partnerID=40&md5=4455cb8c3ff6720b82ff084131bc1073","The current era of earth observation now provides constellations of open-access, multispectral satellite imagery with medium spatial resolution, greatly increasing the frequency of cloud-free data for analysis. The Landsat satellites have a long historical record, while the newer Sentinel-2 (S2) satellites offer higher temporal, spatial and spectral resolution. The goal of this study was to evaluate the relative benefits of single- and multi-seasonal multispectral satellite data for discriminating detailed forest alliances, as defined by the U.S. National Vegetation Classification system, in a Mediterranean-climate landscape (Sonoma County, California). Results were compared to a companion analysis of simulated hyperspectral satellite data (HyspIRI) for the same study site and reference data (Clark et al., 2018). Experiments used real and simulated S2 and Landsat 8 (L8) data. Simulated S2 and L8 were from HyspIRI images, thereby focusing results on differences in spectral resolution rather than other confounding factors. The Support Vector Machine (SVM) classifier was used in a hierarchical classification of land-cover (Level 1), followed by alliances (Level 2) in forest pixels, and included summer-only and multi-seasonal sets of predictor variables (bands, indices and bands plus indices). Both real and simulated multi-seasonal multispectral variables significantly improved overall accuracy (OA) by 0.2–1.6% for Level 1 tree/no tree classifications and 3.6–25.8% for Level 2 forest alliances. Classifiers with S2 variables tended to be more accurate than L8 variables, particularly for S2, which had 0.4–2.1% and 5.1–11.8% significantly higher OA than L8 for Level 1 tree/no tree and Level 2 forest alliances, respectively. Combining multispectral bands and indices or using just bands was generally more accurate than relying on just indices for classification. Simulated HyspIRI variables from past research had significantly greater accuracy than real L8 and S2 variables, with an average OA increase of 8.2–12.6%. A final alliance-level map used for a deeper analysis used simulated multi-seasonal S2 bands and indices, which had an overall accuracy of 74.3% (Kappa = 0.70). The accuracy of this classification was only 1.6% significantly lower than the best HyspIRI-based classification, which used multi-seasonal metrics (Clark et al., 2018), and there were alliances where the S2-based classifier was more accurate. Within the context of these analyses and study area, S2 spectral-temporal data demonstrated a strong capability for mapping global forest alliances, or similar detailed floristic associations, at medium spatial resolutions (10–30 m). © 2019 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57209285216;56278282600;57209616506;","Verification of two machine learning approaches for cloud masking based on reflectance of channel IR3.9 using Meteosat Second Generation over Middle East maritime",2019,"10.1080/01431161.2019.1624863","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068262019&doi=10.1080%2f01431161.2019.1624863&partnerID=40&md5=e3a268082747fad192facc6fe811c519","Most cold channels of Meteosat Second Generation (MSG) satellites can distinguish between the sea and ice cloud tops, except for the IR3.9 channel because of the close reflectance and radiance values of the IR3.9 channel for maritime, low-level cloud and ice cloud tops. In this article, we introduce and evaluate two machine learning methods for cloud masking of Spinning Enhanced Visible and Infrared Imager (SEVIRI) images in the day and night that use the reflectance value of the IR3.9 channel. We reached a good correlation by comparing the results of the modelled cloud masking of Meteosat satellite images with MODIS (Moderate Resolution Imaging Spectroradiometer) and CLM (Cloud Mask product of EUMETSAT) images in a way that the coefficient of determination (R2) value was 92.34%, 89.91% and 83.69%, 78.23% in the cold season and 90.17%, 87.09% and 80.37%, 76.48% in the warm season, respectively, using the CHAID (chi-squared automatic interaction detection) decision tree and RBF (radial basis function) neural network approaches. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57211316093;55574230638;57208802674;","Convective/stratiform precipitation classification using ground-based doppler radar data based on the K-nearest neighbor algorithm",2019,"10.3390/rs11192277","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073475434&doi=10.3390%2frs11192277&partnerID=40&md5=4c7521e7799af47afa88e35a7d48be73","Stratiform and convective rain types are associated with different cloud physical processes, vertical structures, thermodynamic influences and precipitation types. Distinguishing convective and stratiform systems is beneficial to meteorology research and weather forecasting. However, there is no clear boundary between stratiform and convective precipitation. In this study, a machine learning algorithm, K-nearest neighbor (KNN), is used to classify precipitation types. Six Doppler radar (WSR-98D/SA) data sets from Jiangsu, Guangzhou and Anhui Provinces in China were used as training and classification samples, and the 2A23 product of the Tropical Precipitation Measurement Mission (TRMM) was used to obtain the training labels and evaluate the classification performance. Classifying precipitation types using KNN requires three steps. First, features are selected from the radar data by comparing the range of each variable for different precipitation types. Second, the same unclassified samples are classified with different k values to choose the best-performing k. Finally, the unclassified samples are put into the KNN algorithm with the best k to classify precipitation types, and the classification performance is evaluated. Three types of cases, squall line, embedded convective and stratiform cases, are classified by KNN. The KNN method can accurately classify the location and area of stratiform and convective systems. For stratiform classifications, KNN has a 95% probability of detection, 8% false alarm rate, and 87% cumulative success index; for convective classifications, KNN yields a 78% probability of detection, a 13% false alarm rate, and a 69% cumulative success index. These results imply that KNN can correctly classify almost all stratiform precipitation and most convective precipitation types. This result suggests that KNN has great potential in classifying precipitation types. © 2019 by the authors." "57195412150;57205353363;7401526171;55331455800;7005052907;57208765667;6701549596;35509463200;56012593900;","Conditional generative adversarial networks (cGANs) for near real-time precipitation estimation from multispectral GOES-16 satellite imageries-PERSIANN-cGAN",2019,"10.3390/rs11192193","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073408806&doi=10.3390%2frs11192193&partnerID=40&md5=ee563f287877925796524d0b1da5cda2","In this paper, we present a state-of-the-art precipitation estimation framework which leverages advances in satellite remote sensing as well as Deep Learning (DL). The framework takes advantage of the improvements in spatial, spectral and temporal resolutions of the Advanced Baseline Imager (ABI) onboard the GOES-16 platform along with elevation information to improve the precipitation estimates. The procedure begins by first deriving a Rain/No Rain (R/NR) binary mask through classification of the pixels and then applying regression to estimate the amount of rainfall for rainy pixels. A Fully Convolutional Network is used as a regressor to predict precipitation estimates. The network is trained using the non-saturating conditional Generative Adversarial Network (cGAN) and Mean Squared Error (MSE) loss terms to generate results that better learn the complex distribution of precipitation in the observed data. Common verification metrics such as Probability Of Detection (POD), False Alarm Ratio (FAR), Critical Success Index (CSI), Bias, Correlation and MSE are used to evaluate the accuracy of both R/NR classification and real-valued precipitation estimates. Statistics and visualizations of the evaluation measures show improvements in the precipitation retrieval accuracy in the proposed framework compared to the baseline models trained using conventional MSE loss terms. This framework is proposed as an augmentation for PERSIANN-CCS (Precipitation Estimation from Remotely Sensed Information using Artificial Neural Network- Cloud Classification System) algorithm for estimating global precipitation. © 2019 by the authors." "57148128800;6602369340;7004959591;6602211600;","USING NEURAL NETWORKS to DETECT OBJECTS in MLS POINT CLOUDS BASED on LOCAL POINT NEIGHBORHOODS",2019,"10.5194/isprs-annals-IV-2-W7-17-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084637582&doi=10.5194%2fisprs-annals-IV-2-W7-17-2019&partnerID=40&md5=f574691bf7a6115ddbcce0ef6a734c04","This paper presents an approach which uses a PointNet-like neural network to detect objects of certain types in MLS point clouds. In our case, it is used for the detection of pedestrians, but the approach can easily be adapted to other object classes. In the first step, we process local point neighborhoods with the neural network to determine a descriptive feature. This is then further processed to generate two outputs of the network. The first output classifies the neighborhood and determines if it is part of an object of interest. If this is the case, the second output determines where it is located in relation to the object center. This regression output allows us to use a voting process for the actual object detection. This processing step is inspired by approaches based on implicit shape models (ISM). It is able to deal with a certain amount of incorrectly classified neighborhoods, since it combines the results of multiple neighborhoods for the detection of an object. A benefit of our approach as compared to other machine learning methods is its low demand for training data. In our experiments, we achieved a promising detection performance even with less than 1000 training examples. © 2019 ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. All rights reserved." "36243762400;57209469105;55234747900;19337612500;","Development and validation of a supervised machine learning radar Doppler spectra peak-finding algorithm",2019,"10.5194/amt-12-4591-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071753493&doi=10.5194%2famt-12-4591-2019&partnerID=40&md5=d3441d7b1ef09d9020043aa9f338a1fd","In many types of clouds, multiple hydrometeor populations can be present at the same time and height. Studying the evolution of these different hydrometeors in a time-height perspective can give valuable information on cloud particle composition and microphysical growth processes. However, as a prerequisite, the number of different hydrometeor types in a certain cloud volume needs to be quantified. This can be accomplished using cloud radar Doppler velocity spectra from profiling cloud radars if the different hydrometeor types have sufficiently different terminal fall velocities to produce individual Doppler spectrum peaks. Here we present a newly developed supervised machine learning radar Doppler spectra peak-finding algorithm (named PEAKO). In this approach, three adjustable parameters (spectrum smoothing span, prominence threshold, and minimum peak width at half-height) are varied to obtain the set of parameters which yields the best agreement of user-classified and machine-marked peaks. The algorithm was developed for Ka-band ARM zenith-pointing radar (KAZR) observations obtained in thick snowfall systems during the Atmospheric Radiation Measurement Program (ARM) mobile facility AMF2 deployment at Hyytiälä, Finland, during the Biogenic Aerosols - Effects on Clouds and Climate (BAECC) field campaign. The performance of PEAKO is evaluated by comparing its results to existing Doppler peak-finding algorithms. The new algorithm consistently identifies Doppler spectra peaks and outperforms other algorithms by reducing noise and increasing temporal and height consistency in detected features. In the future, the PEAKO algorithm will be adapted to other cloud radars and other types of clouds consisting of multiple hydrometeors in the same cloud volume. © 2019 Copernicus GmbH. All rights reserved." "57206924573;57209398568;57209691961;","Cloud identification and classification from high spectral resolution data in the far infrared and mid-infrared",2019,"10.5194/amt-12-3521-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068450158&doi=10.5194%2famt-12-3521-2019&partnerID=40&md5=f4cbe18aa76ba93e25ed0670659b3204","A new cloud identification and classification algorithm named CIC is presented. CIC is a machine learning algorithm, based on principal component analysis, able to perform a cloud detection and scene classification using a univariate distribution of a similarity index that defines the level of closeness between the analysed spectra and the elements of each training dataset. CIC is tested on a widespread synthetic dataset of high spectral resolution radiances in the far- and mid-infrared part of the spectrum, simulating measurements from the Fast Track 9 mission FORUM (Far-Infrared Outgoing Radiation Understanding and Monitoring), competing for the ESA Earth Explorer programme, which is currently (2018 and 2019) undergoing industrial and scientific Phase A studies. Simulated spectra are representatives of many diverse climatic areas, ranging from the tropical to polar regions. Application of the algorithm to the synthetic dataset provides high scores for clear or cloud identification, especially when optimisation processes are performed. One of the main results consists of pointing out the high information content of spectral radiance in the far-infrared region of the electromagnetic spectrum to identify cloudy scenes, specifically thin cirrus clouds. In particular, it is shown that hit scores for clear and cloudy spectra increase from about 70 % to 90 % when far-infrared channels are accounted for in the classification of the synthetic dataset for tropical regions. © Author(s) 2019." "55574821400;57195530501;57207107574;21741875100;57213194211;57198137843;56071662500;6701728686;57189843313;57213187691;57200272141;","Applying Machine Learning to Earth Observations in A Standards Based Workflow",2019,"10.1109/IGARSS.2019.8898032","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077684387&doi=10.1109%2fIGARSS.2019.8898032&partnerID=40&md5=6eafffde1b2a76e5bff01fcc4fa8b3a5","Earth Observations (EO) enable scientific research, such as the study of meteorology and climate, ecosystems and forests, hydrology and marine life. Applications of EO help protect populations from disasters and improve life in intelligent cities. Increasingly, Machine Learning techniques are seen as key to solve these complex multidisciplinary problems. The scale and dimensionality of data involved often require the definition of processing chains, or workflows. Standards can facilitate the composition, sharing, execution and discovery of these workflows and applications, making them more useful. This paper presents three applications based on Deep Learning: a tree species classifier, a car detector and a flood detector. These applications rely on software containers to package ML framework and algorithms, as well as on workflows to process EO data. We found that these practices allow improved reuse and deployment of research assets in infrastructures. We also note the strong discriminative capabilities of Deep Learning on smaller datasets and the difficulty of gen-eralization to other methods of sensing or regions of interest. © 2019 IEEE." "57192947904;34869963500;6603354695;","Domain Adaptation of Landsat-8 and Proba-V Data Using Generative Adversarial Networks for Cloud Detection",2019,"10.1109/IGARSS.2019.8899193","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076112397&doi=10.1109%2fIGARSS.2019.8899193&partnerID=40&md5=66c070099c71f0ae9a4abd1dba92fa21","Training machine learning algorithms for new satellites requires collecting new data. This is a critical drawback for most remote sensing applications and specially for cloud detection. A sensible strategy to mitigate this problem is to exploit available data from a similar sensor, which involves transforming this data to resemble the new sensor data. However, even taking into account the technical characteristics of both sensors to transform the images, statistical differences between data distributions still remain. This results in a poor performance of the methods trained on one sensor and applied to the new one. In this this work, we propose to use the generative adversarial networks (GANs) framework to adapt the data from the new satellite. In particular, we use Landsat-8 images, with the corresponding ground truth, to perform cloud detection in Proba-V. Results show that the GANs adaptation significantly improves the detection accuracy. © 2019 IEEE." "22980018800;21742642500;24081888700;","Reconstruction of Cloud Vertical Structure With a Generative Adversarial Network",2019,"10.1029/2019GL082532","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068180615&doi=10.1029%2f2019GL082532&partnerID=40&md5=0129a56da6c98b3418f6a862312fd8a9","We demonstrate the feasibility of solving atmospheric remote sensing problems with machine learning using conditional generative adversarial networks (CGANs), implemented using convolutional neural networks. We apply the CGAN to generating two-dimensional cloud vertical structures that would be observed by the CloudSat satellite-based radar, using only the collocated Moderate-Resolution Imaging Spectrometer measurements as input. The CGAN is usually able to generate reasonable guesses of the cloud structure and can infer complex structures such as multilayer clouds from the Moderate-Resolution Imaging Spectrometer data. This network, which is formulated probabilistically, also estimates the uncertainty of its own predictions. We examine the statistics of the generated data and analyze the response of the network to each input parameter. The success of the CGAN in solving this problem suggests that generative adversarial networks are applicable to a wide range of problems in atmospheric science, a field characterized by complex spatial structures and observational uncertainties. ©2019. American Geophysical Union. All Rights Reserved." "56326408200;57203324978;55739545700;57209326925;57209326451;7102410570;57190186331;55716531300;","Reconstruction of ocean color data using machine learning techniques in polar regions: Focusing on off Cape Hallett, Ross Sea",2019,"10.3390/rs11111366","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067395831&doi=10.3390%2frs11111366&partnerID=40&md5=7065de2352e9d6c9ca9e9490aa744aa5","The most problematic issue in the ocean color application is the presence of heavy clouds, especially in polar regions. For that reason, the demand for the ocean color application in polar regions is increased. As a way to overcome such issues, we conducted the reconstruction of the chlorophyll-a concentration (CHL) data using the machine learning-based models to raise the usability of CHL data. This analysis was first conducted on a regional scale and focused on the biologically-valued Cape Hallett, Ross Sea, Antarctica. Environmental factors and geographical information associated with phytoplankton dynamics were considered as predictors for the CHL reconstruction, which were obtained from cloud-free microwave and reanalysis data. As the machine learning models used in the present study, the ensemble-based models such as Random forest (RF) and Extremely randomized tree (ET) were selectedwith 10-fold cross-validation. As a result, both CHL reconstructions fromthe twomodels showed significant agreement with the standard satellite-derived CHL data. In addition, the reconstructed CHLs were close to the actual CHL value even where it was not observed by the satellites. However, there is a slight difference between the CHL reconstruction results from the RF and the ET, which is likely caused by the difference in the contribution of each predictor. In addition, we examined the variable importance for the CHL reconstruction quantitatively. As such, the sea surface and atmospheric temperature, and the photosynthetically available radiation have high contributions to the model developments. Mostly, geographic information appears to have a lower contribution relative to environmental predictors. Lastly, we estimated the partial dependences for the predictors for further study on the variable contribution and investigated the contributions to the CHL reconstruction with changes in the predictors. © 2019 by the authors." "56282082000;27567950000;6604053069;","Estimating and modeling of spatial variability of volumetric attributes of a Nigerian heavy oil and bitumen deposit",2019,"10.2118/198859-MS","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084015654&doi=10.2118%2f198859-MS&partnerID=40&md5=51d77b8eb28093b26266d911bbb23f64","Deposits of heavy oil and natural bitumen have been long-discovered in the Dahomey basin south-western Nigeria. However, inconsistency in estimates of volumes of hydrocarbon contained in these deposits has inhibited commercial interest in the deposits. The inconsistency is attributable to the little or no consideration for spatial variability in those studies. This work is therefore motivated by the need for spatially-coherent geomodels leading to reliable volumetric estimates. An existing database of porosity, depth-to-top and thickness attributes of a section of the deposits located at Agbabu is the subject of this work. This work conducted exploratory spatial data analysis (ESDA) as well as empirical variogram estimation, interpretation and modeling of the attributes. Here, the estimation and interpretation of empirical variogram faced a number of challenges with potentials to render the estimates uninterpretable, unstable and inconsistent with geologic information. These include presence of spatial outlier data, clusteredness of variogram clouds, data paucity, and irregular distribution of point-pairs on variogram clouds. Spatial outliers were either removed or correlated with existing geologic information. The clusteredness issues were resolved using a machine-learning – aided variogram estimation technique recently proposed. Variogram cloud binning approach was deployed to handle irregular distribution of point-pairs. In attempting to deploy an automatic fitting algorithm, cases of insufficient empirical points leading to lack of convergence were encountered. Such cases were resolved by adopting a combination of manual and automatic fitting approaches. Ultimately, this work presents a three-dimensional anisotropic (zonal) porosity variogram model and two-dimensional anisotropic (geometric) models for the depth-to-top and thickness variograms. These models are suitable inputs to spatial interpolation algorithms in generating maps of these volumetric attributes. Copyright 2019, Society of Petroleum Engineers." "57190173611;55458058500;7402317480;57213185612;","An improved automatic pointwise semantic segmentation of a 3D urban scene from mobile terrestrial and airborne LiDAr point clouds: A machine learning approach",2019,"10.5194/isprs-annals-IV-4-W8-139-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077959523&doi=10.5194%2fisprs-annals-IV-4-W8-139-2019&partnerID=40&md5=1407526b397e77da313f1e7c442d045d","Automatic semantic segmentation of point clouds observed in a 3D complex urban scene is a challenging issue. Semantic segmentation of urban scenes based on machine learning algorithm requires appropriate features to distinguish objects from mobile terrestrial and airborne LiDAR point clouds in point level. In this paper, we propose a pointwise semantic segmentation method based on our proposed features derived from Difference of Normal and the features “directional height above” that compare height difference between a given point and neighbors in eight directions in addition to the features based on normal estimation. Random forest classifier is chosen to classify points in mobile terrestrial and airborne LiDAR point clouds. The results obtained from our experiments show that the proposed features are effective for semantic segmentation of mobile terrestrial and airborne LiDAR point clouds, especially for vegetation, building and ground classes in an airborne LiDAR point clouds in urban areas. © Authors 2019. CC BY 4.0 License." "57189220493;6701598631;6603372854;","Automatic detection of objects in 3D point clouds based on exclusively semantic guided processes",2019,"10.3390/ijgi8100442","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075520183&doi=10.3390%2fijgi8100442&partnerID=40&md5=414aabba8b837dd61f37ad2eb86addc2","In the domain of computer vision, object recognition aims at detecting and classifying objects in data sets. Model-driven approaches are typically constrained through their focus on either a specific type of data, a context (indoor, outdoor) or a set of objects. Machine learning-based approaches are more flexible but also constrained as they need annotated data sets to train the learning process. That leads to problems when this data is not available through the specialty of the application field, like archaeology, for example. In order to overcome such constraints, we present a fully semantic-guided approach. The role of semantics is to express all relevant knowledge of the representation of the objects inside the data sets and of the algorithms which address this representation. In addition, the approach contains a learning stage since it adapts the processing according to the diversity of the objects and data characteristics. The semantic is expressed via an ontological model and uses standard web technology like SPARQL queries, providing great flexibility. The ontological model describes the object, the data and the algorithms. It allows the selection and execution of algorithms adapted to the data and objects dynamically. Similarly, processing results are dynamically classified and allow for enriching the ontological model using SPARQL construct queries. The semantic formulated through SPARQL also acts as a bridge between the knowledge contained within the ontological model and the processing branch, which executes algorithms. It provides the capability to adapt the sequence of algorithms to an individual state of the processing chain and makes the solution robust and flexible. The comparison of this approach with others on the same use case shows the efficiency and improvement this approach brings. © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/)." "6505585734;","Machines’ fault detection and tolerance using big data management",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074587141&partnerID=40&md5=76daadcbcef928b88e04af290742e6a9","Internet of Things (IoT) is one of significant topics in business and in research field. Modern societies are adopting IoT as a necessity rather than a luxury in their daily life style. IoT defined as devices and machines are connected to data repositories which mainly reside in the Cloud. One of the most challenging issues related to IoT environments is the huge amount of data, referred to as Big Data. Data generated from IoT environments are generally stored and processed in cloud infrastructure, then it consumed (in analyzed forms) as input to feed the same IoT environments. IoT and Big Data are two faces of the same coin. This article discusses IoT aspects related to big data in cloud, proposes a Big Data Management Center (BDMC) that utilizes accumulated big data transmitted form machines sensors .BDMC can anticipate faults that might occur in machines, besides it can decide which fault can be tolerated and which fault resulted in severe damage in that machine. This prediction will give us advantage in advance to overcome problems before they occur. This article will also go through important aspects relate to IoT such as “storage as cloud”, “wireless connectivity” and data storage using Hadoop format. Finally, this article presents a case study on fault detection on real vehicles tests, then experimental results is provided. © International Research Publication House." "36491173000;7202269863;55313567200;35612130700;","Using UAV for automatic lithological classification of open pit mining front",2019,"10.1590/0370-44672018720122","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069743611&doi=10.1590%2f0370-44672018720122&partnerID=40&md5=37ccfcf51f3d5cee0a6c6a03e0d70000","Mine planning is dependent on the natural lithologic features and on the definition of their limits. The geological model is constantly updated during the life of the mine, based on all the information collected so far, plus the knowledge developed from the exploration stage up to the mine closure. As the mine progresses, the amount of available data increases, as well as the experience of the geological modeller and mine planner who deliver the short, medium, and long-term plans. This classical approach can benefit from the automation of the geological mapping on the mining faces and outcrops, improving the speed of repetitious work and avoiding exposure to intrinsic dangers like mining equipment, falling rocks, high wall proximity, among others. The use of photogrammetry to keep up with surface mining activities boarded in UAVs is a reality and the automated lithological classification using machine learning techniques is a low-cost evolution that might present accuracies above 90% of the contact zones and lithologies based on the automated dense point cloud classification when compared to the manual (or reality) classified model. © 2019, Escola de Minas. All rights reserved." "41461833800;57209197440;","Implementing the real-time data environment for the oil and gas digital transformation",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066780065&partnerID=40&md5=4e962086b1ef030096523e6c9c5d5980","Real-time operations have long provided a crucial decision support and collaboration capability that enhances decision making during the well construction process. Operators, drilling contractors, and service companies use real-time capabilities to improve operations service quality, create efficiencies, understand formation geology, and enhance overall reservoir knowledge. Historical applications have provided domain experts with real-time information to support less-experienced wellsite personnel, provide advice and guidance, and conduct remotely performed operations. However, the information was typically centered on a single well and often with a discrete job-focused execution mindset, even when the same provider delivered multiple services. The industry focus on “digital transformation” creates new opportunities for the real-time environment. The promise of real-time analytics and machine learning increases the potential value of having access to all petrotechnical data. To be successful, these systems should simultaneously integrate data across two dimensions: all services provided throughout a single asset’s life cycle and all wells across an entire basin. While engineers use historical real-time data for offset analysis during their next well design, data typically remain filed away, relatively inaccessible for large-scale analytics. This paper presents a case history on the development and implementation of an enterprise real-time data environment in support of the digital transformation. Key objectives for the environment are to ensure data accessibility for analytics modeling while maintaining the digital coherence necessary during the data collection and distribution processes across multiple services and wells. Functional capabilities include a multi-well, real-time historian, automated data distribution to a big data environment for full-scale analytics, the capability to run analytics models at the edge, and interconnectivity with other field systems for more context-enriched datasets. Key design principles for success are the use of an open architecture to accelerate innovation and a cloud-first, microservices design to maximize implementation flexibility. Key development learnings are also discussed. © 2019 Offshore Mediterranean Conference (OMC). All rights reserved." "23481467500;23491866500;","A single-tree processing framework using terrestrial laser scanning data for detecting forest regeneration",2019,"10.3390/rs11010060","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059951414&doi=10.3390%2frs11010060&partnerID=40&md5=ddf5e0d43930a30490417285279458cf","Direct assessment of forest regeneration from remote sensing data is a previously little-explored problem. This is due to several factors which complicate object detection of small trees in the understory. Most existing studies are based on airborne laser scanning (ALS) data, which often has insufficient point densities in the understory forest layers. The present study uses plot-based terrestrial laser scanning (TLS) and the survey design was similar to traditional forest inventory practices. Furthermore, a framework of methods was developed to solve the difficulties of detecting understory trees for quantifying regeneration in temperate montane forest. Regeneration is of special importance in our montane study area, since large parts are declared as protection forest against alpine natural hazards. Close to nature forest structures were tackled by separating 3D tree stem detection from overall tree segmentation. In support, techniques from 3D mathematical morphology, Hough transformation and state-of-the-art machine learning were applied. The methodical framework consisted of four major steps. These were the extraction of the tree stems, the estimation of the stem diameters at breast height (DBH), the image segmentation into individual trees and finally, the separation of two groups of regeneration. All methods were fully automated and utilized volumetric 3D image information which was derived from the original point cloud. The total amount of regeneration was split into established regeneration, consisting of trees with a height > 130 cm in combination with a DBH < 12 cm and unestablished regeneration, consisting of trees with a height < 130 cm. Validation was carried out against field-based expert estimates of percentage ground cover, differentiating seven classes that were similar to those used by forest inventory. The mean absolute error (MAE) of our method for established regeneration was 1.11 classes and for unestablished regeneration only 0.27 classes. Considering the metrical distances between the class centres, the MAE amounted 8.08% for established regeneration and 2.23% for unestablished regeneration. © 2019 by the authors." "57193851407;57204945981;7004242319;8573340700;","A Versatile Method for Ice Particle Habit Classification Using Airborne Imaging Probe Data",2018,"10.1029/2018JD029163","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85058022375&doi=10.1029%2f2018JD029163&partnerID=40&md5=8987464a75d3f5988ccaa33eff02a697","A versatile method to automatically classify ice particle habit from various airborne optical array probes is presented. The classification is achieved using a multinomial logistic regression model. For each airborne probe, the model determines the particle habit (among six classes) based on a large set of geometrical and textural descriptors extracted from the two-dimensional image of a particle. The technique is applied and evaluated using three probes with significantly different specifications: the high volume precipitation spectrometer, the two-dimensional stereo probe, and the cloud particle imager. Performance and robustness of the method are assessed using standard machine learning tools on the basis of thousands of images manually labeled for each of the considered probes. The three classifiers show good performance characterized by overall accuracies and Heidke skill scores above 90%. Depending on the application and user preferences, the classification scheme can be easily adapted. For a more precise output, intraclass subclassification can be achieved in a nested fashion, illustrated here with columnar crystals and aggregates. A comparative study of the classification output obtained with the three probes is presented for two aircraft flight periods selected when the three probes were operating together. Results are globally consistent in term of proportions of habit identified (once blurry and partial images have been automatically discarded). A perfect agreement is not expected as the three considered probes are sensitive to different particle size range. ©2018. American Geophysical Union. All Rights Reserved." "7402962962;","Outer space and cyber space: Meeting et in the cloud",2018,"10.1017/S1473550416000318","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84980338898&doi=10.1017%2fS1473550416000318&partnerID=40&md5=951d98ef9ba52c38cd0d9ecec98a7ed5","What justifies the astrobiologist's search for post-biological or machine-intelligence in outer space? Four assumptions borrowed from transhumanism (H+) seem to be at work: (1) it is reasonable to speculate that life on Earth will evolve in the direction of post-biological intelligence; (2) if extraterrestrials have evolved longer than we on Earth, then they will be more scientifically and technologically advanced; (3) superintelligence, computer uploads of brains, and dis-embodied mind belong together; and (4) evolutionary progress is guided by the drive toward increased intelligence. When subjected to critical review, these assumptions prove to be weak. Most importantly, evolutionary biologists do not support the idea that evolution is internally directed toward increased intelligence. Without this assumption, justifying the search for ET more intelligent than earthlings is anaemic. Nevertheless, one can still hope that in the near future we will be communicating with new neighbours in the Milky Way. Can sheer hope inspire science? © 2016 Cambridge University Press." "56673446700;55362143700;56783246900;23005893600;","Classification of image matching point clouds over an urban area",2018,"10.1080/01431161.2018.1452069","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054810160&doi=10.1080%2f01431161.2018.1452069&partnerID=40&md5=b0318c442fe84cc14f3ea0c69ddd745d","Airborne laser scanning (ALS) and image matching are the two main techniques for generating point clouds for large areas. While the classification of ALS point clouds has been well investigated, there are few studies that are related to image matching point clouds. In this study, point clouds of multiple resolutions from high-resolution aerial images (ground sampling distance, GSD, of 6 cm) over the city of Vienna were generated and investigated with respect to point density and processing time. Three different study sites with various urban structures are selected from a bigger dataset and classified based on two different approaches: machine learning and a traditional operator-based decision tree. Classification accuracy was evaluated and compared with confusion matrices. In general, the machine learning method results in a higher overall accuracy compared to the simple decision tree method, with accuracies of 87% and 84%, respectively, at the highest resolution. At lower-resolution levels (GSDs of 12 cm and 24 cm), the overall accuracy of machine learning drops by 4% and that of the simple decision tree by 7% for each level. Classifying rasterized data instead of the original point cloud resulted in an accuracy drop of 5%. Thus, using machine learning on point clouds at the highest available resolution is suggested for classification of urban areas. © 2018 Informa UK Limited, trading as Taylor & Francis Group." "55787237000;49964767000;","Efficient cloud Resource Scaling based on prediction approaches",2018,"10.14419/ijet.v7i3.2.14563","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079962183&doi=10.14419%2fijet.v7i3.2.14563&partnerID=40&md5=058e3f1105aec6afedf975b0f7bbb0a2","Resource Scaling is one of the important job in cloud environment while adapting resource configurations due to elasticity mechanism. In the view of cloud computing, resource scaling mechanism hold the assurance of QoS (Quality of Service), So, one of the key challenging task in cloud environment is, resource scaling. Effective scaling mechanism gives an optimal solutions for computational problems while achieving QoS and avoiding SLA (Service Level Agreement) violations. To enhance resource scaling mechanism in cloud environment, predicting future workload to the each application in different manners like number of physical machines, number of virtual machines, number of requests and resource utilization etc., is an essential step. According to the prediction results, resource scaling can be done in the right time, while preventing QoS dropping and SLA violations. To achieve efficient resource scaling, proposed approach lease advantages of fuzzy time series and machine learning algorithms. The proposed approach is able to reach effective resource scaling mechanism with better results. © 2018 Authors." "57194005615;57215463665;57202322668;57202320981;57202320607;","Handover forecasting in 5G using machine learning",2018,"10.14419/ijet.v7i2.31.13401","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047855373&doi=10.14419%2fijet.v7i2.31.13401&partnerID=40&md5=1e57b2659dbefcac187ba588a8e2e431","Communication plays a major role in human's life. Without network communication can't be done, to achieve proper and effective communication different generations of networks are introduced. Each generation has its own features and perspective of communication, but till now there is no network properly makes people to communicate. Many researches says 5G network will rule the network world as it satisfies all the effective network goals. This paper is proposed to obtain all the goals of a communication network by making proper handover with the help of machine learning. Here we have used two main algorithms to make our 5G handover process by clustering and classifying. Clustering is a process of making the datasets into single units of every users and classification is a process of classifying user's clustered datasets into common path using prediction and forecasting. For clustering we are using K-means and for classification we are using Random Forest algorithm. By using the algorithms the datasets which is being predicted and forecasted is stored in the cloud. Here cloud technology is used as a platform for developing datasets associated with internet. 5G network adapts to any form technology easier and here we have used all the essential technologies under machine learning. This paper deals with all the above methodologies effectively with newer combinations of algorithms along with proper solutions. © 2018 Authors." "16305516000;56192121900;57201487640;","Harmonic extension on the point cloud",2018,"10.1137/16M1098747","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045054370&doi=10.1137%2f16M1098747&partnerID=40&md5=4f8df65d9daaff21e416814eb142c4b7","In this paper, we consider the harmonic extension problem, which is widely used in many applications of machine learning. We formulate the harmonic extension as solving a Laplace–Beltrami equation with Dirichlet boundary condition. We use the point integral method (PIM) proposed in [Z. Li, Z. Shi, and J. Sun, Commun. Comput. Phys., 22 (2017), pp. 228–258; Z. Shi and J. Sun, Res. Math. Sci., to appear; Z. Li and Z. Shi, Multiscale Model. Simul., 14 (2016), pp. 874–905] to solve the Laplace–Beltrami equation. The basic idea of the PIM method is to approximate the Laplace equation using an integral equation, which is easy to discretize from points. Based on the integral equation, we found that the traditional graph Laplacian method (GLM) fails to approximate the harmonic functions near the boundary. One important application of the harmonic extension in machine learning is semisupervised learning. We run a popular semisupervised learning algorithm by Zhu, Ghahramani, and Lafferty [Machine Learning, Proceedings of the Twentieth International Conference (ICML 2003), 2003, Washington, DC, 2003, ARAI Press, Menlo Park, CA pp. 912–919] over a couple of well-known datasets and compare the performance of the aforementioned approaches. Our experiments show the PIM performs the best. We also apply PIM to an image recovery problem and show it outperforms GLM. Finally, on the model problem of the Laplace–Beltrami equation with Dirichlet boundary, we prove the convergence of the point integral method. © 2018 Society for Industrial and Applied Mathematics." "7006792019;48661217200;23027278800;57200600874;7102811204;","Feature extraction and machine learning for the classification of active cropland in the Aral Sea Basin",2017,"10.1109/IGARSS.2017.8127326","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041836347&doi=10.1109%2fIGARSS.2017.8127326&partnerID=40&md5=a0732bcaf557128e13a5d2c882e8302e","Agricultural monitoring over large areas such as the irrigation zones of the Aral Sea Basin (ASB) needs Remote Sensing imagery with a high temporal resolution and a large swath width. Satellite sensors such as MODIS or the upcoming Sentinel-3 suits these requirements. A clear discrimination of active cropland coverage however is impossible using low spatial resolution imagery due to mixed pixel effects with other landcover types. In this paper we utilized Landsat 8 data and present the derivation of the active cropland in the ASB for 3 consecutive years (2014-2016). Each observed year comprises two growing seasons (spring and summer) with 40 Landsat scenes in each season, with a total of 240 Landsat scenes. Due to the lack of adequate in-situ data for remote agricultural regions, a feature extraction method based on image segmentation and k-means clustering is applied to delineate samples of agricultural fields which are forwarded to machine learning algorithms for image classification including. A simple markovian logic is used to introduce prior distribution values the following observation year. The results clearly show, that the extraction of active cropland is possible without the need for time series data, which is limited due to clouds and acquisition date irregularity. Model transferability tests were performed to assess machine learning inference on unseen data with training datasets from adjacent or non-adjacent scenes. The extraction of active cropland for each season also yielded secondary information such as the assessment of land use intensities and crop rotation patterns. © 2017 IEEE." "22235457800;7403531523;52264136000;7004364155;","Determination of CERES TOA fluxes using machine learning algorithms. Part I: Classification and retrieval of CERES cloudy and clear scenes",2017,"10.1175/JTECH-D-16-0183.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85032746166&doi=10.1175%2fJTECH-D-16-0183.1&partnerID=40&md5=ecf69b5675c0551235ccb0203aca6fc7","Continuous monitoring of the earth radiation budget (ERB) is critical to the understanding of Earth's climate and its variability with time. The Clouds and the Earth's Radiant Energy System (CERES) instrument is able to provide a long record of ERB for such scientific studies. This manuscript, which is the first of a two-part paper, describes the new CERES algorithm for improving the clear/cloudy scene classification without the use of coincident cloud imager data. This new CERES algorithm is based on a subset of the modern artificial intelligence (AI) paradigm called machine learning (ML) algorithms. This paper describes the development and application of the ML algorithm known as random forests (RF), which is used to classify CERES broadband footprint measurements into clear and cloudy scenes. Results from the RF analysis carried using the CERES Single Scanner Footprint (SSF) data for January and July are presented in the manuscript. The daytime RF misclassification rate (MCR) shows relatively large values (> 30%) for snow, sea ice, and bright desert surface types, while lower values (< 10%) for the forest surface type. MCR values observed for the nighttime data in general show relatively larger values for most of the surface types compared to the daytime MCR values. The modified MCR values show lower values (< 4%) for most surface types after thin cloud data are excluded from the analysis. Sensitivity analysis shows that the number of input variables and decision trees used in the RF analysis has a substantial influence on determining the classification error. © 2017 American Meteorological Society." "37003444300;55496988000;57218522431;55638787200;56430227700;","Improved online sequential extreme learning machine for simulation of daily reference evapotranspiration",2017,"10.24850/j-tyca-2017-02-12","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85030551493&doi=10.24850%2fj-tyca-2017-02-12&partnerID=40&md5=b6286a0fc06b4ce201639d2c900af0f3","The traditional extreme learning machine has significant disadvantages, including slow training, difficulty in selecting parameters, and difficulty in setting the singularity and the data sample. A prediction model of an improved Online Sequential Extreme Learning Machine (IOS-ELM) of daily reference crop evapotranspiration is therefore examined in this paper. The different manipulation of the inverse of the matrix is made according to the optimal solution and using a regularization factor at the same time in the model. The flexibility of the IOS-ELM in ET0 modeling was assessed using the original meteorological data (Tmax, Tm, Tmin, n, Uh, RHm, φ, Z) of the years 1971–2014 in Yulin, Ankang, Hanzhong, and Xi’an of Shaanxi, China. Those eight parameters were used as the input, while the reference evapotranspiration values were the output. In addition, the ELM, LSSVM, Hargreaves, Priestley-Taylor, Mc Cloud and IOS-ELM models were tested against the FAO-56 PM model by the performance criteria. The experimental results demonstrate that the performance of IOS-ELM was better than the ELM and LSSVM and significantly better than the other empirical models. Furthermore, when the total ET0 estimation of the models was compared by the relative error, the results of the intelligent algorithms were better than empirical models at rates lower than 5%, but the gross ET0 empirical models mainly had 12% to 64.60% relative error. This research could provide a reference to accurate ET0 estimation by meteorological data and give accurate predictions of crop water requirements, resulting in intelligent irrigation decisions in Shaanxi." "56012593900;56701578200;35509463200;57199699111;8308107100;9744927700;57207556152;24468168900;","Deep learning for very high-resolution imagery classification",2017,"10.4324/9781315371740","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051813632&doi=10.4324%2f9781315371740&partnerID=40&md5=f44bc29eff1aaeba3177dd20727f0693","Very high-resolution (VHR) land cover classification maps are needed to increase the accuracy of current land ecosystem and climate model outputs. Limited studies are in place that demonstrate the state-of-the-art in deriving VHR land cover products [1-4]. Additionally, most methods heavily rely on commercial softwares that are difficult to scale given the area of study (e.g., continents to globe). Complexities in present methods relate to (1) scalability of the algorithm, (2) large image data processing (compute and memory intensive), (3) computational cost, (4) massively parallel architecture, and (5) machine learning automation. VHR satellite data sets are of the order of terabytes and features extracted from these data sets are of the order of petabytes. This chapter demonstrates the use of a scalable machine learning algorithm using airborne imagery data acquired by the National Agriculture Imagery Program (NAIP) for the Continental United States (CONUS) at an optimal spatial resolution of 1 m [5]. These data come as image tiles (a total of quarter million image scenes with ~ 60 million pixels) that are multispectral in nature (red, green, blue, and near-infrared [NIR] spectral channels) and have a total size of ~ 60 terabytes for an individual acquisition over CONUS. Features extracted from the entire data set would amount to ~ 8 -10 petabytes. In the proposed approach, a novel semiautomated machine learning algorithm rooted on the principles of “deep learning” is implemented to delineate the percentage of canopy tree cover. In order to perform image analytics in such a granular system, it is mandatory to devise an intelligent archiving and query system for image retrieval, file structuring, metadata processing, and filtering of all available image scenes. This chapter showcases an end-to-end architecture for designing the learning algorithm, namely deep 114belief network (DBN) (stacked restricted Boltzmann machines [RBMs] as an unsupervised classifier) followed by a backpropagation neural network (BPNN) for image classification, a statistical region merging (SRM)-based segmentation algorithm to perform unsupervised segmentation, and a structured prediction framework using conditional random field (CRF) that integrates the results of the classification module and the segmentation module to create the final classification labels. In order to scale this process across quarter million NAIP tiles that cover the entire CONUS, we provide two architectures, one using the National Aeronautics and Space Administration (NASA) high-performance computing (HPC) infrastructure [6,7] and the other using the Amazon Web Services (AWS) cloud compute platform [8]. The HPC framework describes the granular parallelism architecture that can be designed to implement the process across multiple cores with low-to-medium memory requirements in a distributed manner. The AWS framework showcases use-case scenarios of deploying multiple AWS services like the Simple Storage Service (S3) for data storage [9], Simple Queuing Service (SQS) [10] for coordinating the worker nodes and the compute-optimized Elastic Cloud Compute (EC2) [11] along with spot instances for implementing the machine learning algorithm © 2017 by Taylor & Francis Group, LLC." "55100412100;11140546100;57203059237;","Constrained, multi-objective, steamflood injection redistribution optimization, using a cloud-distributed, metamodel-assisted, memetic optimization algorithm",2017,"10.2118/186010-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85050390048&doi=10.2118%2f186010-ms&partnerID=40&md5=0326de0f7be8b3edd43331af42644a11","A cloud-distributed optimization algorithm applicable to large scale, constrained, multiobjective, optimization problems, such as steamflood redistribution, is presented. The proposed algorithm utilizes the so-called Metamodel Assisted Evolutionary Algorithm (MAEA) as its algorithmic basis. MAEAs use a generic implementation of an evolutionary algorithm as their main optimization engine and advanced machine learning techniques as metamodels. Metamodels are utilized through the application of an inexact pre-evaluation phase during the optimization, which substantially decreases the evaluations of the problem specific forward model. Additionally, a unification of global search (GS) and local search (LS) is achieved via the use of Lamarckian learning principles applied on top of a MAEA creating, in essence, a Metamodel Assisted Memetic Algorithm (MAMA). MAMAs profit from the abilities of MAEAs to explore the most promising regions of the design space without being trapped in local optima while also utilizing the efficiency of deterministic methods to further refine promising solutions located during GS. At the end of each EA generation, the most promising members of the current populations are selected to undergo LS using a gradient-based method. Further, integration with scalable cloud-distributed computing allows MAMAs (CD-MAMA) to perform rapid and simultaneous evaluation of tens of thousands of operating plans. The proposed algorithm has been statistically validated on two mathematical test cases and, subsequently, used to optimize a field undergoing steamflood under two different oil-price scenarios. Thus, demonstrating that, cloud-distributed MAMAs coupled with efficient reservoir models, allow for steamflood injection redistribution optimization in affordable, by industry, wallclock times (hours). For the field in question production comes from poorly consolidated sands within the Antelope Shale member of the Miocene Monterey formation with porosity averaging 30%, permeability averaging 2,000 mD and net thicknesses typically between 50 and 300 feet. Structural dip is steep at approximately 60 degrees. The reservoirs are shallow, with depths ranging from 200 - 600 feet TVD. Oil gravity is approximately 13° API. Reservoir pressures are well below bubble point and average 50 - 100 PSI. The field has about 200 producers and 30 injectors, producing about 2000 bpd of oil and injecting 8000bpd of steam. The field has been under operation for about 7 years, with most of the continuous steamflood starting about 3.5 years ago. Copyright 2017, Society of Petroleum Engineers." "6602459839;57211129420;8699469900;","Deep data: Discovery and visualization Application to hyperspectral ALMA imagery",2016,"10.1017/S1743921317000175","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85020016036&doi=10.1017%2fS1743921317000175&partnerID=40&md5=489378aa21e2b752e8765fa6c80b161b","Leading-edge telescopes such as the Atacama Large Millimeter and sub-millimeter Array (ALMA), and near-future ones, are capable of imaging the same sky area at hundreds-to-thousands of frequencies with both high spectral and spatial resolution. This provides unprecedented opportunities for discovery about the spatial, kinematical and compositional structure of sources such as molecular clouds or protoplanetary disks, and more. However, in addition to enormous volume, the data also exhibit unprecedented complexity, mandating new approaches for extracting and summarizing relevant information. Traditional techniques such as examining images at selected frequencies become intractable while tools that integrate data across frequencies or pixels (like moment maps) can no longer fully exploit and visualize the rich information. We present a neural map-based machine learning approach that can handle all spectral channels simultaneously, utilizing the full depth of these data for discovery and visualization of spectrally homogeneous spatial regions (spectral clusters) that characterize distinct kinematic behaviors. We demonstrate the effectiveness on an ALMA image cube of the protoplanetary disk HD142527. The tools we collectively name NeuroScope are efficient for Big Data due to intelligent data summarization that results in significant sparsity and noise reduction. We also demonstrate a new approach to automate our clustering for fast distillation of large data cubes. © Copyright International Astronomical Union 2017." "7006802750;6602337949;6505862984;7202201947;7201527458;14031427400;","Automating the estimation of various meteorological parameters using satellite data and machine learning techniques",2002,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-0036028201&partnerID=40&md5=6ffced3f546df96f93f91360d5bf672f","Satellite data from various sensors and platforms are being used to develop automated algorithms to assist in U.S. Navy operational weather assessment and forecasting. Supervised machine learning techniques are used to discover patterns in the data and develop associated classification and parameter estimation algorithms. These methods are applied to cloud classification in GOES imagery, tropical cyclone intensity estimation using SSM/I data, and cloud ceiling height estimation at remote locations using appropriate geostationary and polar orbiting satellite data in conjunction with numerical weather prediction output and climatology. All developed algorithms rely on training data sets that consist of records of attributes (computed from the appropriate data source) and the associated ground truth." "57204526072;57219866769;57212363342;26026749200;","Comparative Assessment of Snowfall Retrieval From Microwave Humidity Sounders Using Machine Learning Methods",2020,"10.1029/2020EA001357","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85095979478&doi=10.1029%2f2020EA001357&partnerID=40&md5=28fc29b84e8d694595c75dbe2c11b1c2","Accurate quantification of snowfall rate from space is important but has remained difficult. Four years (2007–2010) of NOAA-18 Microwave Humidity Sounder (MHS) data are trained and tested with snowfall estimates from coincident CloudSat Cloud Profiling Radar (CPR) observations using several machine learning methods. Among the studied methods, random forest using MHS (RF-MHS) is found to be the best for both detection and estimation of global snowfall. The RF-MHS estimates are tested using independent years of coincident CPR snowfall estimates and compared with snowfall rates from Modern-Era Retrospective analysis for Research and Applications Version 2 (MERRA-2), Atmospheric Infrared Sounder (AIRS), and MHS Goddard Profiling Algorithm (GPROF). It was found that RF-MHS algorithm can detect global snowfall with approximately 90% accuracy and a Heidke skill score of 0.48 compared to independent CloudSat samples. The surface wet bulb temperatures, brightness temperatures at 190 GHz, and 157 GHz channels are found to be the most important features to delineate snowfall areas. The RF-MHS retrieved global snowfall rates are well compared with CPR estimates and show generally better statistics than MERRA-2, AIRS, and GPROF products. A case study over the United States verifies that the RF-MHS estimated snowfall agrees well with the ground-based National Center for Environmental Prediction (NCEP) Stage-IV and MERRA-2 product, whereas a relatively large underestimation is observed with the current GPROF product (V05). MHS snowfall estimated based on RF algorithm, however, shows some underestimation over cold and snow-covered surfaces (e.g., Greenland, Alaska, and northern Russia), where improvements through new sensors or retrieval techniques are needed. ©2020. The Authors." "57194554240;14519510100;43561035200;43561518700;56684747900;","A hybrid data balancing method for classification of imbalanced training data within google earth engine: Case studies from mountainous regions",2020,"10.3390/rs12203301","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092930622&doi=10.3390%2frs12203301&partnerID=40&md5=9c7b2e0955ced4519f475b763bf4cf8a","Distribution of Land Cover (LC) classes is mostly imbalanced with some majority LC classes dominating against minority classes in mountainous areas. Although standard Machine Learning (ML) classifiers can achieve high accuracies for majority classes, they largely fail to provide reasonable accuracies for minority classes. This is mainly due to the class imbalance problem. In this study, a hybrid data balancing method, called the Partial Random Over-Sampling and Random Under-Sampling (PROSRUS), was proposed to resolve the class imbalance issue. Unlike most data balancing techniques which seek to fully balance datasets, PROSRUS uses a partial balancing approach with hundreds of fractions for majority and minority classes to balance datasets. For this, time-series of Landsat-8 and SRTM topographic data along with various spectral indices and topographic data were used over three mountainous sites within the Google Earth Engine (GEE) cloud platform. It was observed that PROSRUS had better performance than several other balancing methods and increased the accuracy of minority classes without a reduction in overall classification accuracy. Furthermore, adopting complementary information, particularly topographic data, considerably increased the accuracy of minority classes in mountainous areas. Finally, the obtained results from PROSRUS indicated that every imbalanced dataset requires a specific fraction(s) for addressing the class imbalance problem, because different datasets contain various characteristics. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57192257968;7006813492;7005734687;7004563021;","See the forest and the trees: Effective machine and deep learning algorithms for wood filtering and tree species classification from terrestrial laser scanning",2020,"10.1016/j.isprsjprs.2020.08.001","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089223435&doi=10.1016%2fj.isprsjprs.2020.08.001&partnerID=40&md5=d849c8012eddc1006d689bcc5fbb482b","Determining tree species composition in natural forests is essential for effective forest management. Species classification at the individual tree level requires fine-scale traits which can be derived through terrestrial laser scanning (TLS) point clouds. A generalizable species classification framework also needs to decouple seasonal foliage variation from deciduous species, for which wood filtering is applicable. Different machine learning and deep learning models are feasible for wood filtering and species classification. We investigated 13 machine learning and deep learning classifiers for 9 species, and 15 classifiers for filtering wood points from TLS plot scans. Each classifier was evaluated using the criteria of mean Intersection over Union accuracy (mIoU), training stability and time cost. On average, deep learning classifiers outperformed machine learning classifiers by 10% and 5% in terms of wood and species classification mIoU, respectively. PointNet++ provided the best species classifier, with the highest mIoU (0.906), stability, and moderate time cost. Among wood classifiers, UNet achieved the top mIoU (0.839) while ResNet-50 was recommended for rapid trial and error testing. Across the classifications, the factors of input resolution, attributes and features were also analyzed. Hot zones of species classification with PointNet++ were visualized to indicate how AI interpret species traits. © 2020" "55636630900;56724707600;6506773247;35192452600;","Mapping mangrove forest using Landsat 8 to support estimation of land-based emissions in Kenya",2020,"10.1007/s40808-020-00778-x","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086737208&doi=10.1007%2fs40808-020-00778-x&partnerID=40&md5=f62a1d2054da056b1d752ff8192955b4","The Kenyan coast is constantly under persistent cloud cover which hinders mapping using optical images.Up-to-date land-cover information in such areas is sometimes missing from national mapping initiatives.This study uses a computed composite image based on a mean of cloud and shadow free Function of Mask masked multi-temporal Landsat 8 images acquired during long-dry season in a pilot area. We test the effectiveness of the composite to map mangrove forest using random forest (RF) and support vector machines (SVM) machine learning algorithms integrated with context from Markov random fields (MRF(s)).MRFs was chosen because it is computationally efficient hence can be scaled out nationally. The MRF frameworks are compared to pixel-based classification using threefold independent validation samples. SVM–MRFs and RF–MRFs methods have the highest overall accuracy compared to pixel-based classification.However, visual assessment of predicted land-cover using aerial photograghs established that SVM–MRFs framework corresponded well to land-cover in the study area. This framework also managed to map classes with limited ground reference data better than RF–MRFs. Generally, context in both techniques played a discriminative role especially in heterogeneous regions. Therefore, scaling out this approaches would go a long way in generating mangrove forest map inventory in persistent cloud cover regions which is useful for land-based emission estimation. © 2020, Springer Nature Switzerland AG." "57216639492;6602209960;55470317400;36015861500;","Efficient Training of Semantic Point Cloud Segmentation Via Active Learning",2020,"10.5194/isprs-annals-V-2-2020-243-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090845790&doi=10.5194%2fisprs-annals-V-2-2020-243-2020&partnerID=40&md5=685f3acb71db43ebd275aa00a938712f","With the development of LiDAR and photogrammetric techniques, more and more point clouds are available with high density and in large areas. Point cloud interpretation is an important step before many real applications like 3D city modelling. Many supervised machine learning techniques have been adapted to semantic point cloud segmentation, aiming to automatically label point clouds. Current deep learning methods have shown their potentials to produce high accuracy in semantic point cloud segmentation tasks. However, these supervised methods require a large amount of labelled data for proper model performance and good generalization. In practice, manual labelling of point clouds is very expensive and time-consuming. Active learning can iteratively select unlabelled samples for manual annotation based on current statistical models and then update the labelled data pool for next model training. In order to effectively label point clouds, we proposed a segment based active learning strategy to assess the informativeness of samples. Here, the proposed strategy uses 40% of the whole training dataset to achieve a mean IoU of 75.2% which is 99.1% of the accuracy in mIoU obtained from the model trained on the full dataset, while the baseline method using same amount of data only reaches 69.6% in mIoU corresponding to 90.9% of the accuracy in mIoU obtained from the model trained on the full dataset. © 2020 Copernicus GmbH. All rights reserved." "57218546375;56263897000;56922261600;57117329200;55207690500;35552595100;","Building extraction using orthophotos and dense point cloud derived from visual band aerial imagery based on machine learning and segmentation",2020,"10.3390/RS12152397","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089524467&doi=10.3390%2fRS12152397&partnerID=40&md5=b4d9ab1dee9a5f54feb4474ca2e229c6","Urban sprawl related increase of built-in areas requires reliable monitoring methods and remote sensing can be an efficient technique. Aerial surveys, with high spatial resolution, provide detailed data for building monitoring, but archive images usually have only visible bands. We aimed to reveal the efficiency of visible orthophotographs and photogrammetric dense point clouds in building detection with segmentation-based machine learning (with five algorithms) using visible bands, texture information, and spectral and morphometric indices in different variable sets. Usually random forest (RF) had the best (99.8%) and partial least squares the worst overall accuracy (~60%). We found that >95% accuracy can be gained even in class level. Recursive feature elimination (RFE) was an efficient variable selection tool, its result with six variables was like when we applied all the available 31 variables. Morphometric indices had 82% producer's and 85% user's Accuracy (PA and UA, respectively) and combining them with spectral and texture indices, it had the largest contribution in the improvement. However, morphometric indices are not always available but by adding texture and spectral indices to red-green-blue (RGB) bands the PA improved with 12% and the UA with 6%. Building extraction from visual aerial surveys can be accurate, and archive images can be involved in the time series of a monitoring. © 2020 by the authors." "57126877800;57215848943;57217155505;57200555819;55469293700;","DANCE-NET: Density-aware convolution networks with context encoding for airborne LiDAR point cloud classification",2020,"10.1016/j.isprsjprs.2020.05.023","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086503837&doi=10.1016%2fj.isprsjprs.2020.05.023&partnerID=40&md5=d34e9ce6e8055c89f41ef75d493a4157","Airborne LiDAR point cloud classification has been a long-standing problem in photogrammetry and remote sensing. Early efforts either combine hand-crafted feature engineering with machine learning-based classification models or leverage the power of conventional convolutional neural networks (CNNs) on projected feature images. Recent proposed deep learning-based methods tend to develop new convolution operators which can be directly applied on raw point clouds for representative point feature learning. Although these methods have achieved satisfying performance for the classification of airborne LiDAR point clouds, they cannot adequately recognize fine-grained local structures due to the uneven density distribution of 3D point clouds. In this paper, to address this challenging issue, we introduce a density-aware convolution module which uses the point-wise density to reweight the learnable weights of convolution kernels. The proposed convolution module can approximate continuous convolution on unevenly distributed 3D point sets. Based on this convolution module, we further develop a multi-scale CNN model with downsampling and upsampling blocks to perform per-point semantic labeling. In addition, to regularize the global semantic context, we implement a context encoding module to predict a global context encoding and formulated a context encoding regularizer to enforce the predicted context encoding to be aligned with the ground truth one. The overall network can be trained in an end-to-end fashion and directly produces the desired classification results in one network forward pass. Experiments on the ISPRS 3D Labeling Dataset and 2019 Data Fusion Contest Dataset demonstrate the effectiveness and superiority of the proposed method for airborne LiDAR point cloud classification. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "7401931279;","Potential of Large-Scale Inland Water Body Mapping from Sentinel-1/2 Data on the Example of Bavaria’s Lakes and Rivers",2020,"10.1007/s41064-020-00111-2","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086029687&doi=10.1007%2fs41064-020-00111-2&partnerID=40&md5=e1e10e6d28329bbdaccf288598b41194","The mapping of water bodies is an important application area of satellite-based remote sensing. In this contribution, a simple framework based on supervised learning and automatic training data annotation is shown, which allows to map inland water bodies from Sentinel satellite data on large scale, i.e. on state level. Using the German state of Bavaria as an example and different combinations of Sentinel-1 SAR and Sentinel-2 multi-spectral imagery as inputs, potentials and limits for the automatic detection of water surfaces for rivers, lakes, and reservoirs are investigated. Both quantitative and qualitative results confirm that fully automatic large-scale inland water body mapping is generally possible from Sentinel data; whereas, the best result is achieved when all available surface-related bands of both Sentinel-1 and Sentinel-2 are fused on a pixel level. The main limitation arises from missed smaller water bodies, which are not observed in bands with a resolution of about 20 m. Given the simplicity of the proposed approach and the open availability of the Sentinel data, the study confirms the potential for a fully automatic large-scale mapping of inland water with cloud-based remote sensing techniques. © 2020, The Author(s)." "57132509000;23003259600;57189378782;16481561000;55597324400;8721610200;56463389100;57052137700;56375286600;55710921100;57208273241;57131646800;56068376200;","Construction of a virtual PM2.5 observation network in China based on high-density surface meteorological observations using the Extreme Gradient Boosting model",2020,"10.1016/j.envint.2020.105801","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085507158&doi=10.1016%2fj.envint.2020.105801&partnerID=40&md5=b5cce83aca2f79b6b73fa50796103d8d","With increasing public concerns on air pollution in China, there is a demand for long-term continuous PM2.5 datasets. However, it was not until the end of 2012 that China established a national PM2.5 observation network. Before that, satellite-retrieved aerosol optical depth (AOD) was frequently used as a primary predictor to estimate surface PM2.5. Nevertheless, satellite-retrieved AOD often encounter incomplete daily coverage due to its sampling frequency and interferences from cloud, which greatly affect the representation of these AOD-based PM2.5. Here, we constructed a virtual ground-based PM2.5 observation network at 1180 meteorological sites across China using the Extreme Gradient Boosting (XGBoost) model with high-density meteorological observations as major predictors. Cross-validation of the XGBoost model showed strong robustness and high accuracy in its estimation of the daily (monthly) PM2.5 across China in 2018, with R2, root-mean-square error (RMSE) and mean absolute error values of 0.79 (0.92), 15.75 μg/m3 (6.75 μg/m3) and 9.89 μg/m3 (4.53 μg/m3), respectively. Meanwhile, we find that surface visibility plays the dominant role in terms of the relative importance of variables in the XGBoost model, accounting for 39.3% of the overall importance. We then use meteorological and PM2.5 data in the year 2017 to assess the predictive capability of the model. Results showed that the XGBoost model is capable to accurately hindcast historical PM2.5 at monthly (R2 = 0.80, RMSE = 14.75 μg/m3), seasonal (R2 = 0.86, RMSE = 12.28 μg/m3), and annual (R2 = 0.81, RMSE = 10.10 μg/m3) mean levels. In general, the newly constructed virtual PM2.5 observation network based on high-density surface meteorological observations using the Extreme Gradient Boosting model shows great potential in reconstructing historical PM2.5 at ~1000 meteorological sites across China. It will be of benefit to filling gaps in AOD-based PM2.5 data, as well as to other environmental studies including epidemiology. © 2020 The Author(s)" "57216464399;55212747200;56472102300;57216460815;56463389100;36070629400;57217650709;","A robust segmentation framework for closely packed buildings from airborne LiDAR point clouds",2020,"10.1080/01431161.2020.1727053","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083579745&doi=10.1080%2f01431161.2020.1727053&partnerID=40&md5=595e1a9129ff07dad2a2fd1b9d90502c","Urban villages (UVs) are commonly found in many Asian cities. These villages contain many closely packed buildings constructed decades ago without proper urban planning. There is a need for those buildings to be identified and put into statistics. In this paper, we present a segmentation framework that invokes multiple machine learning techniques and point cloud/image processing algorithms to segment individual closely packed buildings from large urban scenes. The presented framework consists of two major segmentation processes. The framework first filters out the non-ground objects from the point cloud, then it classified them by using the Random Forest classifier to isolate buildings from the entire scene. After that, the building point clouds will be segmented based on several building attribute analysis methods. This is followed by using the Random Sample Consensus (RANSAC) plane filtering method to expand the space between two closely packed buildings, so that the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) clustering technique can be used to more accurately segment each individual building from the closely packed building areas. Two airborne Light Detection and Ranging (LiDAR) datasets collected in two different cities with some typical closely packed buildings were used to verify the proposed framework. The results show that the framework can effectively identify the closely packed buildings with unified structures from large airborne LiDAR datasets. The overall segmentation accuracy reaches 84% for the two datasets. The proposed framework can serve as a basis for analysis and segmentation of closely packed buildings with a more complicated structure. © 2020, © 2020 Informa UK Limited, trading as Taylor & Francis Group." "57219411340;56276541300;55974891000;57189520569;57209799235;8835983800;","Learning vector quantization neural network for surface water extraction from Landsat OLI images",2020,"10.1117/1.JRS.14.032605","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092674089&doi=10.1117%2f1.JRS.14.032605&partnerID=40&md5=c67878c68d11e08c36cc76bf92513763","There is a growing concern over surface water dynamics due to an increased understanding of water availability and management with current climate trends. Remote sensing has now become an effective means of water extraction due to the availability of an enormous amount of data with diverse spatial, spectral, and temporal resolutions. However, water extraction from optical remote sensing data is associated with several major difficulties, such as the applicability of the extraction method over large areas and complex environments; shadow contamination from clouds, buildings, and mountains; and disclosure of shadowed water and exclusion of floating and submerged plants. To address these difficulties, a learning vector quantization (LVQ) neural network-based method was proposed and implemented to extract water using Landsat 8 imageries. This method is capable of separating water from clouds, build-up areas, shadows, and shadowed water by the ideal input of bands 1 to 7 and normalized difference vegetation index. This model learns water across Sri Lanka. Eight OLI scenes were tested, and the performance was compared with five widely used machine learning algorithms: support vector machine, K-nearest neighbor, discriminant analysis, combination of modified normalized difference water index and modified fuzzy clustering method, and K-means clustering methods. This method performed the best, achieving overall accuracies and the kappa coefficients between 97.8% and 99.7% and between 0.96 and 0.99, respectively. Results have demonstrated robustness, consistency, and preciseness in various dark surfaces, noisiest water environments, and highly water scarce scenes. LVQ revealed a good generalizing ability to detect all types of water with less amount of training samples. This method can be easily adaptable for other sensors and global water to support water resource studies. © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI." "35179029000;57202511009;57202513399;56269844200;7006360842;","Deep reinforcement learning for the management of software-defined networks and network function virtualization in an Edge-IoT architecture",2020,"10.3390/su12145706","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088836863&doi=10.3390%2fsu12145706&partnerID=40&md5=18e4e88674d12b371cea2bcbe24d4e5b","The Internet of Things (IoT) paradigm allows the interconnection of millions of sensor devices gathering information and forwarding to the Cloud, where data is stored and processed to infer knowledge and perform analysis and predictions. Cloud service providers charge users based on the computing and storage resources used in the Cloud. In this regard, Edge Computing can be used to reduce these costs. In Edge Computing scenarios, data is pre-processed and filtered in network edge before being sent to the Cloud, resulting in shorter response times and providing a certain service level even if the link between IoT devices and Cloud is interrupted. Moreover, there is a growing trend to share physical network resources and costs through Network Function Virtualization (NFV) architectures. In this sense, and related to NFV, Software-Defined Networks (SDNs) are used to reconfigure the network dynamically according to the necessities during time. For this purpose, Machine Learning mechanisms, such as Deep Reinforcement Learning techniques, can be employed to manage virtual data flows in networks. In this work, we propose the evolution of an existing Edge-IoT architecture to a new improved version in which SDN/NFV are used over the Edge-IoT capabilities. The proposed new architecture contemplates the use of Deep Reinforcement Learning techniques for the implementation of the SDN controller. © 2020 by the authors." "57209655847;57204570329;57210343076;57217728843;57210343863;55315026400;9277159100;","Automated 3D reconstruction using optimized view-planning algorithms for iterative development of structure-from-motion models",2020,"10.3390/rs12132169","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087546220&doi=10.3390%2frs12132169&partnerID=40&md5=67dc55d346e11be144c7a74bd4c80582","Unsupervised machine learning algorithms (clustering, genetic, and principal component analysis) automate Unmanned Aerial Vehicle (UAV) missions as well as the creation and refinement of iterative 3D photogrammetric models with a next best view (NBV) approach. The novel approach uses Structure-from-Motion (SfM) to achieve convergence to a specified orthomosaic resolution by identifying edges in the point cloud and planning cameras that ""view"" the holes identified by edges without requiring an initial model. This iterative UAV photogrammetric method successfully runs in various Microsoft AirSim environments. Simulated ground sampling distance (GSD) of models reaches as low as 3.4 cm per pixel, and generally, successive iterations improve resolution. Besides analogous application in simulated environments, a field study of a retired municipal water tank illustrates the practical application and advantages of automated UAV iterative inspection of infrastructure using 63% fewer photographs than a comparable manual flight with analogous density point clouds obtaining a GSD of less than 3 cm per pixel. Each iteration qualitatively increases resolution according to a logarithmic regression, reduces holes in models, and adds details to model edges. © 2020 by the authors." "55236333500;55602536800;35186195400;57195959155;57192094610;8684670500;","Machine learning algorithms to predict tree-related microhabitats using airborne laser scanning",2020,"10.3390/rs12132142","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087544650&doi=10.3390%2frs12132142&partnerID=40&md5=c3cc526a714606d8406481340d0a79d6","In the last few years, the occurrence and abundance of tree-related microhabitats and habitat trees have gained great attention across Europe as indicators of forest biodiversity. Nevertheless, observing microhabitats in the field requires time and well-trained staff. For this reason, new efficient semiautomatic systems for their identification and mapping on a large scale are necessary. This study aims at predicting microhabitats in a mixed and multi-layered Mediterranean forest using Airborne Laser Scanning data through the implementation of a Machine Learning algorithm. The study focuses on the identification of LiDAR metrics useful for detecting microhabitats according to the recent hierarchical classification system for Tree-related Microhabitats, from single microhabitats to the habitat trees. The results demonstrate that Airborne Laser Scanning point clouds support the prediction of microhabitat abundance. Better prediction capabilities were obtained at a higher hierarchical level and for some of the single microhabitats, such as epiphytic bryophytes, root buttress cavities, and branch holes. Metrics concerned with tree height distribution and crown density are the most important predictors of microhabitats in a multi-layered forest. © 2020 by the authors." "57204706166;36627288300;57203752278;7003582919;56519829700;57217530999;24825059000;24825059000;56414357900;7006481399;36100657400;23989037500;23989037500;","Information content of jwst nirspec transmission spectra of warm neptunes",2020,"10.3847/1538-3881/ab9176","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087346234&doi=10.3847%2f1538-3881%2fab9176&partnerID=40&md5=da95a5c6e9d823cec49e0b72ac2aa239","Warm Neptunes offer a rich opportunity for understanding exo-atmospheric chemistry. With the upcoming James Webb Space Telescope (JWST), there is a need to elucidate the balance between investments in telescope time versus scientific yield. We use the supervised machine-learning method of the random forest to perform an information content (IC) analysis on a 11-parameter model of transmission spectra from the various NIRSpec modes. The three bluest medium-resolution NIRSpec modes (0.7-1.27 μm, 0.97-1.84 μm, 1.66-3.07 μm) are insensitive to the presence of CO. The reddest medium-resolution mode (2.87-5.10 μm) is sensitive to all of the molecules assumed in our model: CO, CO2, CH4, C2H2, H2O, HCN, and NH3. It competes effectively with the three bluest modes on the information encoded on cloud abundance and particle size. It is also competitive with the low-resolution prism mode (0.6-5.3 μm) on the inference of every parameter except for the temperature and ammonia abundance. We recommend astronomers to use the reddest medium-resolution NIRSpec mode for studying the atmospheric chemistry of 800-1200 K warm Neptunes; its corresponding high-resolution counterpart offers diminishing returns. We compare our findings to previous JWST IC analyses that favor the blue orders and suggest that the reliance on chemical equilibrium could lead to biased outcomes if this assumption does not apply. A simple, pressure-independent diagnostic for identifying chemical disequilibrium is proposed based on measuring the abundances of H2O, CO, and CO2. © 2020. The Author(s). Published by the American Astronomical Society" "57208411135;35576373500;55234984500;35098639300;23110503000;15081795000;57217492400;57217489905;","Wheat Area Mapping in Afghanistan Based on Optical and SAR Time-Series Images in Google Earth Engine Cloud Environment",2020,"10.3389/fenvs.2020.00077","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087306998&doi=10.3389%2ffenvs.2020.00077&partnerID=40&md5=a40a9eb808b6884de49516a04af61e9d","Wheat is cultivated on more than 2.7 million hectares in Afghanistan annually, yet the country is dependent on imports to meet domestic demand. The timely estimation of domestic wheat production is highly critical to address any potential food security issues and has been identified as a priority by the Ministry of Agriculture Irrigation and Livestock (MAIL). In this study, we developed a system for in-season mapping of wheat crop area based on both optical (Sentinel-2) and synthetic aperture radar (SAR, Sentinel-1) data to support estimation of wheat cultivated area for management and food security planning. Utilizing a 2010 Food and Agriculture Organization (FAO) cropland mask, wheat sown area for 2017 was mapped integrating decision trees and machine learning algorithms in the Google Earth Engine cloud platform. Information from provincial crop calendars in addition to training and validation data from field-based surveys, and high-resolution Digitalglobe and Airbus Pleiades images were used for classification and validation. The total irrigated and rainfed wheat area were estimated as 912,525 and 562,611 ha, respectively for 2017. Province-wise accuracy assessments show the maximum accuracy of irrigated (IR) and rainfed (RF) wheat across provinces was 98.76 and 99%, respectively, whereas the minimum accuracy was found to be 48% (IR) and 73% (RF). The lower accuracy is attributed to the unavailability of reference data, cloud cover in the satellite images and overlap of spectral reflectance of wheat with other crops, especially in the opium poppy growing provinces. While the method is designed to provide estimation at different stages of the growing season, the best accuracy is achieved at the end of harvest using time-series satellite data for the whole season. The approach followed in the study can be used to generate wheat area maps for other years to aid in food security planning and policy decisions. © Copyright © 2020 Tiwari, Matin, Qamer, Ellenburg, Bajracharya, Vadrevu, Rushi and Yusafi." "57217281424;57217281247;16068804700;6507509159;","Generating elevation surface from a single RGB remotely sensed image using deep learning",2020,"10.3390/rs12122002","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087000238&doi=10.3390%2frs12122002&partnerID=40&md5=9ed85b9d4bff82ed6c458fa5d78c05c5","Generating Digital Elevation Models (DEM) from satellite imagery or other data sources constitutes an essential tool for a plethora of applications and disciplines, ranging from 3D flight planning and simulation, autonomous driving and satellite navigation, such as GPS, to modeling water flow, precision farming and forestry. The task of extracting this 3D geometry from a given surface hitherto requires a combination of appropriately collected corresponding samples and/or specialized equipment, as inferring the elevation from single image data is out of reach for contemporary approaches. On the other hand, Artificial Intelligence (AI) and Machine Learning (ML) algorithms have experienced unprecedented growth in recent years as they can extrapolate rules in a data-driven manner and retrieve convoluted, nonlinear one-to-one mappings, such as an approximate mapping from satellite imagery to DEMs. Therefore, we propose an end-to-end Deep Learning (DL) approach to construct this mapping and to generate an absolute or relative point cloud estimation of a DEM given a single RGB satellite (Sentinel-2 imagery in this work) or drone image. The model has been readily extended to incorporate available information from the non-visible electromagnetic spectrum. Unlike existing methods, we only exploit one image for the production of the elevation data, rendering our approach less restrictive and constrained, but suboptimal compared to them at the same time. Moreover, recent advances in software and hardware allow us to make the inference and the generation extremely fast, even on moderate hardware. We deploy Conditional Generative Adversarial networks (CGAN), which are the state-of-the-art approach to image-to-image translation. We expect our work to serve as a springboard for further development in this field and to foster the integration of such methods in the process of generating, updating and analyzing DEMs. © 2020 by the authors." "57191851405;","Coupled online learning as a way to tackle instabilities and biases in neural network parameterizations: General algorithms and Lorenz 96 case study (v1.0)",2020,"10.5194/gmd-13-2185-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084740461&doi=10.5194%2fgmd-13-2185-2020&partnerID=40&md5=e8e1286f7da138978f140bd5ee1b6d5b","Over the last couple of years, machine learning parameterizations have emerged as a potential way to improve the representation of subgrid processes in Earth system models (ESMs). So far, all studies were based on the same three-step approach: first a training dataset was created from a high-resolution simulation, then a machine learning algorithm was fitted to this dataset, before the trained algorithm was implemented in the ESM. The resulting online simulations were frequently plagued by instabilities and biases. Here, coupled online learning is proposed as a way to combat these issues. Coupled learning can be seen as a second training stage in which the pretrained machine learning parameterization, specifically a neural network, is run in parallel with a high-resolution simulation. The high-resolution simulation is kept in sync with the neural network-driven ESM through constant nudging. This enables the neural network to learn from the tendencies that the high-resolution simulation would produce if it experienced the states the neural network creates. The concept is illustrated using the Lorenz 96 model, where coupled learning is able to recover the ""true"" parameterizations. Further, detailed algorithms for the implementation of coupled learning in 3D cloud-resolving models and the super parameterization framework are presented. Finally, outstanding challenges and issues not resolved by this approach are discussed.
. © Author(s) 2020." "57216972010;55719407900;56415300800;57216969459;56720614400;","Mapping rice paddy based on machine learning with sentinel-2 multi-temporal data: Model comparison and transferability",2020,"10.3390/rs12101620","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085583302&doi=10.3390%2frs12101620&partnerID=40&md5=ac14d222ed7eeaa8f56a2ed7e308fd5f","Rice is an important agricultural crop in the Southwest Hilly Area, China, but there has been a lack of efficient and accurate monitoring methods in the region. Recently, convolutional neural networks (CNNs) have obtained considerable achievements in the remote sensing community. However, it has not been widely used in mapping a rice paddy, and most studies lack the comparison of classification effectiveness and efficiency between CNNs and other classic machine learning models and their transferability. This study aims to develop various machine learning classification models with remote sensing data for comparing the local accuracy of classifiers and evaluating the transferability of pretrained classifiers. Therefore, two types of experiments were designed: local classification experiments and model transferability experiments. These experiments were conducted using cloud-free Sentinel-2 multi-temporal data in Banan District and Zhongxian County, typical hilly areas of Southwestern China. A pure pixel extraction algorithm was designed based on land-use vector data and a Google Earth Online image. Four convolutional neural network (CNN) algorithms (one-dimensional (Conv-1D), two-dimensional (Conv-2D) and three-dimensional (Conv-3D_1 and Conv-3D_2) convolutional neural networks) were developed and compared with four widely used classifiers (random forest (RF), extreme gradient boosting (XGBoost), support vector machine (SVM) and multilayer perceptron (MLP). Recall, precision, overall accuracy (OA) and F1 score were applied to evaluate classification accuracy. The results showed that Conv-2D performed best in local classification experiments with OA of 93.14% and F1 score of 0.8552 in Banan District, OA of 92.53% and F1 score of 0.8399 in Zhongxian County. CNN-based models except Conv-1D provided more desirable performance than non-CNN classifiers. Besides, among the non-CNN classifiers, XGBoost received the best result with OA of 89.73% and F1 score of 0.7742 in Banan District, SVM received the best result with OA of 88.57% and F1 score of 0.7538 in Zhongxian County. In model transferability experiments, almost all CNN classifiers had low transferability. RF and XGBoost models have achieved acceptable F1 scores for transfer (RF = 0.6673 and 0.6469, XGBoost = 0.7171 and 0.6709, respectively). © 2020 by the authors." "57200795580;22941140200;55885604100;","Feature selection for airborne LiDAR data filtering: a mutual information method with Parzon window optimization",2020,"10.1080/15481603.2019.1695406","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076344149&doi=10.1080%2f15481603.2019.1695406&partnerID=40&md5=64141d135c23cd9eb64fa12a4ce9aaa1","Filtering is one of the key steps for Digital Elevation Model (DEM) generation from airborne Light Detection and Ranging (LiDAR) data. Machine-learning-based filters have emerged as a class of filtering algorithms in recent years. Most existing studies mainly focus on feature generation due to limited available features a point cloud possesses. More than 30 features have been described in the existing literature. But most generated features are based on geometric information of points. Several redundant and irrelevant features may not necessarily improve the filtering accuracy. Hence, this paper proposes a feature-selection method using minimal-Redundancy-Maximal-Relevance (mRMR) combined with Parzen window optimization to deal with both discrete and continuous features. An optimal/suboptimal feature subset is constructed for machine-learning filters in various landscapes. Experimental results based on AdaBoost show that height-related features, particularly height itself, are of the greatest significance in both urban and rural scenes. Moreover, different subsets can be selected from the datasets of the two landscapes by our feature-selection strategy, which increases the data relevance for describing each geographical landscape. This study provides guidelines for the selection of optimal/suboptimal features for point cloud filtering based on machine-learning algorithms. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57194612091;57195695111;55362143700;23005893600;52263795700;57216910936;57216910905;7003276131;57216911316;23397202200;15059097100;","Generalized sparse convolutional neural networks for semantic segmentation of point clouds derived from tri-stereo satellite imagery",2020,"10.3390/RS12081289","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085257464&doi=10.3390%2fRS12081289&partnerID=40&md5=f0f82bc389ed2df4fba4ff7e57aee7cb","We studied the applicability of point clouds derived from tri-stereo satellite imagery for semantic segmentation for generalized sparse convolutional neural networks by the example of an Austrian study area. We examined, in particular, if the distorted geometric information, in addition to color, influences the performance of segmenting clutter, roads, buildings, trees, and vehicles. In this regard, we trained a fully convolutional neural network that uses generalized sparse convolution one time solely on 3D geometric information (i.e., 3D point cloud derived by dense image matching), and twice on 3D geometric as well as color information. In the first experiment, we did not use class weights, whereas in the second we did. We compared the results with a fully convolutional neural network that was trained on a 2D orthophoto, and a decision tree that was once trained on hand-crafted 3D geometric features, and once trained on hand-crafted 3D geometric as well as color features. The decision tree using hand-crafted features has been successfully applied to aerial laser scanning data in the literature. Hence, we compared our main interest of study, a representation learning technique, with another representation learning technique, and a non-representation learning technique. Our study area is located in Waldviertel, a region in Lower Austria. The territory is a hilly region covered mainly by forests, agriculture, and grasslands. Our classes of interest are heavily unbalanced. However, we did not use any data augmentation techniques to counter overfitting. For our study area, we reported that geometric and color information only improves the performance of the Generalized Sparse Convolutional Neural Network (GSCNN) on the dominant class, which leads to a higher overall performance in our case. We also found that training the network with median class weighting partially reverts the effects of adding color. The network also started to learn the classes with lower occurrences. The fully convolutional neural network that was trained on the 2D orthophoto generally outperforms the other two with a kappa score of over 90% and an average per class accuracy of 61%. However, the decision tree trained on colors and hand-crafted geometric features has a 2% higher accuracy for roads. © 2020 by the authors." "57054207300;55780256400;","Integrating landsat-8 and sentinel-2 time series data for yield prediction of sugarcane crops at the block level",2020,"10.3390/RS12081313","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084642257&doi=10.3390%2fRS12081313&partnerID=40&md5=f7b5109e380d54bf42c52c8ae03c654c","Early prediction of sugarcane crop yield at the commercial block level (unit area of a single crop of the same variety, ratoon or planting date) offers significant benefit to growers, consultants, millers, policy makers, crop insurance companies and researchers. This current study explored a remote sensing based approach for predicting sugarcane yield at the block level by further developing a regionally specific Landsat time series model and including individual crop sowing (or previous seasons' harvest) date. For the Bundaberg growing region of Australia this extends over a five months period, from July to November. For this analysis, the sugarcane blocks were clustered into 10 groups based on their specific planting or ratoon commencement date within the specified five months period. These clustered or groups of blocks were named 'bins'. Cloud free (<20%) satellite data from the polar-orbiting Landsat-8 (launched 2013), Sentinel-2A (launched 2015) and Sentinel-2B (launched 2017) sensors were acquired over the cane growing region in Bundaberg (area of 32,983 ha), from the growing season starting in July 2014, with the average green normalised difference vegetation index (GNDVI) derived for each block. The number of images acquired for each season was defined by the number of cloud free acquisitions. Using the Simple Linear Machine Learning (ML) algorithm, the extracted Landsat derived GNDVI values for each of the blocks were converted to Sentinel GNDVI. The average GNDVI of each 'bin' was plotted and a quadratic model was fitted through the time series to identify the peak growth stage defined as the maximum GNDVI value. The model derived maximum GNDVI values for each of the bins were then regressed against the average actual yield (t·ha-1) achieved for the respective bin over the five growing years, producing strong correlations (R2 = 0.92 to 0.99). The quadratic curves developed for the different bins were shifted according to the specific planting or ratoon date of an individual block allowing for the peak GNDVI value of the block to be calculated, regressed against the actual block yield (t·ha-1) and the prediction of yield to be made. To validate the accuracies of the 10 time series algorithms representing each of the 10 bins, 592 individual blocks were selected from the Bundaberg region during the 2019 harvest season. The crops were clustered into the appropriate bins with the respective algorithm applied. From a Sentinel image acquired on the 5 May 2019, the prediction accuracies were encouraging (R2 = 0.87 and RMSE = 11.33 (t·ha-1)) when compared to actual harvested yield, as reported by the mill. The results presented in this paper demonstrate significant progress in the accurate prediction of sugarcane yield at the individual sugarcane block level using a remote sensing, time-series based approach. © 2020 by the authors." "57216790729;57216787681;57205495136;14053992900;55848021900;48261068400;50560994200;57216786213;55273559900;","Retrieval of aerodynamic parameters in rubber tree forests based on the computer simulation technique and terrestrial laser scanning data",2020,"10.3390/RS12081318","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084638859&doi=10.3390%2fRS12081318&partnerID=40&md5=90da3ce53ed63eb29857da3067ab3241","Rubber trees along the southeast coast of China always suffer severe damage from hurricanes. Quantitative assessments of the capacity for wind resistance of various rubber tree clones are currently lacking. We focus on a vulnerability assessment of rubber trees of different clones under wind disturbance impacts by employing multidisciplinary approaches incorporating scanned points, aerodynamics, machine learning and computer graphics. Point cloud data from two typical rubber trees belonging to different clones (PR107 and CATAS 7-20-59) were collected using terrestrial laser scanning, and a connection chain of tree skeletons was constructed using a clustering algorithm of machine learning. The concept of foliage clumps based on the trunk and first-order branches was first proposed to optimize rubber tree plot 3D modelling for simulating the wind field and assessing the wind-related parameters. The results from the obtained phenotypic traits show that the variable leaf area index and included angle between the branches and trunk result in variations in the topological structure and gap fraction of tree crowns, respectively, which are the major influencing factors relevant to the rubber tree's capacity to resist hurricane strikes. The aerodynamics analysis showed that the maximum dynamic pressure, wind velocity and turbulent intensity of the wind-related parameters in rubber tree plots of clone PR107 (300 Pa, 30 m/s and 15%) are larger than that in rubber tree plots of clone CATAS-7-20-59 (120 Pa, 18 m/s and 5%), which results in a higher probability of local strong cyclone occurrence and a higher vulnerability to hurricane damage. © 2020 by the authors." "55694702700;26423907800;56960514500;56368863100;57208054851;","Harmonized landsat 8 and sentinel-2 time series data to detect irrigated areas: An application in Southern Italy",2020,"10.3390/RS12081275","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084588520&doi=10.3390%2fRS12081275&partnerID=40&md5=7c8cb13264f96429dca8af2c17bca9b6","Lack of accurate and up-to-date data associated with irrigated areas and related irrigation amounts is hampering the full implementation and compliance of theWater Framework Directive (WFD). In this paper, we describe the framework that we developed and implemented within the DIANA project to map the actual extent of irrigated areas in the Campania region (Southern Italy) during the 2018 irrigation season. For this purpose, we considered 202 images from the Harmonized Landsat Sentinel-2 (HLS) products (57 images from Landsat 8 and 145 images from Sentinel-2). Such data were preprocessed in order to extract a multitemporal Normalized Difference Vegetation Index (NDVI) map, which was then smoothed through a gap-filling algorithm. We further integrated data coming from high-resolution (4 km) global satellite precipitation Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks (PERSIANN)-Cloud Classification System (CCS) products. We collected an extensive ground truth in the field represented by 2992 data points coming from three main thematic classes: bare soil and rainfed (class 0), herbaceous (class 1), and tree crop (class 2). This information was exploited to generate irrigated area maps by adopting a machine learning classification approach. We compared six different types of classifiers through a cross-validation approach and found that, in general, random forests, support vector machines, and boosted decision trees exhibited the best performances in terms of classification accuracy and robustness to different tested scenarios. We found an overall accuracy close to 90% in discriminating among the three thematic classes, which highlighted promising capabilities in the detection of irrigated areas from HLS products. © 2020 by the authors." "57192504939;57218174545;56181390800;57202231708;57200511212;","Convective Clouds Extraction from Himawari-8 Satellite Images Based on Double-Stream Fully Convolutional Networks",2020,"10.1109/LGRS.2019.2926402","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082885920&doi=10.1109%2fLGRS.2019.2926402&partnerID=40&md5=c8b5406d9ff30a9740eed2e05158380e","Auto-extraction of convective clouds is of great significance. Convective clouds often bring heavy rain, strong winds, and other disastrous weather. Early warning of convection can effectively reduce loss. Using remote sensing images, we can get large-scale cloud information, which provides many effective methods for convective clouds detection. In this letter, we proposed a novel method to extract convective clouds. We introduce a novel deep network using only 1\times 1 convolution (3ONet) to extract the spectral characteristics. We then combine a 3ONet with the symmetrical dense-shortcut deep fully convolutional networks (SDFCNs) with a double-stream fully convolutional network to extract convective clouds. In the experiment, we used 12 000 Himawari-8 satellite image patches to verify the proposed framework. Experimental results with 0.5882 mean intersection over union (mIOU) pointed out the proposed method can extract convective clouds effectively. © 2004-2012 IEEE." "57215001908;35316130500;55882914500;34968949700;12144303500;23667564700;","A multi-resolution air temperature model for France from MODIS and Landsat thermal data",2020,"10.1016/j.envres.2020.109244","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079681488&doi=10.1016%2fj.envres.2020.109244&partnerID=40&md5=1a0e37aacb8aea9703499533265aa401","Understanding and managing the health effects of ambient temperature (Ta) in a warming, urbanizing world requires spatially- and temporally-resolved Ta at high resolutions. This is challenging in a large area like France which includes highly variable topography, rural areas with few weather stations, and heterogeneous urban areas where Ta can vary at fine spatial scales. We have modeled daily Ta from 2000 to 2016 at a base resolution of 1 km2 across continental France and at a 200 × 200 m2 resolution over large urban areas. For each day we predict three Ta measures: minimum (Tmin), mean (Tmean), and maximum (Tmax). We start by using linear mixed models to calibrate daily Ta observations from weather stations with remotely sensed MODIS land surface temperature (LST) and other spatial predictors (e.g. NDVI, elevation) on a 1 km2 grid. We fill gaps where LST is missing (e.g. due to cloud cover) with additional mixed models that capture the relationship between predicted Ta at each location and observed Ta at nearby weather stations. The resulting 1 km Ta models perform very well, with ten-fold cross-validated R2 of 0.92, 0.97, and 0.95, mean absolute error (MAE) of 1.4 °C, 0.9 °C, and 1.4 °C, and root mean square error (RMSE) of 1.9 °C, 1.3 °C, and 1.8 °C (Tmin, Tmean, and Tmax, respectively) for the initial calibration stage. To increase the spatial resolution over large urban areas, we train random forest and extreme gradient boosting models to predict the residuals (R) of the 1 km Ta predictions on a 200 × 200 m2 grid. In this stage we replace MODIS LST and NDVI with composited top-of-atmosphere brightness temperature and NDVI from the Landsat 5, 7, and 8 satellites. We use a generalized additive model to ensemble the random forest and extreme gradient boosting predictions with weights that vary spatially and by the magnitude of the predicted residual. The 200 m models also perform well, with ten-fold cross-validated R2 of 0.79, 0.79, and 0.85, MAE of 0.4, 0.3, and 0.3, and RMSE of 0.6, 0.4, and 0.5 (Rmin, Rmean, and Rmax, respectively). Our model will reduce bias in epidemiological studies in France by improving Ta exposure assessment in both urban and rural areas, and our methodology demonstrates that MODIS and Landsat thermal data can be used to generate gap-free timeseries of daily minimum, maximum, and mean Ta at a 200 × 200 m2 spatial resolution. © 2020 Elsevier Inc." "57203220499;55576827900;56674284600;","Generalization considerations and solutions for point cloud hillslope classifiers",2020,"10.1016/j.geomorph.2020.107039","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077919094&doi=10.1016%2fj.geomorph.2020.107039&partnerID=40&md5=d46d20aa106baf8c869298420a83d4d2","Point cloud classifiers have the potential to rapidly perform landscape characterization for a variety of applications. The generalization (i.e., transferability to new sites) of such classifiers could improve their accessibility and usefulness for both engineers and researchers alike, but guidelines for classifier generalization are lacking in the literature. This study develops and applies a Random Forest machine learning classifier for Terrestrial Laser Scanning (TLS) point clouds, and generalizes the classifier to point clouds from several different locations. The classifier is trained to identify basic hillslope topographic features, including vegetation, soil, talus, and bedrock using multi-scale geometric features of the point cloud. Four rock and soil slopes in western Colorado were scanned using TLS. Generalization experiments were performed testing point density, occlusion, and between-site domain variance factors, and all factors showed a significant influence on generalization accuracy. Several methods for improving classifier generalization accuracy were tested and compared, including combining training data from multiple sites, imposing probability thresholds, and a Domain Adaptation methodology known as Active Learning. It was found that incorporating data from multiple sites resulted in improved generalization accuracy, but in most cases the largest improvements in accuracy were associated with adding new training data from the target site. In this case, using Active Learning resulted in significant accuracy improvements with an over 90% reduction in the number of added training points. The results suggest that scanning characteristics are important factors in classifier generalization accuracy, but their effects can be mitigated by using the techniques described herein. © 2020 Elsevier B.V." "57202218489;57205157831;57192717976;15060638500;","Review of empirical solar radiation models for estimating global solar radiation of various climate zones of China",2020,"10.1177/0309133319867213","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073822617&doi=10.1177%2f0309133319867213&partnerID=40&md5=7e5e3de90ff384db8f067e947aea2561","Utilizing solar energy requires accurate information about global solar radiation (GSR), which is critical for designers and manufacturers of solar energy systems and equipment. This study aims to examine the literature gaps by evaluating recent predictive models and categorizing them into various groups depending on the input parameters, and comprehensively collect the methods for classifying China into solar zones. The selected groups of models include those that use sunshine duration, temperature, dew-point temperature, precipitation, fog, cloud cover, day of the year, and different meteorological parameters (complex models). 220 empirical models are analyzed for estimating the GSR on a horizontal surface in China. Additionally, the most accurate models from the literature are summarized for 115 locations in China and are distributed into the above categories with the corresponding solar zone; the ideal models from each category and each solar zone are identified. Comments on two important temperature-based models that are presented in this work can help the researchers and readers to be unconfused when reading the literature of these models and cite them in a correct method in future studies. Machine learning techniques exhibit performance GSR estimation better than empirical models; however, the computational cost and complexity should be considered at choosing and applying these techniques. The models and model categories in this study, according to the key input parameters at the corresponding location and solar zone, are helpful to researchers as well as to designers and engineers of solar energy systems and equipment. © The Author(s) 2019." "57218571650;57199227774;","Diagnosis of Parkinson disease using sensor data and machine learning approach in mobile cloud",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089606302&partnerID=40&md5=e602e5daee888cb3fb8987fe25f2c8e3","In reality, the utilization of sensors is developing each day to improve the personal satisfaction by giving medicinal services data on clinical diagnostics. There are various sensors like detecting advance sensors, ""electronic gadgets"" physical sensors have been effectively shown in the field of biomedical applications because of sensor great working ability. Cloud computing technology is used to accommodate vast amounts of data. Such medical applications also use the cloud computing platform to store and access the data protected. Parkinson's disease (PD) could also be a chronic long condition of the central systema nervosum that primarily affects the motor system of the patient. The side effects of Parkinsons’ sickness incorporate muscle rigidity, tremors and modifications in speech. The objective of the framework is to analyse the parkinson’s disease in voice detection. As the indications are worsen, the patients motor and non-motor system will get failure. The framework is used for diagnosing the parkinson’s disease. The proposed system concentrates on improving the Parkinson's disease diagnosis using voice signals from patient detected from different sensors and uploaded to the cloud for processing. This frame work concentrates on improving disease diagnosis with experimental results using Support Vector Machine, Logistic Regression, Random Forest and eXtreme Gradient Boosting.Machine learning algorithm mainly concentrates on improving the prediction of parkinson’s disease diagnosis. © 2020 Alpha Publishers. All rights reserved." "57217412518;","Beyond earthworm: Keeping the promise",2020,"10.1785/0220190198","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089527994&doi=10.1785%2f0220190198&partnerID=40&md5=1dfedaec6042d4d4e631301371124920","The founding of the Advanced National Seismic System (ANSS) vision was originally presented in U.S. Geological Survey Circular 1188 (U.S. Geological Survey [USGS], 1999), after many years of discussions and workshops, described in detail by Filson and Arabasz (2016). Much has been accomplished in the ensuing two decades. Disparate and sometimes divergent developments that had been previously explored at individual private and public universities were finally centralized with increased efficiency and coherency of effort. The stated mission of the ANSS is to ""... provide accurate and timely data and information products for seismic events, including their effects on buildings and structures, employing modern monitoring methods and technologies.""In this article, an approach (xQuake) is proposed that does not interfere in any way with the mission of the National Earthquake Information Center and ANSS but instead restores much of the community focus and international collaboration that has been lost over the past two decades. xQuake uses an executable graph framework in a pipeline architecture; this framework can be seamlessly integrated into current ANSS quake monitoring systems. This newapproach incorporatesmodern approaches to computer analytics, includingmultitopic Kafka exchange rings, cloud computing, a self-configuring phase associator, and machine learning. The xGraph system is free for noncommercial use, open source, hardware agnostic (Windows, Linux, Mac), with no requirement for commercial datastores. © Seismological Society of America." "55576700800;57205299839;57200596843;35098748100;54785899100;","All-sky longwave downward radiation from satellite measurements: General parameterizations based on LST, column water vapor and cloud top temperature",2020,"10.1016/j.isprsjprs.2020.01.011","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077787169&doi=10.1016%2fj.isprsjprs.2020.01.011&partnerID=40&md5=b6042f028815807da31f4dbcd4f4c7c2","Remotely sensed surface longwave downward radiation (LWDR) plays an essential role in studying the surface energy budget and greenhouse effect. Most existing satellite-based methods or products depend on variables that are not readily available from space such as, liquid water path, air temperature, vapor pressure and/or cloud-base temperature etc., which seriously restrict the wide applications of satellite data. In this paper, new nonlinear parameterizations and a machine learning-based model for deriving all-sky LWDR are proposed based only on land surface temperature (LST), column water vapor and cloud-top temperature (CTT), that are relatively readily available day and night for most satellite missions. It is the first time to incorporate the CTT in the parameterizations for estimating LWDR under the cloudy-sky conditions. The results reveal that the new models work well and can derive all-sky global LWDR with reasonable accuracies (RMSE <23 W/m2, bias <2.0 W/m2). The convenience of input data makes the new models easy to use, and thus will definitely expand the applicability of remotely sensed measurements in radiation budget fields and many land applications. © 2020" "56768656200;57214920823;55986546100;36663710700;57211922206;55934244200;","Discriminative feature learning constrained unsupervised network for cloud detection in remote sensing imagery",2020,"10.3390/rs12030456","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85080882325&doi=10.3390%2frs12030456&partnerID=40&md5=afa54c8f3017e4109c318b6d2596b840","Cloud detection is a significant preprocessing step for increasing the exploitability of remote sensing imagery that faces various levels of difficulty due to the complexity of underlying surfaces, insufficient training data, and redundant information in high-dimensional data. To solve these problems, we propose an unsupervised network for cloud detection (UNCD) on multispectral (MS) and hyperspectral (HS) remote sensing images. The UNCD method enforces discriminative feature learning to obtain the residual error between the original input and the background in deep latent space, which is based on the observation that clouds are sparse and modeled as sparse outliers in remote sensing imagery. The UNCD enforces discriminative feature learning to obtain the residual error between the original input and the background in deep latent space, which is based on the observation that clouds are sparse and modeled as sparse outliers in remote sensing imagery. First, a compact representation of the original imagery is obtained by a latent adversarial learning constrained encoder. Meanwhile, the majority class with sufficient samples (i.e., background pixels) is more accurately reconstructed than the clouds with limited samples by the decoder. An image discriminator is used to prevent the generalization of out-of-class features caused by latent adversarial learning. To further highlight the background information in the deep latent space, a multivariate Gaussian distribution is introduced. In particular, the residual error with clouds highlighted and background samples suppressed is applied in the cloud detection in deep latent space. To evaluate the performance of the proposed UNCD method, experiments were conducted on both MS and HS datasets that were captured by various sensors over various scenes, and the results demonstrate its state-of-the-art performance. The sensors that captured the datasets include Landsat 8, GaoFen-1 (GF-1), and GaoFen-5 (GF-5). Landsat 8 was launched at Vandenberg Air Force Base in California on 11 February 2013, in a mission that was initially known as the Landsat Data Continuity Mission (LDCM). China launched the GF-1 satellite. The GF-5 satellite captures hyperspectral observations in the Chinese Key Projects of High-Re solution Earth Observation System. The overall accuracy (OA) values for Images I and II from the Landsat 8 dataset were 0.9526 and 0.9536, respectively, and the OA values for Images III and IV from the GF-1 wide field of view (WFV) dataset were 0.9957 and 0.9934, respectively. Hence, the proposed method outperformed the other considered methods. © 2020 by the authors." "7003480828;13607413100;","Re-imagining the future of education in the era of the fourth industrial revolution",2020,"10.1108/WHATT-10-2019-0066","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079449284&doi=10.1108%2fWHATT-10-2019-0066&partnerID=40&md5=8c8c6c6207f57561896d01db0fd7f984","Purpose: Globally, higher education has been, over the years, a source of innovation, policy, new knowledge and a national asset. However, the advent of the Fourth Industrial Revolution (4IR) is having an impact on the principles of learning from primary to tertiary levels. The purpose of this paper is to consider how the 4IR has and will continue to impact education at the various levels of learning. Design/methodology/approach: The paper aims to bridge the perceived information gap and provide insights into the kinds of educational preparation and the skills and qualifications that 4IR jobs require. In response, the following are considered: the need to tweak the curriculum, adopt the right technology for in class and online delivery and the projection of other learning techniques and skills that are often not considered pertinent. Data gathering for the report was by discussion with experts and consultation of relevant articles and write-ups from related websites. Findings: The advent of smart communication systems involving artificial intelligence, internet, robotics, virtual reality and digital textbooks has opened a new vista in relation to how and what is learnt in schools. Just as technologies brought about smart communication systems, the 4IR model of higher education is rapidly evolving and as such, curriculum development and review must be dynamic, and it must keep pace with the technological advances and skills required in the twenty first century. Research limitations/implications: More purposeful research needs to be conducted in universities and industries with the intention of accelerating internal and external innovations so that markets can be expanded. Furthermore, efforts to reduce the cost and time of generating innovations will need to be intensified. Practical implications: The value and emphasis that are placed on the acquisition of degrees and paper qualifications are changing rapidly. Although it is traditional for students to compete for admission to the face-to-face classroom model, it is no longer unusual for a student to take courses online from any part of the world and still be accepted into positions usually reserved for traditional classroom education. Social implications: As at today, examples of 4IR services include Uber, Airbnb, Cloud services, Artificial intelligence, Cyber-security, three-dimensional printers, driverless cars and robotics. Machine learning and drone technology are also of growing significance. As yet, subjects dealing with such inventions and innovations are not part of the curriculum of many institutions and this is a cause for concern. Originality/value: The 4IR era will bring great changes to how students are taught and what students must learn as the tools for transformational learning are already overwhelming. Jobs will be scarce for those without the requisite skills, whereas those with the right skills will have to keep up with the pace of technological development, otherwise they too will be left behind. Schools will increasingly become centres for the generation of innovation and its incubation and in all this, quality learning, teaching and knowledge impartation can easily be carried out online. © 2020, Emerald Publishing Limited." "56789262700;48662599000;8719489000;56654564800;7006430057;23767612400;56421152600;","Forest signal detection for photon counting LiDAR using Random Forest",2020,"10.1080/2150704X.2019.1682708","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074241228&doi=10.1080%2f2150704X.2019.1682708&partnerID=40&md5=ef3686a58bba4b1b5693c679fd3e9a77","ICESat (The Ice, Cloud, and Land Elevation Satellite)-2, as the new generation of NASA (National Aeronautics and Space Administration)’s ICESat mission, had been successfully launched in September 2018. The sensor onboard the satellite is a newly designed photon counting LiDAR (Light Detection And Ranging) system for the first time used in space. From the currently released airborne simulation data, it can be seen that there exist numerous noise photons scattering from the atmosphere to even below the ground, especially for the vegetation areas. Therefore, relevant research on methods to distinguish the signal photons effectively is crucial for further forestry applications. In this paper, a machine learning based approach was proposed to detect the potential signal photons from 14 MATLAS datasets across 3 sites in the USA. We found that k-NN (k-Nearest Neighbour) distance and the reachability of the photon towards the nearby signal centre showed good stability and contributed to a robust model establishment. The relevant quantitative assessment demonstrated that the machine learning approach could achieve high detection accuracy over 85% based on a very limited number of samples even in rough terrain conditions. Further analysis proved the potential of model transferability across different sites. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57196237174;57213029726;55939316400;","MODIS probabilistic cloud masking over the Amazonian evergreen tropical forests: a comparison of machine learning-based methods",2020,"10.1080/01431161.2019.1637963","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068638824&doi=10.1080%2f01431161.2019.1637963&partnerID=40&md5=75faac6f917fac230870f6a65f3e0810","Amazonian tropical forests play a significant role in global water, carbon and energy cycles. Satellite remote sensing is presented as a feasible means in order to monitor these forests. In particular, the Moderate Resolution Imaging Spectroradiometer (MODIS) is amongst major tools for studying this region. Nevertheless, MODIS operative surface variable retrieval was reported to be impacted by cloud contamination effects. A proper cloud masking is a major consideration in order to ensure accuracy when analysing Amazonian tropical forests current and future status. In the present study, the potential of supervised machine learning algorithms in order to overcome this issue is evaluated. In front of global operative MODIS cloud masking algorithms (MYD35 and the Multi-Angle Implementation of Atmospheric Correction Algorithm (MAIAC)) these algorithms benefit from the fact that they can be optimized to properly represent the local cloud conditions of the region. Models considered were: Gaussian Naïve Bayes (GNB), Linear Discriminant Analysis (LDA), Quadratic Discriminant Analysis (QDA), Random Forests (RF), Support Vector Machine (SVM) and Multilayer Perceptron (MLP). These algorithms are able to provide a continuous measure of cloud masking uncertainty (i.e. a probability estimate of each pixel belonging to clear and cloudy class) and therefore can be used for probabilistic cloud masking. Truth reference dataset (a priori knowledge) requirement was satisfied by considering the collocation of Cloud Profiling Radar (CPR) and Cloud Aerosol Lidar with Orthogonal Polarization (CALIOP) observations with MODIS sensor. Model performance was tested using three independent datasets: 1) collocated CPR/CALIOP and MODIS data, 2) MODIS manually classified images and 3) in-situ ground data. For satellite image and in-situ testing results were additionally compared to current operative MYD35 (version 6.1) and MAIAC cloud masking algorithms. Satellite image and in-situ testing results show that machine learning algorithms are able to improve MODIS operative cloud masking performance over the region. MYD35 and MAIAC tend to underestimate and overestimate the cloud cover over the study region, respectively. Amongst the models considered, probabilistic algorithms (LDA, GNB and in less extent QDA) provided better performance than RF, SVM and MLP machine learning algorithms as they were able to better deal with the viewing conditions limitation that resulted from collocating MODIS and CPR/CALIOP observations. In particular, best performance was obtained for LDA with a difference in Kappa coefficient (model minus MODIS operative algorithm) of 0.293/0.155 (MYD35/MAIAC, respectively) considering satellite image testing validation. Worst performance was obtained for MLP with a difference in Kappa coefficient of 0.175/0.037. For in-situ testing, models overall accuracy (OA) and Kappa coefficient values are higher than MYD35/MAIAC respective values. Models are computationally efficient (swath image calculation time between 0.37 and 9.49 s) and thus being able to be implanted for remote-sensing vegetation retrieval processing chains over the Amazonian tropical forests. LDA stands out as the best candidate because of its maximum accuracy and minimum computational associated. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57212584870;57188717852;55277716700;56664552300;57219294173;57219298103;57211501464;56088674100;7404589013;","Improved Mapping Results of 10 m Resolution Land Cover Classification in Guangdong, China Using Multisource Remote Sensing Data with Google Earth Engine",2020,"10.1109/JSTARS.2020.3022210","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092112240&doi=10.1109%2fJSTARS.2020.3022210&partnerID=40&md5=34ab9cea870c0e6677aefb0bf0f5e73b","Land cover information depicting the complex interactions between human activities and surface change is critically essential for nature conservation, social management, and sustainable development. Recent advances have shown great potentials of remote sensing data in generating high-resolution land cover maps, but it remains unclear how different models, data sources, and inclusive features affect the classification results, which hinders its applications in regional studies requiring more accurate land cover data. Informing these issues, here we developed a robust framework to improve the mapping results of 10 m resolution land cover classification in Guangdong Province, China using thousands of manually collected samples, multisource remote sensing data (Sentinel-1, Sentinel-2, and Luojia-1), machine learning algorithms, and a free cloud-based platform of Google Earth Engine. Results showed that an overall accuracy of 86.12% and a Kappa coefficient of 0.84 could be achieved for land cover classification in Guangdong for 2019. We found that random forest models achieved better performance than classification and regression trees, minimum distance, and support vector machine models. We also found that features derived from Sentinel-1 data and Sentinel-2 spectral indices greatly contributed to the classification process, while the feature of Luojia-1 data was not as much important as other configurations. A comparison between our results and several existing land cover products in terms of classification accuracy and visual interpretation further validated the effectiveness and robustness of the proposed framework. Our experiments and findings not only systematically elucidate the role of classification methods and data sources in deriving more accurate and reliable land cover maps but also provide certain guidelines for future land cover mapping from regional to global scales. © 2008-2012 IEEE." "7401526171;57218396843;35975568000;","Improving PERSIANN-CCS Using Passive Microwave Rainfall Estimation",2020,"10.1007/978-3-030-24568-9_21","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089068676&doi=10.1007%2f978-3-030-24568-9_21&partnerID=40&md5=cd6573882f174612b92fc8659e48e681","Re-calibrated PERSIANN-CCS is one of the algorithms used in “Integrated Multi-satellitE Retrievals for GPM” (IMERG) to provide high-resolution precipitation estimations from the NASA Global Precipitation Measurement (GPM) program and retrospective data generation for the period covered by the Tropical Rainfall Measurement Mission (TRMM). This study presents the development of a re-calibrated PERSIANN-CCS algorithm for the next-generation GPM multi-sensor precipitation retrieval algorithm (IMERG). The activities include implementing the probability matching method to update PERSIANN-CCS using passive microwave (PMW) rainfall estimation from low earth orbit (LEO) satellites and validation of precipitation estimation using radar rainfall measurement. Further improvement by the addition of warm rain estimation to the PERSIANN-CCS algorithm using warmer temperature thresholds for cloud image segmentation is also presented. Additionally, developments using multispectral image analysis and machine learning approaches are discussed and proposed for future studies. © 2020, Springer Nature Switzerland AG." "22433897700;57195300026;","Data-driven agriculture for rural smallholdings",2020,"10.5311/JOSIS.2020.20.669","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088784194&doi=10.5311%2fJOSIS.2020.20.669&partnerID=40&md5=842ca6f59783a61aa7ef1fb6ccdab3e3","Spatial information science has a critical role to play in meeting the major challenges facing society in the coming decades, including feeding a population of 10 billion by 2050, addressing environmental degradation, and acting on climate change. Agriculture and agri-food value-chains, dependent on spatial information, are also central. Due to agriculture's dual role as not only a producer of food, fibre, and fuel but also as a major land, water, and energy consumer, agriculture is at the centre of both the food-water-energyenvironment nexus and resource security debates. The recent confluence of a number of advances in data analytics, cloud computing, remote sensing, computer vision, robotic and drone platforms, and Internet of Things (IoT) sensors and networks have lead to a significant reduction in the cost of acquiring and processing data for decision support in the agricultural sector. When combined with costeffective automation through development of swarm farming technologies, these technologies have the potential to decouple productivity and cost efficiency from economies of size, reducing the need to increase farm size to remain economically viable. We argue that these pressures and opportunities are driving agricultural value-chains towards high-resolution data-driven decision-making, where even decisions made by small rural landowners can be data-driven. We survey recent innovations in data, especially focusing on sensor, spatial, and data mining technologies with a view to their agricultural application; discuss economic feasibility for small farmers; and identify some technical challenges that need to be solved to reap the benefits. Flexibly composable information resources, coupled with sophisticated data sharing technologies, and machine learning with transparently embedded spatial and aspatial methods are all required. © by the author(s)." "57200377471;55811589800;57215038513;","Transitioning a legacy reservoir simulator to cloud native services",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085775736&partnerID=40&md5=7210a84b869ed41ebe9062ef222c19f3","The digital transformation journey provides new opportunities for running simulations through cloud computing, with flexibility in hardware resources and availability of a wide array of software tools that can enhance decision- making. This work reviews the benefits of running reservoir simulations in a cloud environment and demonstrates the efficiency and cost savings. Additionally, a workflow for uncertainty analysis and history matching that integrates data analysis and machine-learning tools is presented. First, the hardware architecture should be designed to meet parallel reservoir simulation needs: significant message passing occurs between computer nodes; for satisfactory performance, these nodes should be connected by a low-latency network rather than randomly located. Second, to help ensure portability and easy replication across multiple cloud sites and platforms, the software performing the simulations should be containerized. Third, to reduce the time required to start a new simulation run, the Kubernetes platform is used to optimize resource allocation. Finally, reservoir simulation in the cloud is no longer merely the running of the simulation model; it is integrated with data management and data analysis tools for decision-making. The cloud-based simulation services discussed herein exhibit good results during scale up when a simulation operation requires a larger number of central processing units and/or greater memory and during scale out when thousands of operation scenarios are necessary for history matching. The ""pay as you go"" pricing model reduces the time and capital costs of acquiring the new computing infrastructure to nearly zero, and the effectively unlimited scale-out capability can reduce the elapsed time for history matching by 80%. The availability of data centers in different regions is good for team collaborations. It serves the data management tool well to track history data, perform data mining, extract more information, and make decisions. Compared to traditional reservoir simulation, the cloud-based reservoir simulation software as a service model simplifies the process and reduces hardware acquisition and maintenance costs. Integrating intelligent data analysis with simulation helps quantify the uncertainty in the model and enables improved decisions. Copyright 2020, International Petroleum Technology Conference." "57215574864;7006738371;35414779200;","A multilayer perceptron for obtaining quick parameter estimations of cool exoplanets from geometric albedo spectra",2020,"10.1088/1538-3873/ab740d","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081251876&doi=10.1088%2f1538-3873%2fab740d&partnerID=40&md5=01f0d39d6c778d84ef4a132f03e5052a","Future space telescopes now in the concept and design stage aim to observe reflected light spectra of extrasolar planets. Assessing whether given notional mission and instrument design parameters will provide data suitable for constraining quantities of interest typically requires time consuming retrieval studies in which tens to hundreds of thousands of models are compared to data with a given assumed signal to noise ratio, thereby limiting the rapidity of design iterations. Here we present a machine learning approach employing a Multilayer Perceptron (MLP) trained on model albedo spectra of extrasolar giant planets to estimate a planet’s atmospheric metallicity, gravity, effective temperature, and cloud properties given simulated observed spectra. The stand-alone C++ code we have developed can train new MLP’s on new training sets within minutes to hours, depending upon the dimensions of input spectra, size of the training set, desired output, and desired accuracy. After the MLP is trained, it can classify new input spectra within a second, potentially helping speed observation and mission design planning. Our MLP’s were trained using a grid of model spectra that varied in metallicity, gravity, temperature, and cloud properties. The results show that a trained MLP is an elegant means for reliable in situ estimations when applied to model spectra. We analyzed the effect of using models in a grid range known to have degeneracies. © 2020. The Astronomical Society of the Pacific. All rights reserved." "57189756538;57212864756;57212876123;57217192429;55844950200;","Combining Participatory Mapping, Cloud Computing and Machine Learning for Mapping Climate Induced Landslide Susceptibility in Lembeh Island, North Sulawesi",2019,"10.1088/1755-1315/363/1/012020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077471375&doi=10.1088%2f1755-1315%2f363%2f1%2f012020&partnerID=40&md5=c65dcd92e32d5bb97ff84c13ee4a2c5a","This study explored the use of participatory mapping and several machine learning algorithms (Naïve Bayes, GMO Maxent, SVM, CART, and Random Forest) to map climate induced landslide susceptibility in Lembeh Island, North Sulawesi, based on Earth Observation data available in Google Earth Engine. Participatory mapping on landslide incidents were conducted in three villages, i.e., Kareko, Pintu Kota, and Pasir Panjang. Data used include digital elevation model from SRTM, multispectral imageries from Sentinel 2, and precipitation from CHIRPS. Terrain modelling was done to DEM to come up with elevation, slope, curvature, and aspect. A cloud free mosaic of Sentinel Images was created using the median reducer and then NDVI was calculated. Precipitation data from CHIRPS was sampled and interpolated using kriging and reduced to maximum and mean. Each algorithm was trained using 70% participatory mapping data and then the prediction was tested for accuracy using the last 30%. Results showed that Random Forest, SVM, CART, and GMO Maxent gave 0.98 testing accuracy and Naïve Bayes only 0.90. The map showed that due to the terrain condition, Lembeh Island is prone to Landslide and even though previously BNPB already provide a landslide hazard risk map, there were many areas not included on that map. Therefore, the map could become an input for BNPB and the Bitung City for developing a mitigation and adaptation strategy. Machine learning and cloud computing along with participatory mapping could also complement mechanistic or multi-criteria analysis using GIS model for landslide susceptibility mapping. © Published under licence by IOP Publishing Ltd." "57209862484;57195480773;56193847400;57200567420;55949825400;","Cloud detection from FY-4A's geostationary interferometric infrared sounder using machine learning approaches",2019,"10.3390/rs11243035","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077887807&doi=10.3390%2frs11243035&partnerID=40&md5=fc48466a1b7e26d371b4c12e93d1a928","FengYun-4A (FY-4A)'s Geostationary Interferometric Infrared Sounder (GIIRS) is the first hyperspectral infrared sounder on board a geostationary satellite, enabling the collection of infrared detection data with high temporal and spectral resolution. As clouds have complex spectral characteristics, and the retrieval of atmospheric profiles incorporating clouds is a significant problem, it is often necessary to undertake cloud detection before further processing procedures for cloud pixels when infrared hyperspectral data is entered into assimilation system. In this study, we proposed machine-learning-based cloud detection models using two kinds of GIIRS channel observation sets (689 channels and 38 channels) as features. Due to differences in surface cover and meteorological elements between land and sea, we chose logistic regression (lr) model for the land and extremely randomized tree (et) model for the sea respectively. Six hundred and eighty-nine channels models produced slightly higher performance (Heidke skill score (HSS) of 0.780 and false alarm rate (FAR) of 16.6% on land, HSS of 0.945 and FAR of 4.7% at sea) than 38 channels models (HSSof 0.741 and FAR of 17.7% on land, HSS of 0.912 and FAR of 7.1% at sea). By comparing visualized cloud detection results with the Himawari-8 Advanced Himawari Imager (AHI) cloud images, the proposed method has a good ability to identify clouds under circumstances such as typhoons, snow covered land, and bright broken clouds. In addition, compared with the collocated Advanced Geosynchronous Radiation Imager (AGRI)-GIIRS cloud detection method, the machine learning cloud detection method has a significant advantage in time cost. This method is not effective for the detection of partially cloudy GIIRS's field of views, and there are limitations in the scope of spatial application. © 2019 by the authors." "57207565169;56680839900;57191839329;6506008310;6602842679;55812857100;","Combination of an automated 3D field phenotyping workflow and predictive modelling for high-throughput and non-invasive phenotyping of grape bunches",2019,"10.3390/rs11242953","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077850531&doi=10.3390%2frs11242953&partnerID=40&md5=360a0187fbffbdbe9917308b359fd5bd","In grapevine breeding, loose grape bunch architecture is one of the most important selection traits, contributing to an increased resilience towards Botrytis bunch rot. Grape bunch architecture is mainly influenced by the berry number, berry size, the total berry volume, and bunch width and length. For an objective, precise, and high-throughput assessment of these architectural traits, the 3D imaging sensor Artec® Spider was applied to gather dense point clouds of the visible side of grape bunches directly in the field. Data acquisition in the field is much faster and non-destructive in comparison to lab applications but results in incomplete point clouds and, thus, mostly incomplete phenotypic values. Therefore, lab scans of whole bunches (360°) were used as ground truth. We observed strong correlations between field and lab data but also shifts in mean and max values, especially for the berry number and total berry volume. For this reason, the present study is focused on the training and validation of different predictive regression models using 3D data from approximately 2000 different grape bunches in order to predict incomplete bunch traits from field data. Modeling concepts included simple linear regression and machine learning-based approaches. The support vector machine was the best and most robust regression model, predicting the phenotypic traits with an R2 of 0.70-0.91. As a breeding orientated proof-of-concept, we additionally performed a Quantitative Trait Loci (QTL)-analysis with both the field modeled and lab data. All types of data resulted in joint QTL regions, indicating that this innovative, fast, and non-destructive phenotyping method is also applicable for molecular marker development and grapevine breeding research. © 2019 by the authors." "9335419500;57202985540;57211938397;56425382100;","An azure aces early warning system for air quality index deteriorating",2019,"10.3390/ijerph16234679","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075465582&doi=10.3390%2fijerph16234679&partnerID=40&md5=787334dcc19eb38776e538b6eb62a074","With the development of industrialization and urbanization, air pollution in many countries has become more serious and has affected people’s health. The air quality has been continuously concerned by environmental managers and the public. Therefore, accurate air quality deterioration warning system can avoid health hazards. In this study, an air quality index (AQI) warning system based on Azure cloud computing platform is proposed. The prediction model is based on DFR (Decision Forest Regression), NNR (Neural Network Regression), and LR (Linear Regression) machine learning algorithms. The best algorithm was selected to calculate the 6 pollutants required for the AQI calculation of the air quality monitoring in real time. The experimental results show that the LR algorithm has the best performance, and the method of this study has a good prediction on the AQI index warning for the next one to three hours. Based on the ACES system proposed, it is hoped that it can prevent personal health hazards and help to reduce medical costs in public. © 2019 by the authors. Licensee MDPI, Basel, Switzerland." "57214713846;57193133682;","Sensor data collection and its architecture with internet of things",2019,"10.35940/ijeat.A9564.109119","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074558358&doi=10.35940%2fijeat.A9564.109119&partnerID=40&md5=34402b0aa7e063620e65afc3d72f7928","Sensors are gadgets, which can screen temperature, moistness, weight, commotion levels, setting mindfulness, lighting condition and identify speed, position, and size of an Object. Sensor information are getting accumulated in gigantic amount thus they are overseen utilizing NOSQL. The information will be gathered in an IOT cloud stage where it will be additionally prepared with machine learning methods for prescient examination. What's more, eventually with the required answer for the business structure will be created. This paper explain the proposed system for IoT data collection with AWS (Amazon Web Service) cloud platform. Various system components like Kinesis stream, M2M platform, Notification service and secured IoT service layout. The complete BMS system architecture is detailed in this paper. © BEIESP." "55575342100;35226698000;57208121325;7103389378;7801621724;7102811204;","Biomass Assessment of Agricultural Crops Using Multi-temporal Dual-Polarimetric TerraSAR-X Data",2019,"10.1007/s41064-019-00076-x","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074137730&doi=10.1007%2fs41064-019-00076-x&partnerID=40&md5=20fc2e59826d1939c53e1105d797fbe3","The biomass of three agricultural crops, winter wheat (Triticum aestivum L.), barley (Hordeum vulgare L.), and canola (Brassica napus L.), was studied using multi-temporal dual-polarimetric TerraSAR-X data. The radar backscattering coefficient sigma nought of the two polarization channels HH and VV was extracted from the satellite images. Subsequently, combinations of HH and VV polarizations were calculated (e.g. HH/VV, HH + VV, HH × VV) to establish relationships between SAR data and the fresh and dry biomass of each crop type using multiple stepwise regression. Additionally, the semi-empirical water cloud model (WCM) was used to account for the effect of crop biomass on radar backscatter data. The potential of the Random Forest (RF) machine learning approach was also explored. The split sampling approach (i.e. 70% training and 30% testing) was carried out to validate the stepwise models, WCM and RF. The multiple stepwise regression method using dual-polarimetric data was capable to retrieve the biomass of the three crops, particularly for dry biomass, with R2 > 0.7, without any external input variable, such as information on the (actual) soil moisture. A comparison of the random forest technique with the WCM reveals that the RF technique remarkably outperformed the WCM in biomass estimation, especially for the fresh biomass. For example, the R2 > 0.68 for the fresh biomass estimation of different crop types using RF whereas WCM show R2 < 0.35 only. However, for the dry biomass, the results of both approaches resembled each other. © 2019, Deutsche Gesellschaft für Photogrammetrie, Fernerkundung und Geoinformation (DGPF) e.V." "57207828620;55440959500;","Predictive ability of covariate-dependent Markov models and classification tree for analyzing rainfall data in Bangladesh",2019,"10.1007/s00704-019-02812-0","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85062964084&doi=10.1007%2fs00704-019-02812-0&partnerID=40&md5=6d9ad96bcaa31acff328cd0445d036f6","This study attempts to make comparison between different parametric regressive models for the bivariate binary data with a machine learning technique. The data on sequential occurrence of rainfall in consecutive days is considered. The outcomes are classified as rainfall in both days, rainfall in one of the consecutive days, and no rainfall in both days. The occurrence of rainfall in consecutive days is analyzed by using statistical models with covariate dependence and classification tree for the period from 1980 to 2014. We have used relative humidity, minimum temperature, maximum temperature, sea level pressure, sunshine hour, and cloud cover in the model as covariates. The binary outcome variable is defined as the occurrence or non-occurrence of rainfall. Five regions of Bangladesh are considered in this study and one station from each region is selected on the basis of two criteria: (i) contains fewer missing values and (ii) representative of the regional characteristics geographically. Several measures are used to compare the models based on Markov chain and classification tree. It is found that for yearly data, both the Markov model and classification tree performed satisfactorily. However, the seasonal data show variation of rainfall. In some seasons, both models perform equally good such as monsoon, pre-monsoon, and post-monsoon, but in the winter season, the Markov model works poorly whereas classification tree fails to work. Additionally, we also observe that the Markov model performed consistently for each season and performs better compared with the classification tree. It has been demonstrated that the covariate-dependent Markov models can be used as classifiers alternative to the classification tree. It is revealed that the predictive ability of the covariate-dependent Markov model based on Markovian assumption performs either better or equally good compared with the classification tree. The joint models also consistently showed better predictive performance compared with regressive model for whole year data as well as for several seasonal data. © 2019, Springer-Verlag GmbH Austria, part of Springer Nature." "57215487421;14053460100;57217204638;57217199418;","A weighted normalized likelihood procedure for empirical land change modeling",2019,"10.1007/s40808-019-00584-0","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086732005&doi=10.1007%2fs40808-019-00584-0&partnerID=40&md5=282d1e940129a74d77460aec681d7d66","A critical foundation for empirical land change modeling is the mapping of transition potentials—quantitative evaluations of the readiness or suitability of land to go through a transition. This paper presents a procedure based on empirically determined normalized likelihoods of transition. It shows that these normalized likelihoods equate to posterior probabilities if case–control sampling is carried out among historical instances of change and persistence. The posterior probabilities can then be aggregated at the pixel level across multiple covariates using linear opinion pooling where the pixel-specific weight for each covariate is determined locally by its ability to distinguish between the alternatives of change or persistence. Thus, covariates with spatially varying diagnostic ability can be productively incorporated. The resulting algorithm is shown to have a skill comparable to that of a multi-layer perceptron approach with the advantage of high efficiency and amenability to distributed processing in a cloud environment. © 2019, Springer Nature Switzerland AG." "8109934200;57215870896;57211137457;56992503700;57215489736;36069985300;57213216388;57214356147;57196711847;24783260400;6602980349;21935606200;","SATVAM: Toward an iot cyber-infrastructure for low-cost urban air quality monitoring",2019,"10.1109/eScience.2019.00014","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081005328&doi=10.1109%2feScience.2019.00014&partnerID=40&md5=094bff23645d2ecd1e9a2d887eb44343","Air pollution is a public health emergency in large cities. The availability of commodity sensors and the advent of Internet of Things (IoT) enable the deployment of a city-wide network of 1000's of low-cost real-Time air quality monitors to help manage this challenge. This needs to be supported by an IoT cyber-infrastructure for reliable and scalable data acquisition from the edge to the Cloud. The low accuracy of such sensors also motivates the need for data-driven calibration models that can accurately predict the science variables from the raw sensor signals. Here, we offer our experiences with designing and deploying such an IoT software platform and calibration models, and validate it through a pilot field deployment at two mega-cities, Delhi and Mumbai. Our edge data service is able to even-out the differential bandwidths from the sensing devices and to the Cloud repository, and recover from transient failures. Our analytical models reduce the errors of the sensors from a best-case of 63% using the factory baseline to as low as 21%, and substantially advances the state-of-The-Art in this domain. © 2019 IEEE." "57193933577;7004479957;","Single-Column Emulation of Reanalysis of the Northeast Pacific Marine Boundary Layer",2019,"10.1029/2019GL083646","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071267445&doi=10.1029%2f2019GL083646&partnerID=40&md5=c7dce819b65137a844ceb16dcba60ea9","An artificial neural network is trained to reproduce thermodynamic tendencies and boundary layer properties from European Center for Medium-Range Weather Forecasts Reanalysis 5th Generation high resolution realization reanalysis data over the summertime northeast Pacific stratocumulus to trade cumulus transition region. The network is trained prognostically using 7-day forecasts rather than using diagnosed instantaneous tendencies alone. The resulting model, Machine-Assisted Reanalysis Boundary Layer Emulation, skillfully reproduces the boundary layer structure and cloud properties of the reanalysis data in 7-day single-column prognostic simulations over withheld testing periods. Radiative heating profiles are well simulated, and the mean climatology and variability of the stratocumulus to cumulus transition are accurately reproduced. Machine-Assisted Reanalysis Boundary Layer Emulation more closely tracks the reanalysis than does a comparable configuration of the underlying forecast model. ©2019. American Geophysical Union. All Rights Reserved." "56097323700;35182211000;37123014300;","Development of a high spatiotemporal resolution cloud-type classification approach using Himawari-8 and CloudSat",2019,"10.1080/01431161.2019.1594438","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063149357&doi=10.1080%2f01431161.2019.1594438&partnerID=40&md5=7b074027ffec2bc1075691f27123bda8","With the cloud scenario products from CloudSat, we developed a high spatiotemporal resolution cloud-type classification procedure for Himawari-8 multispectral datasets using maximum-likelihood estimation (MLE) and random forests (RF) classification. The training and classification procedures were processed independently, and both algorithms provided cloud-type results with a good performance. Validation indicated that the use of the visible (VIS) channel significantly improved the cloud-type identification capabilities, while the use of three or more channels simultaneously resulted in considerable improvements over the use of bispectral combinations. The comparison among different classifiers also revealed that RF was more sensitive than MLE to the quality and distribution of the training data. After retraining the RF using MLE-based clustered samples, we produced two more-reasonable and efficient classifiers that can be used during the day and night. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "55715413100;36664709300;14028364400;57207459537;","Gobal forest cover mapping using landsat and google earth engine cloud computing",2019,"10.1109/Agro-Geoinformatics.2019.8820469","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072927621&doi=10.1109%2fAgro-Geoinformatics.2019.8820469&partnerID=40&md5=333c08a02b5855495d6f2c1f78f660bc","Nowadays, the access of Landsat data-sets and the ever-lowering costs of computing make it feasible to monitor the Earth's land cover at Landsat resolutions of 30 meter. However, the rapid forest-covered products on a large scale, such as intercontinental or global, is still challenging. By utilizing the huge catalog of satellite imagery as well as the high-performance computing capacity of Google Earth Engine, we proposed an automated pipeline for generating 30-meter resolution global-scale forest map from time-series of Landsat images, and a novel 30-meter resolution global forest map of 2018 is released. In this paper, we describe the methods to create products of forest cover at Landsat resolutions. First, we partitioned the landscapes into sub-regions of similar forest type and spatial continuity, thus maximizing spectral differentiation, simplifying classifier model and improving classification accuracy. Then, with the existing forest cover, which come from a variety of sources, a multi-source forest/non-forest sample set was established for machine algorithm learning training. Finally, a machine learning algorithm was used to obtain samples automatically, extract the characteristics of satellite images and establish the forest / non-forest classifier models. Taking the Landsat8 images in 2018 as a case, selecting satellite image features based on the study of forest reflectance, including onboard reflectivity, the index of forest vegetation and the texture features of each band, using established forest eco-zoning and multi-source forest / non-forest sample points, we realized automated learning and classification of forest cover for three initial zones. The accuracy verification of forest cover products in the three region was carried on two aspects: collecting verification points on high resolution satellite imagery (e.g. google earth), and cross-validating the current globally disclosed forest cover products. These two methods will illustrate the accuracy of the forest cover product. © 2019 IEEE." "57205765622;57191616594;56050235900;57205541304;56915184300;36711526700;","Farming on the edge: Architectural goals",2019,"10.1109/Agro-Geoinformatics.2019.8820424","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072921206&doi=10.1109%2fAgro-Geoinformatics.2019.8820424&partnerID=40&md5=c3fbe1274a9e36a04953f92a1765fecc","This research investigates how advances in Internet of Things (IoT) and availability of internet connection would enable Edge Solutions to promote smart utilization of existing machines at the edge. The presented results are based on experiments performed in real scenarios using the proposed solution. Whereas scenarios were cloned from real environments it is important to have in mind that experiments were performed with low load in terms of data and small number of devices in terms of distribution. As result of extensive architecture investigation for an optimal edge solution and its possible correlation to industrial applications, this paper will provide evidences supporting the use of edge solutions in challenging conditions which arise at the edge, including smart factories and smart agriculture. The present work assumes that the reader has some exposition to Edge computing, Cloud computing and software development. The paper will present some important findings on this area, compare main architectural aspects and will provide a broad view of how edge solutions might be built for this particular scenario. Having discussed how the ideal architecture works and having provided an overview about how it may be applied to industrial plants, the final section of this paper addresses how artificial intelligence will fit into edge solutions, forming a new source of 'smart capabilities' to existing environments. © 2019 IEEE." "57208372273;57205760598;57208597081;","Using Convolutional Neural Networks for Cloud Detection from Meteor-M No. 2 MSU-MR Data",2019,"10.3103/S1068373919070045","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070222008&doi=10.3103%2fS1068373919070045&partnerID=40&md5=bceac4b26f4af32c7aa883edc850638b","A method for cloud detection using the machine-learning algorithm based on a convolutional neural network is presented. Input data are satellite images received from the MSU-MR multispectral low-resolution scanning unit onboard the Meteor-M No. 2 satellite. The developed method can be an alternative to the traditional algorithms of cloud detection based on the calculation of differential indices and thresholds. The algorithm is verified using the machine-learning metrics, comparing the resulting cloud mask with the reference one obtained by interpreting the satellite image by an experienced meteorologist. It was also compared (for verification) with a similar product based on VIIRS spectroradiometer data. The cloud mask computed using the algorithm allows the automatic thematic processing of satellite images. © 2019, Allerton Press, Inc." "57202977053;57209296376;57201667638;56544915700;26030052000;15923105200;","A high-speed particle phase discriminator (PPD-HS) for the classification of airborne particles, as tested in a continuous flow diffusion chamber",2019,"10.5194/amt-12-3183-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067259848&doi=10.5194%2famt-12-3183-2019&partnerID=40&md5=ff24ecdc6e7bba88bf2dde7df78ee37f","A new instrument, the High-speed Particle Phase Discriminator (PPD-HS), developed at the University of Hertfordshire, for sizing individual cloud hydrometeors and determining their phase is described herein. PPD-HS performs an in situ analysis of the spatial intensity distribution of near-forward scattered light for individual hydrometeors yielding shape properties. Discrimination of spherical and aspherical particles is based on an analysis of the symmetry of the recorded scattering patterns. Scattering patterns are collected onto two linear detector arrays, reducing the complete 2-D scattering pattern to scattered light intensities captured onto two linear, one-dimensional strips of light sensitive pixels. Using this reduced scattering information, we calculate symmetry indicators that are used for particle shape and ultimately phase analysis. This reduction of information allows for detection rates of a few hundred particles per second. Here, we present a comprehensive analysis of instrument performance using both spherical and aspherical particles generated in a well-controlled laboratory setting using a vibrating orifice aerosol generator (VOAG) and covering a size range of approximately 3-32 μm. We use supervised machine learning to train a random forest model on the VOAG data sets that can be used to classify any particles detected by PPD-HS. Classification results show that the PPD-HS can successfully discriminate between spherical and aspherical particles, with misclassification below 5% for diameters >3μm. This phase discrimination method is subsequently applied to classify simulated cloud particles produced in a continuous flow diffusion chamber setup. We report observations of small, near-spherical ice crystals at early stages of the ice nucleation experiments, where shape analysis fails to correctly determine the particle phase. Nevertheless, in the case of simultaneous presence of cloud droplets and ice crystals, the introduced particle shape indicators allow for a clear distinction between these two classes, independent of optical particle size. From our laboratory experiments we conclude that PPD-HS constitutes a powerful new instrument to size and discriminate the phase of cloud hydrometeors. The working principle of PPD-HS forms a basis for future instruments to study microphysical properties of atmospheric mixed-phase clouds that represent a major source of uncertainty in aerosol-indirect effect for future climate projections.. © Author(s) 2019." "57210343217;57210336943;57210340168;26639841900;","Implementation of machine learning techniques applied to the network intrusion detection system",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070482626&partnerID=40&md5=d34b66f1d3f8adff05593710ed100ab2","An immense amount of data is being generated every second due to technological advancement and reforms. Social networking and cloud computing are generating a huge amount of data every second. Every minute data is being captured in the computing world from the click of the mouse to video people tend to watch generating an immediate recommendation. Everything a user is doing on the internet is being captured in different ways for multiple intents. Now it all ends up to monitor the system and network and, secure lines and servers. This mechanism is called Intrusion Detection System(IDS). Hacker uses multiple numbers of ways to attack the system which can be detected through a number of algorithm and techniques. A comprehensive survey of some major techniques of machine learning implemented on intrusion Detection was done where techniques based on kmeans, K-means with principal component analysis, Random Forest algorithm Extreme learning the ma-chine, techniques, classification algorithms such as Naive Bayes algorithm, Hoeffding Tree algorithm.Also, Accuracy Updated Ensemble algorithm, Accuracy Weighted Ensemble algorithm, Support Vector Machine, Genetic algorithm and Deep learning were studied. Now some of these algorithms are applied upon the NSL-KDD data set and compared on the basis of their accuracy. © BEIESP." "57210211142;57210204611;36467916800;","Video summarization using adaptive thresholding by machine learning for distributed cloud storage",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069926529&partnerID=40&md5=857110a07a1bea9cd7d72df977cb602a","The growth in the video content-based communication and information management have motivated the applications dealing with the video contents to be placed on cloud-based data-storages, which are associated with datacentres. The applications ranging from social media to distance education and surveillance to business communication or any applications related to the e-governess are considering the video data as one of the most preferred mode of communication. The video data is enriched with audio and visual contents, which makes analysis or expressing information high convenient. Computerized video is an electronic portrayal of moving visual pictures as encoded advanced information. This is as opposed to simple video, which speaks to moving visual pictures with a simple sign. The advanced video contains a progression of computerized pictures showed in quick progression. The number of applications, as mentioned, dealing with video data is increasing and as a result a large amount of video content is generated every day. Hence, the complexity of retrieving the information from the video contents are also increasing. The bottleneck of the video retrieval process is for a lower sized segment of the video content can be retrieved in low time complexity and if the information to be preserved to a higher extend, then the retrieval time complexity increases. Thus, a good number of parallel researchers have introduced various methods for video content summarization and retrieval using summarization with efficient searching methods, but most of the parallel research outcomes are criticized for either higher time complexity or for higher information loss. This problem can be ideally solved by finding the highly accurate ratio of key information video frames from the total video content. Henceforth, this work, presents a novel machine learning method for identifying the key frames, not only based on the information available in the frame, also validating the key frames with the thresholds of the objects or changes in the frame. The work is again enhanced by considering the adaptive thresholding method for distributed and collaborative video information. The measures taken in the proposed algorithm produces a 98% accuracy for video information representation and nearly 99% reduction in the video frames, which results into nearly 99% reduction in the time complexity. © BEIESP." "57210207708;56623155300;57210206706;","Cloud based healthcare framework for criticality level analysis",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069912205&partnerID=40&md5=13f695c05b4814864bc7c9041c952cf3","In a Cloud based development Amazon Web Service(AWS) is a platform which secures the cloud services, offers database storage, content delivery, computer power and also provides other functionality to help business scale and grow. This proposed work aims that medical healthcare inputs from various sensors have been automatically retrieved and directly loaded into the cloud. Once the data has been loaded creating the inference engine, setting up a big data cloud environment and store the data into a cloud based dataset. Medical data is calculated using machine learning algorithms such as K-Nearest Neighbor (KNN), Naïve Bayes and Support Vector Machine (SVM) through R shiny web application. The cloud system stores the health care data and transmitted to practitioners through the web service network. Based on these medical data the score value of a patient is calculated and displays criticality of patient. © BEIESP." "57209139058;55694076900;57191433477;57191247319;56029052900;8529553300;57071147200;7701309571;","Detecting Comma-Shaped Clouds for Severe Weather Forecasting Using Shape and Motion",2019,"10.1109/TGRS.2018.2887206","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066611927&doi=10.1109%2fTGRS.2018.2887206&partnerID=40&md5=bf6cd6bd28b389265d5429ffea35a4d7","Meteorologists use shapes and movements of clouds in satellite images as indicators of several major types of severe storms. Yet, because satellite image data are in increasingly higher resolution, both spatially and temporally, meteorologists cannot fully leverage the data in their forecasts. Automatic satellite image analysis methods that can find storm-related cloud patterns are thus in demand. We propose a machine learning and pattern recognition-based approach to detect 'comma-shaped' clouds in satellite images, which are specific cloud distribution patterns strongly associated with cyclone formulation. In order to detect regions with the targeted movement patterns, we use manually annotated cloud examples represented by both shape and motion-sensitive features to train the computer to analyze satellite images. Sliding windows in different scales ensure the capture of dense clouds, and we implement effective selection rules to shrink the region of interest among these sliding windows. Finally, we evaluate the method on a hold-out annotated comma-shaped cloud data set and cross match the results with recorded storm events in the severe weather database. The validated utility and accuracy of our method suggest a high potential for assisting meteorologists in weather forecasting. © 2019 IEEE." "56979231400;55634326100;7201369119;","Semi-automated open water iceberg detection from Landsat applied to Disko Bay, West Greenland",2019,"10.1017/jog.2019.23","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065863688&doi=10.1017%2fjog.2019.23&partnerID=40&md5=335f25d70d68f10cf8a8eeef497226fb","Changes in Greenland's marine-terminating outlet glaciers have led to changes in the flux of icebergs into Greenland's coastal waters, yet icebergs remain a relatively understudied component of the ice-ocean system. We developed a simple iceberg delineation algorithm for Landsat imagery. A machine learning-based cloud mask incorporated into the algorithm enables us to extract iceberg size distributions from open water even in partially cloudy scenes. We applied the algorithm to the Landsat archive covering Disko Bay, West Greenland, to derive a time series of iceberg size distributions from 2000-02 and 2013-15. The time series captures a change in iceberg size distributions, which we interpret as a result of changes in the calving regime of the parent glacier, Sermeq Kujalleq (Jakobshavn Isbræ). The change in calving style associated with the disintegration and disappearance of Sermeq Kujalleq's floating ice tongue resulted in the production of more small icebergs. The increased number of small icebergs resulted in increasingly negative power law slopes fit to iceberg size distributions in Disko Bay, suggesting that iceberg size distribution time series provide useful insights into changes in calving dynamics. © 2019 The Author(s)." "7006413284;57214912047;7006940184;56928095700;57214910980;","Data driven smart monitoring for pipeline integrity assessment",2019,"10.2118/197327-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088200548&doi=10.2118%2f197327-ms&partnerID=40&md5=3901aba2ba55ea3a50eb9eb7d4c15b49","Efficiency and safety are primary requirements for oil & gas fluid filled transportation system. However, the complexity of the asset makes it challenging to derive a theoretical framework for managing the control parameters. The current frontier for a real time monitoring exploits the ""digital tansformation"", i.e. the acquisition and the analysis of large datasets recorded along the whole asset lifecycle, which are used to infer ""data driven"" relations and to predict the evolution of the asset integrity. This paper presents some results of a research project for the design, implementation and testing of a ""machine learning"" approach to vibroacoustic data recorded continuously by acquisition units installed every 10-20 km along a pipeline. In a fluid transportation system, vibroacoustic signals are generated by the flow regulation equipment (i.e. pumping, valves, metering), by the fluid flowing (i.e. turbulence, cavitation, bubbles), by third party interference (i.e. spillage, sabotage, illegal tapping), by internal inspection using PIGs operations), and by natural hazards (i.e. microseismic, subsidence, landslides). The basic principle of machine learning is to ""observe"", for an appropriate time interval, a series of descriptors, in this stage related to vibroacoustic signals but that can be integrated with other physical data (i.e. temperature, density, viscosity), in order to ""learn"" their safe range of variation or, when properly fed to a classification procedure, to obtain automatically a discrete set of operational status. The classification criteria are then applied to new data, highlighting the presence of system anomalies. The paper considers vibroacoustic signals collected at the flow stations of an oil trunkline in Nigeria. The vibroacoustic signals are the static pressure, the acceleration and the pressure transients recorded at the departure and at the arrival terminals. More than one year of data is available. Derived smart indicators are defined, which are directly linked to the asset parameters: for instance, the cross-correlation of the pressure transients at adjacent measuring locations permits to estimate the fluid channel continuity (correlation value), the sound velocity (time of correlation peak), and the sound attenuation (amplitude versus frequency amplitude decay). A portion of the data during normal operation is used for training and tuning a reference model. After that, new data are compared with the model, and anomalies are automatically detected. Two kind of errors are raised: i) sensors; ii) alerts. Sensor errors are referred to missing or corrupted sensors data. Alerts are raised when the measured physical quantities are not coherent with the functional and known service behaviors of the transport system. The system model is not static over time, and in fact it can be updated by the operators' feedback, that can tag false alarms and thus, automatically, re-define the set of operational scenarios of the upstream system. The medium-long term construction and update of data driven models is effective for predictive maintenance, automatic anomalies detection, optimization of operational procedures. Moreover, the new policy of data management and the opportunity of gaining awareness by interconnecting the monitoring experience of different assets leverages the introduction of new technologies (cloud, big data), new professional figures (smart data scientist), new operational and business models. © 2019, Society of Petroleum Engineers" "56338549300;57210363428;57216230666;57208208641;57200248379;57210364558;35617712500;","Enhanced reservoir geosteering and geomapping from refined models of ultra-deep LWD resistivity inversions using machine-learning algorithms",2019,"10.30632/T60ALS-2019_EE","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083952227&doi=10.30632%2fT60ALS-2019_EE&partnerID=40&md5=02ecbd0b7afc2ffe43398486ff3e437e","Ultradeep logging-while-drilling (LWD) resistivity tools have been widely used in the borehole resistivity applications. Previous field examples successfully demonstrated a detection range more than 200 ft away from a wellbore. As a result of the ultradeep detection capability, the complexity of the inversion process increases to accommodate a larger number of layered models, as compared to the conventional approaches. Cloud-based distributed solutions were implemented in the algorithms to efficiently provide in-time geological inverted models for real-time decisions. These available software platforms and successful field data from the ultradeep-reading tools initialized further studies and evaluations of the advanced, machine-learning algorithms applied into the existing inversion process. Based on a large database of ultradeep resistivity measurements from past successful field jobs and pre-well modeling, this paper presents several deep-learning algorithms to improve the existing inversion process for extracting more geological information (Payrazyan et al., 2017; Xu et al., 2019). The proposed methods identify similarities among numerous solutions attained by individual steps of the existing inversion process. Then, most likely distributions are detected within a detection range of the inversion to remove outlier signals and models, and to further produce more geologically reasonable representations. The proposed methods also enable automatic boundary-picking of layers with a major resistivity contrast between them. The determined connections learned from the previous sets of measurements are used to train any future processes based on a new set of measurements, enabling more efficient evaluations and calculations. Both modeling and field examples establish better geological interpolations acquired from the presented machine-learning algorithms than the original inversion approach. The use of the machine-learning concepts on the ultradeep resistivity measurements efficiently enhances the quality of the final geological interpretation over long detection distances into the formations. This enhancement benefits operations by optimizing reservoir development, maximizing assets, and reducing overall operational cost. Copyright 2019, held jointly by the Society of Petrophysicists and Well Log Analysts (SPWLA) and the submitting authors." "57208528190;55970576100;17434697300;57209911880;","The transparency of big data, data harvesting and digital Twins",2019,"10.1007/978-3-030-11289-9_6","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076088790&doi=10.1007%2f978-3-030-11289-9_6&partnerID=40&md5=336127021231df4d6177aa9b8d779ea5","Computer storage and cloud computing has become more powerful with multiple algorithms running complex data analysis looking at intelligence trends, user behaviour, profiling and ways to make use of these outputs. Added with the artificial intelligence (AI) interaction has meant a new and dynamic method to create models forging analysis to be more clinical, proficient and continually seeking more improvement with the self-learning and intelligent programming of machine learning (ML). In the healthcare sector there is deep interest in collecting, curating the data and making the best use of silo’d data through methods such as blockchain. This can then lead to a multitude of innovations such as precision based medicine, targeting individual variability in genes, their environment, etc. It also means that big data analytics in healthcare is evolving into providing these insights from very large data sets and improving outcomes while reducing costs and inefficiencies. However, there also are some ethical impacts in the process of Digital Twins which can lead to segmentation and discrimination. Or perhaps the data that is automatically collected from healthcare sensors in IoMT and what type of governance are they scrutinized to. It is clear that data is the most important asset of not just an organisation but also to the individual and why the General Data Protection Regulation (GDPR) has taken an important stance in data protection by design and default, that all organisations needs to follow. This chapter aims to highlight some of the concerns. © Springer Nature Switzerland AG 2019." "56292149100;57200597357;6602942477;","Development of an operational algorithm for automated deforestation mapping via the Bayesian integration of long-term optical and microwave satellite data",2019,"10.3390/rs11172038","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071993587&doi=10.3390%2frs11172038&partnerID=40&md5=d6d190f979b2cb6d1902ef9ee8bf3a47","The frequent fine-scale monitoring of deforestation using satellite sensors is important for the sustainable management of forests. Traditional optical satellite sensors suffer from cloud interruption, particularly in tropical regions, and recent active microwave sensors (i.e., synthetic aperture radar) demonstrate the difficulty in data interpretation owing to their inherent sensor noise and complicated backscatter features of forests. Although the sensor integration of optical and microwave sensors is of compelling research interest, particularly in the conduct of deforestation monitoring, this topic has not been widely studied. In this paper, we introduce an operational algorithm for automated deforestation mapping using long-term optical and L-band SAR data, including a simple time-series analysis of Landsat stacks and a multilayered neural network with Advanced Spaceborne Thermal Emission and Reflection Radiometer and Phased Array-type L-band Synthetic Aperture Radar-2, followed by sensor integration based on the Bayesian Updating of Land-Cover. We applied the algorithm over a deciduous tropical forest in Cambodia in 2003-2018 for validation, and the algorithm demonstrated better accuracy than existing approaches, which only depend on optical data or SAR data. Owing to the cloud penetration ability of SAR, observation gaps of optical data under cloudy conditions were filled, resulting in a prompter detection of deforestation even in the tropical rainy season. We also investigated the effect of posterior probability constraints in the Bayesian approach. The land-cover maps (forest/deforestation) created by the well-tuned Bayesian approach achieved 94.0% ± 4.5%, 80.0% ± 10.1%, and 96.4% ± 1.9% for the user's accuracy, producer's accuracy, and overall accuracy, respectively. In the future, small-scale commission errors in the resultant maps should be improved by using more sophisticated machine-learning approaches and considering the reforestation effects in the algorithm. The application of the algorithm to other landscapes with other sensor combinations is also desirable. © 2019 by the authors." "35173655800;55916474900;6506175565;7004175711;","Calibrating long-period variables as standard candles with machine learning",2019,"10.1093/mnras/sty3495","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067187700&doi=10.1093%2fmnras%2fsty3495&partnerID=40&md5=391b362f078591f5e5fb08ee9fb34d5b","Variable stars with well-calibrated period-luminosity (PL) relationships provide accurate distance measurements to nearby galaxies and are therefore a vital tool for cosmology and astrophysics. While these measurements typically rely on samples of Cepheid and RR-Lyrae stars, abundant populations of luminous variable stars with longer periods of 10-1000 d remain largely unused. We apply machine learning to derive a mapping between light-curve features of these variable stars and their magnitude to extend the traditional PL relation commonly used for Cepheid samples. Using photometric data for long-period variable stars in the Large Magellanic Cloud (LMC), we demonstrate that our predictions produce residual errors comparable to those obtained on the corresponding Cepheid population. We show that our model generalizes well to other samples by performing a blind test on photometric data from the Small Magellanic Cloud (SMC). Our predictions on the SMC again show small residual errors and biases, comparable to results that employ PL relations fitted on Cepheid samples. The residual biases are complementary between the long-period variable and Cepheid fits, which provides exciting prospects to better control sources of systematic error in cosmological distance measurements. We finally show that the proposed methodology can be used to optimize samples of variable stars as standard candles independent of any prior variable star classification. © 2018 The Author(s)." "57205427051;57205427084;57205427187;57205427288;","An applied machine learning approach to subsea asset inspection",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059991653&partnerID=40&md5=4c939405f4c6d9e1ad6b5cb5f590388f","Subsea infrastructure inspection campaigns are conducted on a fixed time schedule to ensure the integrity of offshore assets while guaranteeing compliance with government and other stakeholders' legislation. These thorough assessments of subsea assets are achieved by acquiring and processing survey results alongside the relevant engineering data and environmental conditions. The current approach is not ideal. The infrastructure operators would prefer to have more frequent, on-demand field visits and base the assessments on an up-to-date high resolution, precise asset model. Unfortunately, this has historically not been technically feasible. The underwater inspection vehicles moved relatively slow, the positioning accuracy was limited, and the underwater sensors' accuracy and resolution was not comparable with those used in air. This is not the case anymore. We now operate inspection vehicles traveling with speeds over six knots! We can control them remotely from the office-based command center. Next to the acoustic technology, the vehicles are additionally equipped with optic sensors, dynamically collecting point clouds, images and videos of unprecedented resolution. Gigabytes of data hits the vessel every few minutes. And that's where it used to end… Processing and, so called, ""Digital Twin"" modelling from this data has traditionally been taking weeks for every day of inspection. In many cases it was simply not technically possible since the data files were prohibitively big. Data required subsampling, more data processors had to go offshore, asset operators had to wait longer. Given the increasing capacity of satellite links, limitless potential of commoditized cloud computing and sophisticated Deep Learning algorithms, we are now able to move subsea asset management to the new era. Through one of their latest research & development programmes - Roames, Fugro has been able to revolutionize the traditional workflows using the state-of-the-art digital technology. The resultant product is a bespoke, web-based service that enables asset (initially pipelines) inspection data to be uploaded to the secure cloud environment, processed using Machine Learning, verified by experts on shore and visualized in an intuitive 4D web viewer. All delivered to the geographically spread stakeholders, operators and decision makers in near real time. This approach greatly reduces the cost of infrastructure management practices, lowers the risk exposure and contributes to extended life of the asset. The Roames, Machine Learning based methodology was validated using a vast archive of survey data prior to testing the workflow on a live project. Detailed method statement and field results are presented in this paper. © Copyright 2018, Society of Petroleum Engineers." "57205425235;57205425268;","Enabling the best by preparing for the worst: Lessons from disaster response for industrial IoT in oil and gas",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059972627&partnerID=40&md5=eceb60791925ee46536223aa632fd684","Objectives/Scope: As more oil and gas companies develop Internet of Things (IoT) strategies and beginning their digital transformation to Industry 4.0 or Smart Manufacturing, they face challenges in adopting technologies due to regulatory restrictions for highly combustible atmospheres such as exist in some of the world's largest and most critical industries - oil & gas, chemical, pharmaceutical, energy and others. In Zone 1 classified hazardous areas worldwide, up to 15% of personnel do not have access to mobile computing devices unless they are certified ""intrinsically safe,"" or incapable of causing a spark that could ignite a combustible environment. Thus, the human ""sensor"" in hazardous area operations, who could conceivably detect perceived anomalies or problems in the maintenance, workflow, process or function of these operations, is relegated to recording observations with pencil and paper and then entering data manually into ERP systems hours or days later. Such lack of real-time communication and data management results in inefficiency, increased costs and elevated safety and asset risk, causing potential down-time and even loss of life in extreme cases. Methods, Procedures, Process: By deploying new IoT technologies that allow people to use technology inside Zone 1 hazardous areas, humans can actively interact with machines in real time to dramatically improve productivity, safety and the bottom line in hazardous operations. A new style of IoT platform built especially for oil & gas hazardous area operations, would need to include various and affordable types of sensors to cover vast spaces, real-time communications, cloud computing, machine learning, rights management, security, big data storage, analytics and user-friendly visualization, all functioning in highly explosive conditions. This paper considers the advantages for productivity and safety of an IoT Platform for Hazardous Locations, based on hands-on research conducted by AegexTechnologies, Verizon, Nokia and multiple technology partners that tested various edge technologies with first responders in realistic disaster scenarios during two annual events, Operation Convergent Response 2017 (#OCR2017) and Operation Convergent Response 2018 (#OCR2018 - to take place 5-8 November 2018)). The events provide unparalleled opportunities to test IoTunder extreme conditions with real people, such as a staged refinery collapse caused by an earthquake. Results, Observations, Conclusions: #OCR2017 and #OCR2018 showed how enabling real-time communications and data management via cutting-edge technologies, such as intrinsically safe tablets and IoT sensors, can strategically assist first responders to better handle emergencies. The studies' results give insight into the need for continued collaboration on IoT capabilities that can better manage not only emergency response, but everyday operations in hazardous industries such as oil and gas. Novel/Additive Information: Equipping oil and gas facilities with pervasive, smart IoT data-sensing capabilities, and equipping oil and gas personnel with real-time communications and data management tools, could result in dramatic improvements in productivity, safety, emergency response and disaster mitigation. © Copyright 2018, Society of Petroleum Engineers." "29867490900;8213128600;6603354695;8213128500;57197087688;6603888005;","Pattern Recognition Scheme for Large-Scale Cloud Detection over Landmarks",2018,"10.1109/JSTARS.2018.2863383","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051767950&doi=10.1109%2fJSTARS.2018.2863383&partnerID=40&md5=12fc72c5a419b57417e2751d014371de","Landmark recognition and matching is a critical step in many image navigation and registration models for geostationary satellite services, as well as to maintain the geometric quality assessment in the instrument data processing chain of earth observation satellites. Matching the landmark accurately is of paramount relevance, and the process can be strongly impacted by the cloud contamination of a given landmark. This paper introduces a complete pattern recognition methodology able to detect the presence of clouds over landmarks using Meteosat second generation (MSG) data. The methodology is based on the ensemble combination of dedicated support vector machines dependent on the particular landmark and illumination conditions. This divide-and-conquer strategy is motivated by the data complexity and follows a physically based strategy that considers variability both in seasonality and illumination conditions along the day to split observations. In addition, it allows training the classification scheme with millions of samples at an affordable computational costs. The image archive was composed of 200 landmark test sites with near 7 million multispectral images that correspond to MSG acquisitions during 2010. Results are analyzed in terms of cloud detection accuracy and computational cost. We provide illustrative source code and a portion of the huge training data to the community. © 2018 IEEE." "6602574676;56227666500;56151374100;57061262300;36098286300;24166367300;7006393267;","Snow-covered area using machine learning techniques",2018,"10.1109/IGARSS.2018.8519443","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064194187&doi=10.1109%2fIGARSS.2018.8519443&partnerID=40&md5=5e1fe13d5a1d90894f14a226f6e1de07","In this study, we used an artificial neural network method to estimate the fractional snow cover area (fSCA), which is fast and accurate, and that can be easily adapted to different remote sensing instruments. We tested our approach using SnowEx data from NASA's Cloud Absorption Radiometer (CAR) over Grand Mesa; one of the largest flat-topped mountains in the world, which features sufficient forested stands with a range of density and height (and a variety of other forest conditions); a spread of snow depth/snow water equivalent conditions over sufficiently flat snowcovered terrain. The retrieved fractional snowcovered area from CAR compares reasonably with a Sentinel-2 image over the same location and demonstrates CAR's unique capability to improve the retrieval of snow properties using machine learning. The retrieved snow fraction parameter from our method is expected to minimize the error associated with the traditional binary snow detection scheme, and improve the retrieval quality of key parameters such as surface albedo. © 2018 IEEE." "13409639400;57203231038;7405934557;","The potential of sentinel satellites for large area aboveground forest biomass mapping",2018,"10.1109/IGARSS.2018.8517597","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063154821&doi=10.1109%2fIGARSS.2018.8517597&partnerID=40&md5=3f4f6582feadd31b318f5a0367b6ec07","Estimation of aboveground forest biomass is critical for regional carbon policies and sustainable forest management. Both passive optical remote sensing and active microwave remote sensing can play an important role in the monitoring of forest biomass. In this study, the recently launched Sentinel-2 Multi Spectral Instrument satellite and Sentinel-1 SAR satellite systems were evaluated and integrated to investigate the relative strengths of each sensor for mapping aboveground forest biomass at a regional scale. The Australian state of Victoria, with its wide range of forest vegetation was chosen as the study area to demonstrate the scalability and transferability of the approach. In this study aboveground forest biomass (AGB) was defined as the tons of carbon per hectare for the aboveground components (stem, branches, leaves) of all live large trees greater than 10 cm in diameter at breast height (DBHOB). Sentinel-2 and Sentinel-1 data were fused within a machine learning framework using a boosted regression tree model and high-quality ground survey data. Multi-criteria evaluations showed the use of the two independent and fundamentally different Sentinel satellite systems were able to provide robust estimates (R 2 of 0.62, RMSE of 32.2 t.C.ha -1 ) of aboveground forest biomass, with each sensor compensating for the weakness (cloud perturbations and spectral saturation for Sentinel 2, and sensitivity to ground moisture for Sentinel 1) of each other. As archives for Sentinel-2 and Sentinel-1 continue to grow, mapping aboveground forest biomass and dynamics at moderate resolution over large regions should become increasingly feasible. © 2018 IEEE" "55083178100;57202647645;57192690229;","Towards extraction of lianas from terrestrial lidar scans of tropical forests",2018,"10.1109/IGARSS.2018.8517634","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063144970&doi=10.1109%2fIGARSS.2018.8517634&partnerID=40&md5=0aef4f09fd61fe99b51a9e277003cc99","Increased liana abundance results in reduced tree growth and increased tree mortality in tropical forest. The impact of lianas on forest-wide carbon storage has been a special interest for many researchers. The vertical and horizontal spatial distribution of lianas in tropical forest will determine the interaction with trees and the forest carbon cycle. In this study, we will introduce an algorithm to extract lianas from terrestrial laser scanning (TLS) data of a tropical forest. We developed a classification method for separating liana points from other points in a point cloud under canopy. We used a Random Forests machine learning algorithm for the classification of liana points from the other points. The leaf-wood and liana-tree classification accuracies are 90.69% and 94.42%, respectively. The results show the potential of TLS data for analysis the spatial distribution of lianas in forest stands and we explore the potential of extracting lianas from TLS point clouds. © 2018 IEEE" "57208219543;57189034361;","Study on Feature Selection and Feature Deep Learning Model for Big Data",2018,"10.1109/ICSCSE.2018.00171","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065861133&doi=10.1109%2fICSCSE.2018.00171&partnerID=40&md5=3b9d777ba6856795dfa1c2ba04095318","In the era of big data, agricultural big data effectively mined in the agricultural cloud service platform helps to provide intelligent services for agricultural production and management[1]. Two big data feature selection methods are proposed based on the presence or absence labels of big data and potential value of data is mined better by effectively utilizing big data feature learning technologies. Transformation of data from data with primitive rough extraction features to data with features with stronger separability and high-level semantic features is of great significance for target task learning. © 2018 IEEE." "56416697700;38863569800;24544853800;14618921800;57191157218;7003503895;7005618744;57203162576;","Artificial intelligence based directional mesh network design for spectrum efficiency",2018,"10.1109/AERO.2018.8396558","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85049869729&doi=10.1109%2fAERO.2018.8396558&partnerID=40&md5=dab355616fd2f3799805985d66794970","The paper presents a novel directional mesh network (DMN) design that can distribute the limited radio spectrum resources more efficiently for a DMN by applying artificial intelligence machine learning (ML) techniques. The proposed DMN framework analyzes time-sensitive signal data close to the signal source using fog computing with different types of ML techniques. Depending on the computational capabilities of the fog nodes, different feature extraction methods such as energy detection, match filter, and cyclostationary detection are selected to optimize spectrum allocation. The proposed system also takes the antenna power gain into consideration, which can further reduce probability of detection and interference of the DMN system. Also, the computing nodes send the periodic signal summary which is much smaller than the original signal to the cloud so that the overall system spectrum source allocation strategies are dynamically updated. Instead of just detecting the spectrum holes for secondary users to transmit the signal, the proposed system can optimize the signal transmission path from the cloud to the end user under the interference and relay constraints. The distributed nodes can further improve the strategy based on the sensing information from the fog. Applying fog computing, the system is more adaptive to the local environment and robust to spectrum changes. It will significantly improve the network reliability, resiliency, and flexibility. Designing the proposed system doesn't necessary need change much of the current communication network platform. As most of the signal data is processed at the fog level, it further strengthens the system security by reducing the communication burden of the communications network. © 2018 IEEE." "57200418714;55557536600;57037229400;","COMPARISON of SINGLE and MULTI-SCALE METHOD for LEAF and WOOD POINTS CLASSIFICATION from TERRESTRIAL LASER SCANNING DATA",2018,"10.5194/isprs-annals-IV-3-217-2018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046755955&doi=10.5194%2fisprs-annals-IV-3-217-2018&partnerID=40&md5=2550243857f10376fe9d5c2d908d3212","The classification of leaf and wood points is an essential preprocessing step for extracting inventory measurements and canopy characterization of trees from the terrestrial laser scanning (TLS) data. The geometry-based approach is one of the widely used classification method. In the geometry-based method, it is common practice to extract salient features at one single scale before the features are used for classification. It remains unclear how different scale(s) used affect the classification accuracy and efficiency. To assess the scale effect on the classification accuracy and efficiency, we extracted the single-scale and multi-scale salient features from the point clouds of two oak trees of different sizes and conducted the classification on leaf and wood. Our experimental results show that the balanced accuracy of the multi-scale method is higher than the average balanced accuracy of the single-scale method by about 10 % for both trees. The average speed-up ratio of single scale classifiers over multi-scale classifier for each tree is higher than 30. © Authors 2018." "57202990666;57200762357;57203899789;57203899549;57203898383;","Enabling autonomous well optimization via using IoT-enabled devices and machine learning in bakken horizontal wells",2018,"10.2118/190955-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088405164&doi=10.2118%2f190955-ms&partnerID=40&md5=5673b223eb4a7e17e6e0f9a4fbba14f3","During the mid-twentieth century, Dr. Sam Gibbs developed math that continues today to serve as the basis for the downhole monitoring and control of many wells. However, analytical methods based on this calculation, including wave equation techniques, assume vertical, single-well pads with little to no side loading. Modern E&P development has shifted to multi-well horizontal well pads with shallow kickoffs uphole and backbuilds downhole with high dog-leg severity, which causes high friction and dynamic conditions in the pumping system. The Gibbs wave equation and fillage calculation do not take into account key wellbore forces such as mechanical or Coulomb friction arising from rod on tubing wear due to deviation in the well. Friction due to deviation can result in downhole dynamometer card shapes that distort fillage calculations and downhole card analysis. Additionally, most rod pump control systems are based on Programmable Logic Control systems (PLC). PLC systems are simple to program but are limited in their computation capabilities and are unable to accommodate sophisticated mathematics. Increased computational capabilities are required to execute higher-order mathematics that accurately calculate downhole parameters and enable well autonomy. One approach to driving autonomous well classification and optimization of setpoints is the deployment of a system that is capable of real-time analysis and higher-order mathematics. An Internet of Things (IoT) device with high-performance computational capabilities and direct communication with a cloud-based analytics software platform was developed with the capabilities to execute higher-order mathematics, artificial intelligence and machine learning on high resolution data, sampled in real-time from the rod pump control system. Equinor deployed this technology on 50 wells in the Bakken. The 50 wells chosen are highly representative of ""typical"" Bakken horizontal wells. The device was connected into the legacy rod pump controller via Modbus connection. Immediate differences in key downhole parameters were observed when comparing the results from the traditional rod pump controller to the IoT device. The higher-accuracy physics-based inputs feed into machine learning algorithms, which dynamically classify wells into key operating states of under-pumping, over-pumping, and dialed in. Using improved downhole information, Equinor was able to automate well optimization setpoint decisions, resulting in reduced well volatility, better pump efficiency, and increased pump fillage. Equinor was able to achieve these improvements while maintaining production in all cases. By identifying wells that were over-pumping and under-pumping to optimize SPM setpoints, Equinor was able to achieve higher efficiency outcomes with equal or increased production. For wells that were under-pumping, Equinor was able to increase oil production by up to 33%. For wells that were over-pumping, Equinor was able to decrease the number of strokes by 11% and increase pump efficiency by 14%. Copyright 2018, Society of Petroleum Engineers." "7101806774;8308170700;22233735600;35776720900;36718621600;55448544300;","Towards automation of satellite-based radar imagery for iceberg surveillance - Machine learning of ship and iceberg discrimination",2018,"10.4043/29130-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088073432&doi=10.4043%2f29130-ms&partnerID=40&md5=782f76cc42bb033cea8ba25972b77aef","Drifting icebergs can threaten navigation and marine operations and are prevalent in a number of regions that have active oil and gas exploration and development. Satellite synthetic aperture radar (SAR) is naturally applicable to map and monitor icebergs and sea ice due its ability to capture images day or night, as well as through cloud, fog and various wind conditions. There are several notable examples of its use to support operations, including Grand Banks, Barents Sea, offshore Greenland and Kara Sea. New constellations of satellites and the increasing volume of satellite data becoming available present a new paradigm for ice surveillance, in terms of persistence, reliability and cost. To fully extract the value of the data from these constellations, automation and cloud-based processing must be implemented. This will allow more timely and efficient processing, lowering monitoring costs by at least an order of magnitude. The increase in data persistence and processing capability allows large regions to be monitored daily for ice incursions, thus increasing safety and efficiency during offshore operations in those regions. The process of automating SAR-based iceberg surveillance involves creating a process flow that is robust and requires limited human intervention. The process flow involves land-masking, target detection, target discrimination and product dissemination. Land masking involves the removal of high-clutter land from the imagery to eliminate false detection from these locations. Target detection usually involves an adaptive threshold to separate true targets from the background ocean clutter. A constant false alarm rate (CFAR) is a standard technique used in radar image processing for this purpose. Target discrimination involves an examination of the distinct features of a target to determine if they match the features of icebergs, vessels or other ‘false alarms’ (e.g., marine wildlife, clutter). The final stage is the production of an output surveillance product, which can be a standard iceberg chart (e.g., MANICE) or something that can be ingested into a GIS system (e.g., ESRI shapefile, Google KML). The target discrimination phase is one of the most important phases because it provides feedback to operations about the presence of targets of interest (icebergs and vessels). The authors have used computer vision techniques successfully to train target classifiers. Standard techniques usually result in classifier accuracies of between 85%-95%, depending on the resolution of the SAR (higher resolutions produce more accurate results) and the availability of multiple polarizations. To see if new machine learning techniques could be applied to increase classifier accuracy, a dataset of 5000 ship and iceberg targets were extracted from Sentinel-1 multi-channel data (HH,HV). The images were collected in several regions (Greenland, Grand Banks, and Strait of Gibraltar). Validation either came by way of supporting information from the offshore operations, or was inferred by location. An online machine learning competition was hosted by Kaggle, a company that conducts online competitions on behalf of their clients. The detection data were made available by Kaggle to the broad internet community. Kaggle has a loyal following of data scientists who regularly participate in Kaggle competitions. The competition was hosted over a three-month period; over 3300 teams participated in the competition. The competition produced an improved classifier over standard computer vision techniques; the top three competitors had 4-5 stage classifiers that increased classification accuracy by approximately 5%. Copyright 2018, Offshore Technology Conference." "57204047405;57201703592;57203344412;57204049934;","IOT based advanced medicine dispenser integrated with an interactive web application",2018,"10.14419/ijet.v7i4.10.20704","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85054278045&doi=10.14419%2fijet.v7i4.10.20704&partnerID=40&md5=2a76b2622d3185fc060dc2e8da1a8ff4","Internet of Things (IOT) is a development of the internet which plays a major role in integrating human-machine interaction by allowing everyday objects to send and receive data in a variety of applications. Using IOT in healthcare monitoring provides an avenue for doctors and patients to interact and to track the dosage of medication administered. The paper presents an interactive, user friendly network integrated with an automated medicine dispenser which uses IOT, cloud computing and machine learning. The network was built on a python tornado framework with a front end developed using materialise CSS. The feasibility of this approach was validated by building a prototype and conducting a survey. © 2018 Authors." "56427952900;57203866180;57209970192;","Machine learning for the communication optimization in distributed systems",2018,"10.14419/ijet.v7i4.1.19491","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85053241855&doi=10.14419%2fijet.v7i4.1.19491&partnerID=40&md5=5ddd4c4a69dc0310a755262f48808fd6","The objective of the work that is presented in this paper was the problem of the communication optimization and detection of the issues of computing resources performance degradation [1, 2] with the usage of machine learning techniques. Computer networks transmit payload data and the meta-data from numerous sources towards vast number of destinations, especially in multitenant environments [3, 4]. Meta data describes the payload data and could be analyzed for anomalies detection in the communication patterns. Communication patterns depend on the payload itself and technical protocol used. The technical patterns are the research target as their analysis could spotlight the vulnerable behavior, for example: unusual traffic, extra load transported and etc. There was a big data used to train model with a supervised machine learning. Dataset was collected from the network interfaces of the distributed application infrastructure. Machine Learning tools had been retained from the cloud services provider - Amazon Web Services. The stochastic gradient descent technique was utilized for the model training, so that it could represent the communication patterns in the system. The learning target parameter was a packet length, the regression was performed to understand the relationship between packet meta-data (timestamp, protocol, the source server) and its length. The root mean square error calculation was applied to evaluate the learning efficiency. After model was prepared using training dataset, the model was tested with the test dataset and then applied on the target dataset (dataset for prediction) to check whether it was capable to detect anomalies. The experimental part showed the applicability of machine learning for the communication optimization in the distributed application environment. By means of the trained artificial intelligence model, it was possible to predict target parameters of traffic and computing resources usage with purpose to avoid service degradation. Additionally, one could reveal anomalies in the transferred traffic between application components. The application of techniques is envisioned in information security field and in the field of efficient network resources planning. Further research could be in application machine learning techniques for more complicated distributed environments and enlarging the number of protocols to prepare communication patterns. © 2018 Authors." "57201990332;57201661007;54894167400;57201974960;","Cold Chain Management Using Model Based Design, Machine Learning Algorithms and Data Analytics",2018,"10.4271/2018-01-1201","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85046654459&doi=10.4271%2f2018-01-1201&partnerID=40&md5=8f676dd5b40e4c7bc79d193d92bafd75","In the food industry, there is an increased demand for generic pharmaceutical products and perishable food without compromising with the changes in texture and taste that occur in the transit. With this demand, there is a need for better visibility of products in the logistics network, to minimize wastage, to ensure product integrity, influence productivity, transparently track the fleet and to identify pathogens before a potential outbreak. In Cold Chain Management, information is power: with potentially billions of dollars' worth of cargo (such as food items, vaccines, serums, tests or chemicals) at stake worldwide. Hence, careful live monitoring, inspection, supervision, validation and documentation of business-critical information is essential. In this paper, we have proposed a framework for Cold Chain Management using Internet of Things (IoT) combined with other technological innovations such as: Cloud Computing, Machine Learning and Big Data Analytics to revolutionize the cold transport industry. By establishing such an architecture, we have tried to monitor, visualize, track and control various platform dependent parameters thereby providing a complete solution across the fleet cycle with assured freshness and palpability. © 2018 SAE International. All Rights Reserved." "22941140200;57200795580;55885604100;","Comparison of the filtering models for airborne LiDAR data by three classifiers with exploration on model transfer",2018,"10.1117/1.JRS.12.016021","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85042439463&doi=10.1117%2f1.JRS.12.016021&partnerID=40&md5=84b019e516d3fa422e055685320d277a","This paper discusses airborne light detection and ranging (LiDAR) point cloud filtering (a binary classification problem) from the machine learning point of view. We compared three supervised classifiers for point cloud filtering, namely, Adaptive Boosting, support vector machine, and random forest (RF). Nineteen features were generated from raw LiDAR point cloud based on height and other geometric information within a given neighborhood. The test datasets issued by the International Society for Photogrammetry and Remote Sensing (ISPRS) were used to evaluate the performance of the three filtering algorithms; RF showed the best results with an average total error of 5.50%. The paper also makes tentative exploration in the application of transfer learning theory to point cloud filtering, which has not been introduced into the LiDAR field to the authors' knowledge. We performed filtering of three datasets from real projects carried out in China with RF models constructed by learning from the 15 ISPRS datasets and then transferred with little to no change of the parameters. Reliable results were achieved, especially in rural area (overall accuracy achieved 95.64%), indicating the feasibility of model transfer in the context of point cloud filtering for both easy automation and acceptable accuracy. © 2018 Society of Photo-Optical Instrumentation Engineers (SPIE)." "57194471739;57200559318;57015781300;57202084374;","Resource allocation of cloud application through machine learning: A case study",2017,"10.1109/ICGI.2017.52","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041695403&doi=10.1109%2fICGI.2017.52&partnerID=40&md5=17881c3c968b8dabf4ab555b41b238fc","With the rapid development of WEB applications, the demand for dynamically adjusting computing resources based on the load variation is increasing. However, most of the traditional WEB systems have limited ability to respond to load changes. In order to solve the problem, software self-Adaptation technology has been applied to the resource management of WEB systems. Many researchers have tried to propose various software self-Adaptation models, all of which contain the control loop 'Monitor-Analyze-Plan-Execute'. However, the 'Knowledge-base' of these models is based on predefined strategy or configuration, which increases the complexity of system development and maintenance. In this paper, machine learning is applied to software self-Adaptation through a case study, and the 'Knowledge-base' is given by machine learning, which greatly reduces the workload of system maintenance and rule configuration. The results show the case based on machine learning can construct the corresponding management strategy according to WEB system runtime status, and conduct software self-Adaptive management. © 2017 IEEE." "35094561600;57016429700;57192366005;13410467900;22952160500;","Demonstrating PlanetSense: Gathering geo-spatial intelligence from crowd-sourced and social-media data",2016,"10.1145/2996913.2996975","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85011030479&doi=10.1145%2f2996913.2996975&partnerID=40&md5=b8cb71c65f605744f3650d8259cefa4f","Crowd-sourced and volunteered information, social media, and participatory sensors are capable of providing real-time activity data. Monitoring these sources in time of relevance and then using them to gather operational knowledge is important during crisis management. Beyond that, it's important to curate this information for geo-spatial research purposes, including land use classification and population occupancy analysis. In this demonstration, we will showcase PlanetSense - a geo-spatial research platform built to harness the existing power of archived data and add to that, the dynamics of heterogeneous real-time streaming data from social media and volunteered sources, seamlessly integrated with sophisticated machine learning algorithms and visualization tools. A demonstration will focus on - 1) Recent initiative emphasizing the need to harness crowdsources, volunteered, and social media data at scale; 2) Anatomy and insight into data collection workflow. We will show the ability to harvest and process several terabytes of raw data in realtime; 3) A detailed discussion with insight into more than 20 sources of data will be given. These sources include text, sensors, as well as imagery data; 4) PlanetSense's end to end distributed architecture will be discussed with focus on collecting and processing high-volumes of streaming data in a Geo-Data Cloud. Data fusion methods and algorithms for integrating disparate data sources with existing legacy products. Data analytics and machine learning methods for generating operational intelligence on the fly; 5) In addition, PlanetSense ""App"" platform will be shown with hands-on application enabling interested audience to quickly develop and deploy solutions. 6) Several case studies will be discussed relevant to, land use classification, monitoring transient population, high-resolution occupancy analysis, mapping special events population, ability to uncover global breaking events and reactions in near-real time, ability to track protest, unrest, and monitor other societal turbulences as they happen, and real-time monitoring of infrastructure outages. © 2016 ACM." "57191839724;55648722300;55105432400;","Automated equipment recognition and classification from scattered point clouds for construction management applications",2016,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84994323171&partnerID=40&md5=77dba3ba80adb2df840d73eb46f7b6c2","Recognizing construction assets from as-is point cloud data of construction environment provides essential information for engineering and management applications including progress monitoring, safety management, supply-chain management, and quality control. This study proposes a fast and automated processing pipeline for construction target assets recognition from scattered as-is point clouds. The recognition tasks can be subdivided into object detection, which involves computing the bounding box around each construction equipment, and object classification, which involves labelling point cloud clusters from discrete equipment categories, such as backhoe loader, bulldozer, dump truck, excavator and front loader. The object detection step consists of point cloud down-sampling, segmentation and clustering. For the object classification step, machine learning methods were employed to determine class membership probability using features derived from the ESF (Ensemble of Shape Functions) descriptor. The classifiers were trained on synthetic point clouds generated from CAD (Computer-aided Design) models. The method was validated using laser scanned point clouds from an equipment yard. The test results demonstrate promising advancements towards semantic labelling and scene understanding of point cloud data." "56330233100;","Predictive policing",2014,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84906231906&partnerID=40&md5=66c4231e00cd8e96a480ae16c6031b2a","The article examines how machine learning can be combined with geospatial technologies to predict where crimes will happen next. Current law enforcement trends reflect an interest in predictive analysis, increasing adoption of cloud services, pervasive location information and the proliferation of mobile devices. These trends are emerging in extremely constrained budget environments for many agencies and communities. Azavea recently set out to take advantage of these trends by combining geographic analysis with contemporary machine learning algorithms and cloud computing to help police departments make more effective use of their limited resources. The team developing HunchLab sought not only to implement the yearly warning system represented by the original prototypes, but also to use crime patterns to predict areas at higher than normal risk. The goal was to help officers prevent spikes in crime from even occurring. The team reviewed the academic literature and decided to implement two analytic techniques: near-repeat pattern analysis and cyclic-load forecasting." "55932962700;6506085008;56516456400;8984213200;57217293480;","A new framework for classification of distributed denial of service (DDOS) attack in cloud computing by machine learning techniques",2014,"10.1166/asl.2014.5288","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84887857071&doi=10.1166%2fasl.2014.5288&partnerID=40&md5=c37e2c58568fa5442cc93480d7645401","The growing nature of cloud computing, increases the urgency of appropriate solution for its relevant threats. Security of cloud is the top challenge among all challenges that cloud computing deals with. Indeed, security issues of cloud have a reverse relation with cloud users' desires to move their assets to cloud environment. Distributed Denial of Service (DDOS) attack is one of the most popular attacks for security experts, which is easy to be carried out; however, it has some catastrophic effects. On the other hand, if a kind of attack like DDOS happens in cloud environment, cloud providers are not eager to share this information with their users due to loss of reputation. However, cloud users demand a mechanism to find out whether their assets are secure or not. To address the aforementioned problem, in this paper a new framework for classification of DDOS attack by utilization of machine learning techniques has been proposed. For this purpose, we will collect a dataset from the resources that are dedicated to user's virtual machine. This file will be delivered to a machine learning tool for future activities. © 2014 American Scientific Publishers All rights reserved." "14831081900;57219272291;","A two-moment machine learning parameterization of the autoconversion process",2021,"10.1016/j.atmosres.2020.105269","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091987463&doi=10.1016%2fj.atmosres.2020.105269&partnerID=40&md5=1ef132ae04b9cd1b3f16472ed3806bf6","Autoconversion is the mass transfer from cloud to precipitation water in an early stage of cloud development, and is the dominant process in the formation of embryonic droplets that trigger precipitation formation. The accurate parameterization of this process is key, in order to improve the interaction between cloud microphysics and cloud dynamics for models from cloud scale to the global climate scale. For model based parameterizations of the auto-conversion process, the usual approach to develop an autoconversion parameterization is by curve fitting the autoconversion rates obtained from simulations or numerical solutions of the kinetic collection equation under a wide range of initial conditions. However, in this case, the autoconversion is modeled by a function that is a nonlinear product of liquid water content and droplet concentration and depends on a small number of parameters. As a result, a large amount of scatter around the actual values can be obtained, indicating a weak relationship between actual and fitted autoconversion rates. The purpose of this paper is to analyze whether neural networks are better than traditional curve fitting or regression to obtain parameterizations of autoconversion. Then, a deep neural network was trained from an autconversion rates dataset generated by solving the kinetic collection equation for a wide range of droplet concentrations and liquid water contents. The obtained machine learned parameterization shows a very good match with actual rates calculated from the kinetic collection equation. © 2020 Elsevier B.V." "57219955783;36620043400;57215559368;57213355177;57216965386;","Extraction of built-up area using multi-sensor data—A case study based on Google earth engine in Zhejiang Province, China",2021,"10.1080/01431161.2020.1809027","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096244274&doi=10.1080%2f01431161.2020.1809027&partnerID=40&md5=57ed9e48eaf1e8ff57f62890d363981c","Accurate and up-to-date built-up area mapping is of great importance to the science community, decision-makers, and society. Therefore, satellite-based, built-up area (BUA) extraction at medium resolution with supervised classification has been widely carried out. However, the spectral confusion between BUA and bare land (BL) is the primary hindering factor for accurate BUA mapping over large regions. Here we propose a new methodology for the efficient BUA extraction using multi-sensor data under Google Earth Engine cloud computing platform. The proposed method mainly employs intra-annual satellite imagery for water and vegetation masks, and a random-forest machine learning classifier combined with auxiliary data to discriminate between BUA and BL. First, a vegetation mask and water mask are generated using NDVI (normalized differenced vegetation index) max in vegetation growth periods and the annual water-occurrence frequency. Second, to accurately extract BUA from unmasked pixels, consisting of BUA and BL, random-forest-based classification is conducted using multi-sensor features, including temperature, night-time light, backscattering, topography, optical spectra, and NDVI time-series metrics. This approach is applied in Zhejiang Province, China, and an overall accuracy of 92.5% is obtained, which is 3.4% higher than classification with spectral data only. For large-scale BUA mapping, it is feasible to enhance the performance of BUA mapping with multi-temporal and multi-sensor data, which takes full advantage of datasets available in Google Earth Engine. © 2020 Informa UK Limited, trading as Taylor & Francis Group." "36117793500;16178105300;6504046934;","Toward making canopy hemispherical photography independent of illumination conditions: A deep-learning-based approach",2021,"10.1016/j.agrformet.2020.108234","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85095700535&doi=10.1016%2fj.agrformet.2020.108234&partnerID=40&md5=17ed4e609f91323e12beefc7f222ee16","Hemispherical photography produces the most accurate results when working with well-exposed photographs acquired under diffuse light conditions (diffuse-light images). Obtaining such data can be prohibitively expensive when surveying hundreds of plots is required. A relatively inexpensive alternative is using photographs acquired under direct sunlight (sunlight images). However, this practice leads to high errors since the standard processing algorithms expect diffuse-light imagery. Here, instead of using classification algorithms, which is the unique dominant practice, we approached the processing of sunlight images using deep learning (DL) regression. We implemented DL systems by using the general-purpose convolutional neural networks known as VGGNet 16, VGGNet 19, Res-Net, and SE-ResNet. We trained them with 608 samples acquired in a South American temperate forest populated by Nothofagus pumilio. For their evaluation, we used 113 independent samples. Each sample (X, Y) consisted of one or several sunlight images (X), and the plant area index (PAI) and effective PAI (PAIe) extracted from a diffuse-light image (Y). The sunlight images include clear sky and broken clouds with sun elevation from 15° to 47°. We obtained the best results with the SE-ResNet architecture. The system requires a low-resolution input reprojected to cylindrical, and it can make predictions with 10% root mean square error, even from pictures acquired with automatic exposure, which challenge previous findings. Furthermore, similar results (R2= 0.9, n = 104) can be obtained by feeding the system with photographs acquired with an inexpensive fisheye converter attached to a smartphone. Altogether, results indicate that our approach is a cost-efficient option for surveying hundreds of plots under direct sunlight. Therefore, combining our method with the traditional procedures provides processing solutions for virtually all kinds of illumination conditions. © 2020" "57214306009;55782832600;57189906017;","Roof bolt identification in underground coal mines from 3D point cloud data using local point descriptors and artificial neural network",2021,"10.1080/2150704X.2020.1809734","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096040035&doi=10.1080%2f2150704X.2020.1809734&partnerID=40&md5=82ebc9a4117da16f3e2c6c298cf7930d","Roof bolts are commonly used to provide structural support in underground mines. Frequent and automated assessment of roof bolt is critical to closely monitor any change in the roof conditions while preventing major hazards such as roof fall. However, due to challenging conditions at mine sites such as sub-optimal lighting and restrictive access, it is difficult to routinely assess roof bolts by visual inspection or traditional surveying. To overcome these challenges, this study presents an automated method of roof bolt identification from 3D point cloud data, to assist in spatio-temporal monitoring efforts at mine sites. An artificial neural network was used to classify roof bolts and extract them from 3D point cloud using local point descriptors such as the proportion of variance (POV) over multiple scales, radial surface descriptor (RSD) over multiple scales and fast point feature histogram (FPFH). Accuracy was evaluated in terms ofprecision, recall and quality metric generally used in classification studies. The generated results were compared against other machine learning algorithms such as weighted k-nearest neighbours (k-NN), ensemble subspace k-NN, support vector machine (SVM) and random forest (RF), and was found to be superior by up to 8% in terms of the achieved quality metric. © 2020 Informa UK Limited, trading as Taylor & Francis Group." "57210946877;56770984700;6701474321;","A machine learning approach for the detection of supporting rock bolts from laser scan data in an underground mine",2021,"10.1016/j.tust.2020.103656","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092391600&doi=10.1016%2fj.tust.2020.103656&partnerID=40&md5=9d04bdc4a5e8c3acecae449065671350","Rock bolts are a crucial part of underground infrastructure support; however, current methods to locate and record their positions are manual, time consuming and generally incomplete. This paper describes an effective method to automatically locate supporting rock bolts from a 3D laser scanned point cloud. The proposed method utilises a machine learning classifier combined with point descriptors based on neighbourhood properties to classify all data points as either ‘bolt’ or ‘not-bolt’ before using the Density Based Spatial Clustering of Applications with Noise (DBSCAN) algorithm to divide the results into candidate bolt objects. The centroids of these objects are then computed and output as simple georeferenced 3D coordinates to be used by surveyors, mine managers and automated machines. Two classifiers were tested, a random forest and a shallow neural network, with the neural network providing the more accurate results. Alongside the different classifiers, different input feature types were also examined, including the eigenvalue based geometric features popular in the remote sensing community and the point histogram based features more common in the mobile robotics community. It was found that a combination of both feature sets provided the strongest results. The obtained precision and recall scores were 0.59 and 0.70 for the individual laser points and 0.93 and 0.86 for the bolt objects. This demonstrates that the model is robust to noise and misclassifications, as the bolt is still detected even if edge points are misclassified, provided that there are enough correct points to form a cluster. In some cases, the model can detect bolts which are not visible to the human interpreter. © 2020" "57204794658;6603846580;57215661230;57214862416;54421097700;57163606900;57194165611;57193830300;56786235200;","Soil variability and quantification based on Sentinel-2 and Landsat-8 bare soil images: A comparison",2021,"10.1016/j.rse.2020.112117","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091914235&doi=10.1016%2fj.rse.2020.112117&partnerID=40&md5=55094a26d6493e88c53586c0e3024e55","There is a worldwide need for detailed spatial information to support soil mapping, mainly in the tropics, where main agricultural areas are concentrated. In this line, satellite images are useful tools that can assist in obtaining soil information from a synoptic point of view. This study aimed at evaluating how satellite images at different resolutions (spatial, spectral and temporal) can influence the representation of soil variability over time, the percentage of bare soil areas and spatial predictions of soil properties in southeastern Brazil. We used single-date and multi-temporal images (SYSI, Synthetic Soil Images) of bare soil pixels from the Sentinel2-MultiSpectral Instrument (S2-MSI) and the Landsat-8 Operational Land Imager (L8-OLI) to conduct this research. Two SYSIs were obtained from images acquired in four years (2016–2019) for each satellite (SYSI S2-MSI and SYSI L8-OLI) and a third SYSI, named SYSI Combined, was obtained by combining the images from both satellites. The single-date images for each satellite was acquired in September, when the influence of clouds was low and bare soil pixels was predominant. Single-date images and SYSIs were compared by means of their spectral patterns and ability to predict topsoil properties (clay, sand, silt, and organic matter contents and soil color) using the Cubist algorithm. We found that the SYSIs outperformed single-date images and that the SYSI Combined and SYSI L8-OLI provided the best prediction performances. The SYSIs also had the highest percentage of areas with bare soil pixels (~30–50%) when compared to the single-date images (~20%). Our results suggest that bare soil images obtained by combining Landsat-8 and Sentinel-2 images are more important for soil mapping than spatial or spectral resolutions. Soil maps obtained via satellite images are important tools for soil survey, land planning, management and precision agriculture. © 2020 Elsevier Inc." "25931139100;6701481007;6506992267;","Global monitoring of deep convection using passive microwave observations",2021,"10.1016/j.atmosres.2020.105244","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090568601&doi=10.1016%2fj.atmosres.2020.105244&partnerID=40&md5=ea1918a664ad2600dc803abe0b4b1d09","In this study, we present the DEEPSTORM (DEEP moiSt aTmospheric cOnvection from micRowave radioMeter) algorithm, able to retrieve ice water path (IWP) and to detect deep moist atmospheric convection (DC) from 80°S to 80°N using observations from four spaceborne passive microwave radiometers. DEEPSTORM is based on a machine learning approach and is fitted against observations from the CPR (Cloud Profiling Radar) spaceborne radar on-board CloudSat. IWP predictions show an average root mean square error of 0.27 kg/m2 and a correlation index of 0.87. DC occurrence is detected with a probability of 59% and a false alarm rate of 24%. The prediction accuracy of IWP and DC is significantly better when the IWP exceeds 0.5 kg/m2 showing that DEEPSTORM is well suited to detect and characterise the strongest DC events. Overall DC detection is more accurate in the tropics than in mid-latitudes while the IWP retrieval works better in the mid-latitudes. Two examples illustrating the potential of DEEPSTORM are presented: the IWP is retrieved during Hurricane Matthew in 2016, and a climatology of DC occurrences between September 2016 and December 2016 is presented. This work will allow building a quasi-worldwide and 20-year long database of DC occurrence and intensity. © 2020 Elsevier B.V." "57208864702;57201315852;7401683250;35337273500;36136336500;57190869353;12781228700;57204719415;57205301324;","Observations of single-stroke flashes from five isolated small thunderstorms in East China",2020,"10.1016/j.jastp.2020.105441","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092165196&doi=10.1016%2fj.jastp.2020.105441&partnerID=40&md5=a044bcc6bdb6384824f4406653f283fb","This paper presents the ground truth datasets of negative cloud-to-ground (NCG) flashes with one single stroke in five isolated small thunderstorms. By developing a machine-learning method based on the convolutional neural network, we identify the return strokes during the whole life of thunderstorms with the wideband electric field waveform of lightning discharges detected by Jianghuai Area Sferic Array. The distribution of flash multiplicity in these thunderstorms exhibits an exponential decrease pattern. The single-stroke flashes (with multiplicity = 1) account for more than 30% of all NCG flashes on a thunderstorm basis, and the proportion of single-stroke flashes tended to be most abundant. When observed at a 20-min interval, the percentage of single-stroke flashes tends to change dramatically during thunderstorm developments, which seems to show an opposite trend with time to the maximum flash multiplicity. Single-stroke flashes tended to be associated with a weaker peak current of initial stroke compared to that of multiple-stroke flashes. It is inferred that the horizontal scale of negative charge regions in thunderclouds might play an important role in enhancing the flash multiplicity. © 2020 Elsevier Ltd" "55939190800;57053470200;35518435600;57211533913;56245008400;22960711400;57150278600;16029719200;13905662200;7007022537;51863613500;","Ensemble-based deep learning for estimating PM2.5 over California with multisource big data including wildfire smoke",2020,"10.1016/j.envint.2020.106143","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091646921&doi=10.1016%2fj.envint.2020.106143&partnerID=40&md5=2e5b08366201ecce30c17488798fef97","Introduction: Estimating PM2.5 concentrations and their prediction uncertainties at a high spatiotemporal resolution is important for air pollution health effect studies. This is particularly challenging for California, which has high variability in natural (e.g, wildfires, dust) and anthropogenic emissions, meteorology, topography (e.g. desert surfaces, mountains, snow cover) and land use. Methods: Using ensemble-based deep learning with big data fused from multiple sources we developed a PM2.5 prediction model with uncertainty estimates at a high spatial (1 km × 1 km) and temporal (weekly) resolution for a 10-year time span (2008–2017). We leveraged autoencoder-based full residual deep networks to model complex nonlinear interrelationships among PM2.5 emission, transport and dispersion factors and other influential features. These included remote sensing data (MAIAC aerosol optical depth (AOD), normalized difference vegetation index, impervious surface), MERRA-2 GMI Replay Simulation (M2GMI) output, wildfire smoke plume dispersion, meteorology, land cover, traffic, elevation, and spatiotemporal trends (geo-coordinates, temporal basis functions, time index). As one of the primary predictors of interest with substantial missing data in California related to bright surfaces, cloud cover and other known interferences, missing MAIAC AOD observations were imputed and adjusted for relative humidity and vertical distribution. Wildfire smoke contribution to PM2.5 was also calculated through HYSPLIT dispersion modeling of smoke emissions derived from MODIS fire radiative power using the Fire Energetics and Emissions Research version 1.0 model. Results: Ensemble deep learning to predict PM2.5 achieved an overall mean training RMSE of 1.54 μg/m3 (R2: 0.94) and test RMSE of 2.29 μg/m3 (R2: 0.87). The top predictors included M2GMI carbon monoxide mixing ratio in the bottom layer, temporal basis functions, spatial location, air temperature, MAIAC AOD, and PM2.5 sea salt mass concentration. In an independent test using three long-term AQS sites and one short-term non-AQS site, our model achieved a high correlation (>0.8) and a low RMSE (<3 μg/m3). Statewide predictions indicated that our model can capture the spatial distribution and temporal peaks in wildfire-related PM2.5. The coefficient of variation indicated highest uncertainty over deciduous and mixed forests and open water land covers. Conclusion: Our method can be generalized to other regions, including those having a mix of major urban areas, deserts, intensive smoke events, snow cover and complex terrains, where PM2.5 has previously been challenging to predict. Prediction uncertainty estimates can also inform further model development and measurement error evaluations in exposure and health studies. © 2020 The Author(s)" "57218881802;56531367400;7005742190;55628589750;56203143700;","Determinants of fog and low stratus occurrence in continental central Europe – a quantitative satellite-based evaluation",2020,"10.1016/j.jhydrol.2020.125451","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090574057&doi=10.1016%2fj.jhydrol.2020.125451&partnerID=40&md5=c950c27b3e1f9a37b765a645a8fb51c8","The formation and development of fog and low stratus clouds (FLS) depend on meteorological and land surface conditions and their interactions with each other. While analyses of temporal and spatial patterns of FLS in Europe exist, the interactions between FLS determinants underlying them have not been studied explicitly and quantitatively at a continental scale yet. In this study, a state-of-the-art machine learning technique is applied to model FLS occurrence over continental Europe, using meteorological and land surface parameters from geostationary satellite and reanalysis data. Spatially explicit model units are created to test for spatial and seasonal differences in model performance and FLS sensitivities to changes in predictors, and effects of different data preprocessing procedures are evaluated. The statistical models show good performance in predicting FLS occurrence during validation, with R2>0.9 especially in winter high pressure situations.The predictive skill of the models seems to be dependent on data availability, data preprocessing, time period, and geographic characteristics. It is shown that atmospheric proxies are more important determinants of FLS presence than surface characteristics, in particular mean sea level pressure, near-surface wind speed and evapotranspiration are crucial, together with FLS occurrence on the previous day. Higher wind speeds, higher land surface temperatures and higher evapotranspiration tend to be negatively related to FLS. Spatial patterns of feature importance show the dominant influence of mean sea level pressure on FLS occurrence throughout the central European domain. When only high pressure situations are considered, wind speed (in the western study region) and evapotranspiration (in the eastern study region) gain importance, highlighting the influence of moisture advection on FLS occurrence in the western parts of the central European domain. This study shows that FLS occurrence can be accurately modeled using machine learning techniques in large spatial domains based on meteorological and land surface predictors. The statistical models used in this study provide a novel analysis tool for investigating empirical relationships in the FLS – land surface system and possibly infer processes. © 2020 Elsevier B.V." "36617930000;57204769007;35517567400;57212508254;57218645527;57218652245;57218652947;","PODPAC: open-source Python software for enabling harmonized, plug-and-play processing of disparate earth observation data sets and seamless transition onto the serverless cloud by earth scientists",2020,"10.1007/s12145-020-00506-0","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089885051&doi=10.1007%2fs12145-020-00506-0&partnerID=40&md5=8909c5c52e617c6c6133947b87cd7de3","In this paper, we present the Pipeline for Observational Data Processing, Analysis, and Collaboration (PODPAC) software. PODPAC is an open-source Python library designed to enable widespread exploitation of NASA earth science data by enabling multi-scale and multi-windowed access, exploration, and integration of available earth science datasets to support analysis and analytics; automatic accounting for geospatial data formats, projections, and resolutions; simplified implementation and parallelization of geospatial data processing routines; standardized sharing of data and algorithms; and seamless transition of algorithms and data products from local development to distributed, serverless processing on commercial cloud computing environments. We describe the key elements of PODPAC’s architecture, including Nodes for unified encapsulation of disparate scientific data sources; Algorithms for plug-and-play processing and harmonization of multiple data source Nodes; and Lambda functions for serverless execution and sharing of new data products via the cloud. We provide an overview of our open-source code implementation and testing process for development and deployment of PODPAC. We describe our interactive, JupyterLab-based end-user documentation including quick-start examples and detailed use case studies. We conclude with examples of PODPAC’s application to: encapsulate data sources available on Amazon Web Services (AWS) Open Data repository; harmonize processing of multiple earth science data sets for downscaling of NASA Soil Moisture Active Passive (SMAP) soil moisture data; and deploy a serverless SMAP-based drought monitoring application for use access from mobile devices. We postulate that PODPAC will also be an effective tool for wrangling and standardizing massive earth science data sets for use in model training for machine learning applications. © 2020, Springer-Verlag GmbH Germany, part of Springer Nature." "8575756400;57193212462;57217205306;57217202313;8510127900;","Enabling radiation tolerant heterogeneous GPU-based onboard data processing in space",2020,"10.1007/s12567-020-00321-9","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086737276&doi=10.1007%2fs12567-020-00321-9&partnerID=40&md5=5a642eef8e8bc1c608758f0a19bd9290","The last decade has seen a dramatic increase in small satellite missions for commercial, public, and government intelligence applications. Given the rapid commercialization of constellation-driven services in Earth Observation, situational domain awareness, communications including machine-to-machine interface, exploration etc., small satellites represent an enabling technology for a large growth market generating truly Big Data. Examples of modern sensors that can generate very large amounts of data are optical sensing, hyperspectral, Synthetic Aperture Radar (SAR), and Infrared imaging. Traditional handling and downloading of Big Data from space requires a large onboard mass storage and high bandwidth downlink with a trend towards optical links. Many missions and applications can benefit significantly from onboard cloud computing similarly to Earth-based cloud services. Hence, enabling space systems to provide near real-time data and enable low latency distribution of critical and time sensitive information to users. In addition, the downlink capability can be more effectively utilized by applying more onboard processing to reduce the data and create high value information products. This paper discusses current implementations and roadmap for leveraging high performance computing tools and methods on small satellites with radiation tolerant hardware. This includes runtime analysis with benchmarks of convolutional neural networks and matrix multiplications using industry standard tools (e.g., TensorFlow and PlaidML). In addition, a ½ CubeSat volume unit (0.5U) (10 × 10 × 5 cm3) cloud computing solution, called SpaceCloud™ iX5100 based on AMD 28 nm APU technology is presented as an example of heterogeneous computer solution. An evaluation of the AMD 14 nm Ryzen APU is presented as a candidate for future advanced onboard processing for space vehicles. © 2020, The Author(s)." "26422389600;57208228475;6602942477;35473149000;","Landslide detection in mountainous forest areas using polarimetry and interferometric coherence",2020,"10.1186/s40623-020-01191-5","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084662765&doi=10.1186%2fs40623-020-01191-5&partnerID=40&md5=588acb58605878691f49f9ccdb777f5e","The cloud-free, wide-swath, day-and-night observation capability of synthetic aperture radar (SAR) has an important role in rapid landslide monitoring to reduce economic and human losses. Although interferometric SAR (InSAR) analysis is widely used to monitor landslides, it is difficult to use that for rapid landslide detection in mountainous forest areas because of significant decorrelation. We combined polarimetric SAR (PolSAR), InSAR, and digital elevation model (DEM) analysis to detect landslides induced by the July 2017 Heavy Rain in Northern Kyushu and by the 2018 Hokkaido Eastern Iburi Earthquake. This study uses fully polarimetric L-band SAR data from the ALOS-2 PALSAR-2 satellite. The simple thresholding of polarimetric parameters (alpha angle and Pauli components) was found to be effective. The study also found that supervised classification using PolSAR, InSAR, and DEM parameters provided high accuracy, although this method should be used carefully because its accuracy depends on the geological characteristics of the training data. Regarding polarimetric configurations, at least dual-polarimetry (e.g., HH and HV) is required for landslide detection, and quad-polarimetry is recommended. These results demonstrate the feasibility of rapid landslide detection using L-band SAR images.[Figure not available: see fulltext.] © 2020, The Author(s)." "57219925946;57217865914;","Using machine learning to derive cloud condensation nuclei number concentrations from commonly available measurements",2020,"10.5194/acp-20-12853-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096103971&doi=10.5194%2facp-20-12853-2020&partnerID=40&md5=4a8649af36ec39f211d06d7e193b37d7","Cloud condensation nuclei (CCN) number concentrations are an important aspect of aerosol-cloud interactions and the subsequent climate effects; however, their measurements are very limited. We use a machine learning tool, random decision forests, to develop a random forest regression model (RFRM) to derive CCN at 0.4% supersaturation ([CCN0.4]) from commonly available measurements. The RFRM is trained on the long-Term simulations in a global size-resolved particle microphysics model. Using atmospheric state and composition variables as predictors, through associations of their variabilities, the RFRM is able to learn the underlying dependence of [CCN0.4] on these predictors, which are as follows: eight fractions of PM2:5 (NH4, SO4, NO3, secondary organic aerosol (SOA), black carbon (BC), primary organic carbon (POC), dust, and salt), seven gaseous species (NOx , NH3, O3, SO2, OH, isoprene, and monoterpene), and four meteorological variables (temperature (T), relative humidity (RH), precipitation, and solar radiation). The RFRM is highly robust: it has a median mean fractional bias (MFB) of 4:4% with 96:33% of the derived [CCN0.4] within a good agreement range of-60% MFB C60% and strong correlation of Kendall's coefficient 0:88. The RFRM demonstrates its robustness over 4 orders of magnitude of [CCN0.4] over varying spatial (such as continental to oceanic, clean to polluted, and nearsurface to upper troposphere) and temporal (from the hourly to the decadal) scales. At the Atmospheric Radiation Measurement Southern Great Plains observatory (ARM SGP) in Lamont, Oklahoma, United States, long-Term measurements for PM2:5 speciation (NH4, SO4, NO3, and organic carbon (OC)), NOx , O3, SO2, T, and RH, as well as [CCN0.4] are available. We modify, optimize, and retrain the developed RFRM to make predictions from 19 to 9 of these available predictors. This retrained RFRM (RFRM-ShortVars) shows a reduction in performance due to the unavailability and sparsity of measurements (predictors); it captures the [CCN0.4] variability and magnitude at SGP with 67:02% of the derived values in the good agreement range. This work shows the potential of using the more commonly available measurements of PM2:5 speciation to alleviate the sparsity of CCN number concentrations' measurements. © 2020 Copernicus GmbH. All rights reserved." "57219947521;57203656607;","A hybrid spatio-temporal prediction model for solar photovoltaic generation using numerical weather data and satellite images",2020,"10.3390/rs12223706","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096228596&doi=10.3390%2frs12223706&partnerID=40&md5=26b4d7f4a5274091fadb3739918a065c","Precise and accurate prediction of solar photovoltaic (PV) generation plays a major role in developing plans for the supply and demand of power grid systems. Most previous studies on the prediction of solar PV generation employed only weather data composed of numerical text data. The numerical text weather data can reflect temporal factors, however, they cannot consider the movement features related to the wind direction of the spatial characteristics, which include the amount of both clouds and particulate matter (PM) among other weather features. This study aims developing a hybrid spatio-temporal prediction model by combining general weather data and data extracted from satellite images having spatial characteristics. A model for hourly prediction of solar PV generation is proposed using data collected from a solar PV power plant in Incheon, South Korea. To evaluate the performance of the prediction model, we compared and performed ARIMAX analysis, which is a traditional statistical time-series analysis method, and SVR, ANN, and DNN, which are based on machine learning algorithms. The models that reflect the temporal and spatial characteristics exhibited better performance than those using only the general weather numerical data or the satellite image data. © MDPI AG. All rights reserved." "57219949429;23487156500;57219950276;56372626300;57219950011;","Combining phenological camera photos and modis reflectance data to predict gpp daily dynamics for alpine meadows on the tibetan plateau",2020,"10.3390/rs12223735","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096227333&doi=10.3390%2frs12223735&partnerID=40&md5=162222b642ee9f9a7108bee37dfb00e4","Gross primary production (GPP) is the overall photosynthetic fixation of carbon per unit space and time. Due to uncertainties resulting from clouds, snow, aerosol, and topography, it is a challenging task to accurately estimate daily GPP. Daily digital photos from a phenological camera record vegetation daily greenness dynamics with little cloud or aerosol disturbance. It can be fused with satellite remote sensing data to improve daily GPP prediction accuracy. In this study, we combine the two types of datasets to improve the estimation accuracy of GPP for alpine meadow on the Tibetan Plateau. To examine the performance of different methods and vegetation indices (VIs), three experiments were designed. First, GPP was estimated with the light use efficiency (LUE) model with the green chromatic coordinate (GCC) from the phenological camera and vegetation index from MODIS, respectively. Second, GPP was estimated with the Backpropagation neural network machine learning algorithm (BNNA) method with GCC from the phenological camera and vegetation index from MODIS, respectively. Finally, GPP was estimated with the BNNA method using GCC and vegetation index as inputs at the same time. Compared with eddy covariance GPP, GPP predicted by the BNNA method with GCC and vegetation indices as inputs at the same time showed the highest accuracy of all the experiments. The results indicated that GCC had a higher accuracy than NDVI and EVI when only one vegetation index data was used in the LUE model or the BNNA method. The R2 of GPP estimated by BNNA and GPP from eddy covariance increased by 0.12 on average, RMSE decreased by 1.13 g C·m−2·day −1 on average, and MAD decreased by 0.87 g C·m−2·day −1 on average compared with GPP estimated by the traditional LUE model and GPP from eddy covariance. This study puts forth a new way to improve the estimation accuracy of GPP on the Tibetan Plateau. With the emergence of a large number of phenological cameras, this method has great potential for use on the Tibetan Plateau, which is heavily affected by clouds and snow. © MDPI AG. All rights reserved." "57219937609;35763111200;","Object-oriented lulc classification in google earth engine combining snic, glcm, and machine learning algorithms",2020,"10.3390/rs12223776","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096173988&doi=10.3390%2frs12223776&partnerID=40&md5=e36dbd8f108e196677d7ce1e93936f40","Google Earth Engine (GEE) is a versatile cloud platform in which pixel-based (PB) and object-oriented (OO) Land Use–Land Cover (LULC) classification approaches can be implemented, thanks to the availability of the many state-of-art functions comprising various Machine Learning (ML) algorithms. OO approaches, including both object segmentation and object textural analysis, are still not common in the GEE environment, probably due to the difficulties existing in concatenating the proper functions, and in tuning the various parameters to overcome the GEE computational limits. In this context, this work is aimed at developing and testing an OO classification approach combining the Simple Non-Iterative Clustering (SNIC) algorithm to identify spatial clusters, the Gray-Level Co-occurrence Matrix (GLCM) to calculate cluster textural indices, and two ML algorithms (Random Forest (RF) or Support Vector Machine (SVM)) to perform the final classification. A Principal Components Analysis (PCA) is applied to the main seven GLCM indices to synthesize in one band the textural information used for the OO classification. The proposed approach is implemented in a user-friendly, freely available GEE code useful to perform the OO classification, tuning various parameters (e.g., choose the input bands, select the classification algorithm, test various segmentation scales) and compare it with a PB approach. The accuracy of OO and PB classifications can be assessed both visually and through two confusion matrices that can be used to calculate the relevant statistics (producer’s, user’s, overall accuracy (OA)). The proposed methodology was broadly tested in a 154 km2 study area, located in the Lake Trasimeno area (central Italy), using Landsat 8 (L8), Sentinel 2 (S2), and PlanetScope (PS) data. The area was selected considering its complex LULC mosaic mainly composed of artificial surfaces, annual and permanent crops, small lakes, and wooded areas. In the study area, the various tests produced interesting results on the different datasets (OA: PB RF (L8 = 72.7%, S2 = 82%, PS = 74.2), PB SVM (L8 = 79.1%, S2 = 80.2%, PS = 74.8%), OO RF (L8 = 64%, S2 = 89.3%, PS = 77.9), OO SVM (L8 = 70.4, S2 = 86.9%, PS = 73.9)). The broad code application demonstrated very good reliability of the whole process, even though the OO classification process resulted, sometimes, too demanding on higher resolution data, considering the available computational GEE resources. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "16485521200;7102795549;7004063850;","Exploring the Constraints on Simulated Aerosol Sources and Transport Across the North Atlantic With Island-Based Sun Photometers",2020,"10.1029/2020EA001392","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096467787&doi=10.1029%2f2020EA001392&partnerID=40&md5=95c6345f6b7f0a381c6fee9263b0aff7","Atmospheric aerosol over the North Atlantic Ocean impacts regional clouds and climate. In this work, we use a set of sun photometer observations of aerosol optical depth (AOD) located on the Graciosa and Cape Verde islands, along with the GEOS-Chem chemical transport model to investigate the sources of these aerosol and their transport over the North Atlantic Ocean. At both locations, the largest simulated contributor to aerosol extinction is the local source of sea-salt aerosol. In addition to this large source, we find that signatures consistent with long-range transport of anthropogenic, biomass burning, and dust emissions are apparent throughout the year at both locations. Model simulations suggest that this signal of long-range transport in AOD is more apparent at higher elevation locations; the influence of anthropogenic and biomass burning aerosol extinction is particularly pronounced at the height of Pico Mountain, near the Graciosa Island site. Using a machine learning approach, we further show that simulated observations at these three sites (near Graciosa, Pico Mountain, and Cape Verde) can be used to predict the simulated background aerosol imported into cities on the European mainland, particularly during the local winter months, highlighting the utility of background AOD monitoring for understanding downwind air quality. ©2020. The Authors." "56113469400;57203048222;","Modeling cloud-to-ground lightning probability in Alaskan tundra through the integration of Weather Research and Forecast (WRF) model and machine learning method",2020,"10.1088/1748-9326/abbc3b","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096331143&doi=10.1088%2f1748-9326%2fabbc3b&partnerID=40&md5=1dbb3fe7419d156e7d27ba454b1c498c","Wildland fires exert substantial impacts on tundra ecosystems of the high northern latitudes (HNL), ranging from biogeochemical impact on climate system to habitat suitability for various species. Cloud-to-ground (CG) lightning is the primary ignition source of wildfires. It is critical to understand mechanisms and factors driving lightning strikes in this cold, treeless environment to support operational modeling and forecasting of fire activity. Existing studies on lightning strikes primarily focus on Alaskan and Canadian boreal forests where land-atmospheric interactions are different and, thus, not likely to represent tundra conditions. In this study, we designed an empirical-dynamical method integrating Weather Research and Forecast (WRF) simulation and machine learning algorithm to model the probability of lightning strikes across Alaskan tundra between 2001 and 2017. We recommended using Thompson 2-moment and Mellor-Yamada-Janjic schemes as microphysics and planetary boundary layer parameterizations for WRF simulations in the tundra. Our modeling and forecasting test results have shown a strong capability to predict CG lightning probability in Alaskan tundra, with the values of area under the receiver operator characteristics curves above 0.9. We found that parcel lifted index and vertical profiles of atmospheric variables, including geopotential height, dew point temperature, relative humidity, and velocity speed, important in predicting lightning occurrence, suggesting the key role of convection in lightning formation in the tundra. Our method can be applied to data-scarce regions and support future studies of fire potential in the HNL. © 2020 The Author(s). Published by IOP Publishing Ltd." "57202739456;57203383511;57201106974;23089892300;","Discovery of new stellar groups in the Orion complex: Towards a robust unsupervised approach",2020,"10.1051/0004-6361/201935955","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85096035966&doi=10.1051%2f0004-6361%2f201935955&partnerID=40&md5=c1e257b7038de2cbd1d76053c03b0ee5","We test the ability of two unsupervised machine learning algorithms, EnLink and Shared Nearest Neighbor (SNN), to identify stellar groupings in the Orion star-forming complex as an application to the 5D astrometric data from Gaia DR2. The algorithms represent two distinct approaches to limiting user bias when selecting parameter values and evaluating the relative weights among astrometric parameters. EnLink adopts a locally adaptive distance metric and eliminates the need for parameter tuning through automation. The original SNN relies only on human input for parameter tuning so we modified SNN to run in two stages. We first ran the original SNN 7000 times, each with a randomly generated sample according to within-source co-variance matrices provided in Gaia DR2 and random parameter values within reasonable ranges. During the second stage, we modified SNN to identify the most repeating stellar groups from the 25 798 we obtained in the first stage. We recovered 22 spatially and kinematically coherent groups in the Orion complex, 12 of which were previously unknown. The groups show a wide distribution of distances extending as far as about 150 pc in front of the star-forming Orion molecular clouds, to about 50 pc beyond them, where we, unexpectedly, find several groups. Our results reveal the wealth of sub-structure in the OB association, within and beyond the classical Blaauw Orion OBI sub-groups. A full characterization of the new groups is essential as it offers the potential to unveil how star formation proceeds globally in large complexes such as Orion. © ESO 2020." "57217524312;37117655800;36473337800;15842202000;","Automatic mapping of rice growth stages using the integration of sentinel-2, mod13q1, and sentinel-1",2020,"10.3390/rs12213613","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85095771467&doi=10.3390%2frs12213613&partnerID=40&md5=9ae21c4f0087dfa2015a893c35c1dae2","Rice (Oryza sativa L.) is a staple food crop for more than half of the world’s population. Rice production is facing a myriad of problems, including water shortage, climate, and land-use change. Accurate maps of rice growth stages are critical for monitoring rice production and assessing its impacts on national and global food security. Rice growth stages are typically monitored by coarse-resolution satellite imagery. However, it is difficult to accurately map due to the occurrence of mixed pixels in fragmented and patchy rice fields, as well as cloud cover, particularly in tropical countries. To solve these problems, we developed an automated mapping workflow to produce near real-time multi-temporal maps of rice growth stages at a 10-m spatial resolution using multisource remote sensing data (Sentinel-2, MOD13Q1, and Sentinel-1). This study was investigated between 1 June and 29 September 2018 in two (wet and dry) areas of Java Island in Indonesia. First, we built prediction models based on Sentinel-2, and fusion of MOD13Q1/Sentinel-1 using the ground truth information. Second, we applied the prediction models on all images in area and time and separation between the non-rice planting class and rice planting class over the cropping pattern. Moreover, the model’s consistency on the multitemporal map with a 5–30-day lag was investigated. The result indicates that the Sentinel-2 based model classification gives a high overall accuracy of 90.6% and the fusion model MOD13Q1/Sentinel-1 shows 78.3%. The performance of multitemporal maps was consistent between time lags with an accuracy of 83.27–90.39% for Sentinel-2 and 84.15% for the integration of Sentinel-2/MOD13Q1/Sentinel-1. The results from this study show that it is possible to integrate multisource remote sensing for regular monitoring of rice phenology, thereby generating spatial information to support local-, national-, and regional-scale food security applications. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57219842847;26664885100;35145722600;15220487300;6601942051;","A survey on the usage of blockchain technology for cyber-threats in the context of industry 4.0",2020,"10.3390/su12219179","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85095688402&doi=10.3390%2fsu12219179&partnerID=40&md5=16ddaf5169707b26c0dc07ef0296f325","A systematic review of the literature is presented related to the usage of blockchain technology (BCT) for cyber-threats in the context of Industry 4.0. BCT plays a crucial role in creating smart factories and it is recognized as a core technology that triggers a disruptive revolution in Industry 4.0. Beyond security, authentication, asset tracking and the exchange of smart contracts, BCTs allow terminals to exchange information according to mutually agreed rules within a secured manner. Consequently, BCT can play a crucial role in industrial sustainability by preserving the assets and the environment and by enhancing the quality of life of citizens. In this work, a classification of the most important cyber-attacks that occurred in the last decade in Industry 4.0 is proposed based on four classes. The latter classes cover scanning, local to remote, power of root and denial of service (DoS). BCT is also defined and various types belong to BCT are introduced and highlighted. Likewise, BCT protocols and implementations are discussed as well. BCT implementation includes linear structure and directed acyclic graph (DAG) technology. Then, a comparative study of the most relevant works based on BCT in Industry 4.0 is conducted in terms of confidentiality, integrity, availability, privacy and multifactor authentication features. Our review shows that the integration of BCT in industry can ensure data confidentiality and integrity and should be enforced to preserve data availability and privacy. Future research directions towards enforcing BCT in the industrial field by considering machine learning, 5G/6G mobile systems and new emergent technologies are presented. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "56816836100;7402706393;57208431792;6701726073;","Multi-temporal predictive modelling of sorghum biomass using uav-based hyperspectral and lidar data",2020,"10.3390/rs12213587","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85094980527&doi=10.3390%2frs12213587&partnerID=40&md5=44bac44a7a31c17d2fd6e11ed7a33821","High-throughput phenotyping using high spatial, spectral, and temporal resolution remote sensing (RS) data has become a critical part of the plant breeding chain focused on reducing the time and cost of the selection process for the “best” genotypes with respect to the trait(s) of interest. In this paper, the potential of accurate and reliable sorghum biomass prediction using visible and near infrared (VNIR) and short-wave infrared (SWIR) hyperspectral data as well as light detection and ranging (LiDAR) data acquired by sensors mounted on UAV platforms is investigated. Predictive models are developed using classical regression-based machine learning methods for nine experiments conducted during the 2017 and 2018 growing seasons at the Agronomy Center for Research and Education (ACRE) at Purdue University, Indiana, USA. The impact of the regression method, data source, timing of RS and field-based biomass reference data acquisition, and the number of samples on the prediction results are investigated. R2 values for end-of-season biomass ranged from 0.64 to 0.89 for different experiments when features from all the data sources were included. Geometry-based features derived from the LiDAR point cloud to characterize plant structure and chemistry-based features extracted from hyperspectral data provided the most accurate predictions. Evaluation of the impact of the time of data acquisition during the growing season on the prediction results indicated that although the most accurate and reliable predictions of final biomass were achieved using remotely sensed data from mid-season to end-of-season, predictions in mid-season provided adequate results to differentiate between promising varieties for selection. The analysis of variance (ANOVA) of the accuracies of the predictive models showed that both the data source and regression method are important factors for a reliable prediction; however, the data source was more important with 69% significance, versus 28% significance for the regression method. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "56677549900;9233163800;6507594945;7004181239;57206924573;57209398568;57193788172;6701412834;","Characterization of the far infrared properties and radiative forcing of antarctic ice and water clouds exploiting the spectrometer-lidar synergy",2020,"10.3390/rs12213574","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85094928144&doi=10.3390%2frs12213574&partnerID=40&md5=5345b21e4cc169ea21b4b6cd2e4259e6","Optical and microphysical cloud properties are retrieved from measurements acquired in 2013 and 2014 at the Concordia base station in the Antarctic Plateau. Two sensors are used synergistically: a Fourier transform spectroradiometer named REFIR-PAD (Radiation Explorer in Far Infrared-Prototype for Applications and Developments) and a backscattering-depolarization LiDAR. First, in order to identify the cloudy scenes and assess the cloud thermodynamic phase, the REFIR-PAD spectral radiances are ingested by a machine learning algorithm called Cloud Identification and Classification (CIC). For each of the identified cloudy scenes, the nearest (in time) LiDAR backscattering profile is processed by the Polar Threshold (PT) algorithm that allows derivation of the cloud top and bottom heights. Subsequently, using the CIC and PT results as external constraints, the Simultaneous Atmospheric and Clouds Retrieval (SACR) code is applied to the REFIR-PAD spectral radiances. SACR simultaneously retrieves cloud optical depth and effective dimensions and atmospheric vertical profiles of water vapor and temperature. The analysis determines an average effective diameter of 28 µm with an optical depth of 0.76 for the ice clouds. Water clouds are only detected during the austral Summer, and the retrieved properties provide an average droplet diameter of 9 µm and average optical depth equal to four. The estimated retrieval error is about 1% for the ice crystal/droplet size and 2% for the cloud optical depth. The sensitivity of the retrieved parameters to the assumed crystal shape is also assessed. New parametrizations of the optical depth and the longwave downwelling forcing for Antarctic ice and water clouds, as a function of the ice/liquid water path, are presented. The longwave downwelling flux, computed from the top of the atmosphere to the surface, ranges between 70 and 220 W/m2 . The estimated cloud longwave forcing at the surface is (31 ± 7) W/m2 and (29 ± 6) W/m2 for ice clouds and (64 ± 12) and (62 ± 11) W/m2 for water clouds, in 2013 and 2014, respectively. The total average cloud forcing for the two years investigated is (46 ± 9) W/m2 . © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57202643803;7004870145;6507811592;7102254283;57203178537;","Weather types affect rain microstructure: Implications for estimating rain rate",2020,"10.3390/rs12213572","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85094871679&doi=10.3390%2frs12213572&partnerID=40&md5=b2e2da0dd65cab4085fb69249fbeed89","Quantitative precipitation estimation (QPE) through remote sensing has to take rain microstructure into consideration, because it influences the relationship between radar reflectivity Z and rain intensity R. For this reason, separate equations are used to estimate rain intensity of convective and stratiform rain types. Here, we investigate whether incorporating synoptic scale meteorology could yield further QPE improvements. Depending on large-scale weather types, variability in cloud condensation nuclei and the humidity content may lead to variation in rain microstructure. In a case study for Bavaria, we measured rain microstructure at ten locations with laser-based disdrometers, covering a combined 18,600 h of rain in a period of 36 months. Rain was classified on a temporal scale of one minute into convective and stratiform based on a machine learning model. Large-scale wind direction classes were on a daily scale to represent the synoptic weather types. Significant variations in rain microstructure parameters were evident not only for rain types, but also for wind direction classes. The main contrast was observed between westerly and easterly circulations, with the latter characterized by smaller average size of drops and a higher average concentration. This led to substantial variation in the parameters of the radar rain intensity retrieval equation Z–R. The effect of wind direction on Z–R parameters was more pronounced for stratiform than convective rain types. We conclude that building separate Z–R retrieval equations for regional wind direction classes should improve radar-based QPE, especially for stratiform rain events. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "56684747900;56898969100;57194760236;57204481038;57218657147;57190089926;56528895600;10139058400;55946316700;7003505161;16070064500;","Application of google earth engine cloud computing platform, sentinel imagery, and neural networks for crop mapping in Canada",2020,"10.3390/rs12213561","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85094844639&doi=10.3390%2frs12213561&partnerID=40&md5=b59a405e6dfdc6c87553ff564e2257cb","The ability of the Canadian agriculture sector to make better decisions and manage its operations more competitively in the long term is only as good as the information available to inform decision-making. At all levels of Government, a reliable flow of information between scientists, practitioners, policy-makers, and commodity groups is critical for developing and supporting agricultural policies and programs. Given the vastness and complexity of Canada’s agricultural regions, space-based remote sensing is one of the most reliable approaches to get detailed information describing the evolving state of the country’s environment. Agriculture and Agri-Food Canada (AAFC)-the Canadian federal department responsible for agriculture-produces the Annual Space-Based Crop Inventory (ACI) maps for Canada. These maps are valuable operational space-based remote sensing products which cover the agricultural land use and non-agricultural land cover found within Canada’s agricultural extent. Developing and implementing novel methods for improving these products are an ongoing priority of AAFC. Consequently, it is beneficial to implement advanced machine learning and big data processing methods along with open-access satellite imagery to effectively produce accurate ACI maps. In this study, for the first time, the Google Earth Engine (GEE) cloud computing platform was used along with an Artificial Neural Networks (ANN) algorithm and Sentinel-1,-2 images to produce an object-based ACI map for 2018. Furthermore, different limitations of the proposed method were discussed, and several suggestions were provided for future studies. The Overall Accuracy (OA) and Kappa Coefficient (KC) of the final 2018 ACI map using the proposed GEE cloud method were 77% and 0.74, respectively. Moreover, the average Producer Accuracy (PA) and User Accuracy (UA) for the 17 cropland classes were 79% and 77%, respectively. Although these levels of accuracies were slightly lower than those of the AAFC’s ACI map, this study demonstrated that the proposed cloud computing method should be investigated further because it was more efficient in terms of cost, time, computation, and automation. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57203228090;57202987792;57213283167;57219670652;57216680871;22996553100;34769664200;15728256600;","Ffau—framework for fully autonomous uavs",2020,"10.3390/rs12213533","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85094681335&doi=10.3390%2frs12213533&partnerID=40&md5=3d2268b30d8d1bd9b81451ac6711fdfc","Unmanned Aerial Vehicles (UAVs), although hardly a new technology, have recently gained a prominent role in many industries being widely used not only among enthusiastic consumers, but also in high demanding professional situations, and will have a massive societal impact over the coming years. However, the operation of UAVs is fraught with serious safety risks, such as collisions with dynamic obstacles (birds, other UAVs, or randomly thrown objects). These collision scenarios are complex to analyze in real-time, sometimes being computationally impossible to solve with existing State of the Art (SoA) algorithms, making the use of UAVs an operational hazard and therefore significantly reducing their commercial applicability in urban environments. In this work, a conceptual framework for both stand-alone and swarm (networked) UAVs is introduced, with a focus on the architectural requirements of the collision avoidance subsystem to achieve acceptable levels of safety and reliability. The SoA principles for collision avoidance against stationary objects are reviewed and a novel approach is described, using deep learning techniques to solve the computational intensive problem of real-time collision avoidance with dynamic objects. The proposed framework includes a web-interface allowing the full control of UAVs as remote clients with a supervisor cloud-based platform. The feasibility of the proposed approach was demonstrated through experimental tests using a UAV, developed from scratch using the proposed framework. Test flight results are presented for an autonomous UAV monitored from multiple countries across the world. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57219671352;55927784300;","Forecasting spatio-temporal dynamics on the land surface using earth observation data—a review",2020,"10.3390/rs12213513","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85094636124&doi=10.3390%2frs12213513&partnerID=40&md5=e22d19be8815166c62ec3896abf35ab3","Reliable forecasts on the impacts of global change on the land surface are vital to inform the actions of policy and decision makers to mitigate consequences and secure livelihoods. Geospatial Earth Observation (EO) data from remote sensing satellites has been collected continuously for 40 years and has the potential to facilitate the spatio-temporal forecasting of land surface dynamics. In this review we compiled 143 papers on EO-based forecasting of all aspects of the land surface published in 16 high-ranking remote sensing journals within the past decade. We analyzed the literature regarding research focus, the spatial scope of the study, the forecasting method applied, as well as the temporal and technical properties of the input data. We categorized the identified forecasting methods according to their temporal forecasting mechanism and the type of input data. Time-lagged regressions which are predominantly used for crop yield forecasting and approaches based on Markov Chains for future land use and land cover simulation are the most established methods. The use of external climate projections allows the forecasting of numerical land surface parameters up to one hundred years into the future, while auto-regressive time series modeling can account for intra-annual variances. Machine learning methods have been increasingly used in all categories and multivariate modeling that integrates multiple data sources appears to be more popular than univariate auto-regressive modeling despite the availability of continuously expanding time series data. Regardless of the method, reliable EO-based forecasting requires high-level remote sensing data products and the resulting computational demand appears to be the main reason that most forecasts are conducted only on a local scale. In the upcoming years, however, we expect this to change with further advances in the field of machine learning, the publication of new global datasets, and the further establishment of cloud computing for data processing. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "56237086200;55628589750;56531367400;56735478500;57203925011;","A new satellite-based retrieval of low-cloud liquid-water path using machine learning and meteosat seviri data",2020,"10.3390/rs12213475","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093977443&doi=10.3390%2frs12213475&partnerID=40&md5=aee607a40b64b9aa3936a32d688c9574","Clouds are one of the major uncertainties of the climate system. The study of cloud processes requires information on cloud physical properties, in particular liquid water path (LWP). This parameter is commonly retrieved from satellite data using look-up table approaches. However, existing LWP retrievals come with uncertainties related to assumptions inherent in physical retrievals. Here, we present a new retrieval technique for cloud LWP based on a statistical machine learning model. The approach utilizes spectral information from geostationary satellite channels of Meteosat Spinning-Enhanced Visible and Infrared Imager (SEVIRI), as well as satellite viewing geometry. As ground truth, data from CloudNet stations were used to train the model. We found that LWP predicted by the machine-learning model agrees substantially better with CloudNet observations than a current physics-based product, the Climate Monitoring Satellite Application Facility (CM SAF) CLoud property dAtAset using SEVIRI, edition 2 (CLAAS-2), highlighting the potential of such approaches for future retrieval developments. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57219573318;54782254200;57211326042;57219570795;6603095001;","High-resolution soybean yield mapping across the us midwest using subfield harvester data",2020,"10.3390/rs12213471","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093972903&doi=10.3390%2frs12213471&partnerID=40&md5=d231007c25bc9aeebf57af335fbed26f","Cloud computing and freely available, high-resolution satellite data have enabled recent progress in crop yield mapping at fine scales. However, extensive validation data at a matching resolution remain uncommon or infeasible due to data availability. This has limited the ability to evaluate different yield estimation models and improve understanding of key features useful for yield estimation in both data-rich and data-poor contexts. Here, we assess machine learning models’ capacity for soybean yield prediction using a unique ground-truth dataset of high-resolution (5 m) yield maps generated from combine harvester yield monitor data for over a million field-year observations across the Midwestern United States from 2008 to 2018. First, we compare random forest (RF) implementations, testing a range of feature engineering approaches using Sentinel-2 and Landsat spectral data for 20-and 30-m scale yield prediction. We find that Sentinel-2-based models can explain up to 45% of out-of-sample yield variability from 2017 to 2018 (r2 = 0.45), while Landsat models explain up to 43% across the longer 2008–2018 period. Using discrete Fourier transforms, or harmonic regressions, to capture soybean phenology improved the Landsat-based model considerably. Second, we compare RF models trained using this ground-truth data to models trained on available county-level statistics. We find that county-level models rely more heavily on just a few predictors, namely August weather covariates (vapor pressure deficit, rainfall, temperature) and July and August near-infrared observations. As a result, county-scale models perform relatively poorly on field-scale validation (r2 = 0.32), especially for high-yielding fields, but perform similarly to field-scale models when evaluated at the county scale (r2 = 0.82). Finally, we test whether our findings on variable importance can inform a simple, generalizable framework for regions or time periods beyond ground data availability. To do so, we test improvements to a Scalable Crop Yield Mapper (SCYM) approach that uses crop simulations to train statistical models for yield estimation. Based on findings from our RF models, we employ harmonic regressions to estimate peak vegetation index (VI) and a VI observation 30 days later, with August rainfall as the sole weather covariate in our new SCYM model. Modifications improved SCYM’s explained variance (r2 = 0.27 at the 30 m scale) and provide a new, parsimonious model. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57207771158;57218911059;57218911847;57218909762;35751750800;56108483100;57218910175;","Platform design for lifelog-based smart lighting control",2020,"10.1016/j.buildenv.2020.107267","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090725653&doi=10.1016%2fj.buildenv.2020.107267&partnerID=40&md5=afe07a24c9a15793c7fd063b1f6b6506","Lifelog refers to daily life records of individuals and includes their environmental, activity, emotional, and biometric information. The utilization of lifelogs makes it possible to provide a personalized lighting environment in line with a user's characteristics. However, no such customized lighting environment has been proposed thus far because methods for collecting and classifying the necessary data and a platform for synthesizing the data are not available. Therefore, in this study, the concept of the lifelog-based smart lighting control is introduced, and lifelog collection methods and a method to recommend the appropriate lighting environment for an individual by analyzing the collected information are proposed. Sensors, lighting controllers, and control interfaces required for lifelog-based smart lighting control were installed in a mock-up space, a machine learning server was set up, and a platform for implementing the optimal lighting environment was constructed by connecting these devices to the cloud. The platform constructed in this study creates a personalized lighting environment by utilizing lifelogs that are divided into emotional information collected via instant message (IM) text analysis, activity information collected using a location and activity tracker, and environmental information collected from weather data. © 2020 Elsevier Ltd" "57210706438;55893287300;57216237989;7004287554;","Vocalisation of the rare and flagship species Pharomachrus mocinno (Aves: Trogonidae): implications for its taxonomy, evolution and conservation",2020,"10.1080/09524622.2019.1647877","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071264664&doi=10.1080%2f09524622.2019.1647877&partnerID=40&md5=21995e7597855aa6c544e3b8fd6b8230","The Resplendent Quetzal Pharomachrus mocinno is a rare Neotropical bird included in the IUCN red list as Near Threatened. Fragmentation of its habitat, the cloud forest, is considered as the principal threat. Two subspecies are currently recognised but genetic and morphometric studies suggested they could be considered as full species. We assessed whether male vocalisation would support a species delimitation hypothesis. We recorded in the field and downloaded from sound archives vocalisation of 57 individuals from 30 different localities distributed in 11 countries. We estimated the acoustic differences of all the Pharomachrus taxa with multivariate analyses and machine learning techniques. Our results show vocal differences between P. m. mocinno and P. m. costaricensis that could have a molecular basis, potentially due to genetic drift developed during the more than three million years of separation of P. m. mocinno (from Mexico to Nicaragua) and P. m. costaricensis (Costa Rica and Panama). We therefore suggest that P. mocinno could potentially be divided into two species. A possible separation of these taxa into two species could have important consequences for the conservation status of the Resplendent Quetzals, and redirect conservation efforts for these taxa. © 2019 Informa UK Limited, trading as Taylor & Francis Group." "36718754900;7501959001;53663285500;7404403569;57218555441;57218546431;8939135600;57218546374;","Machine learning approaches for rice crop yield predictions using time-series satellite data in Taiwan",2020,"10.1080/01431161.2020.1766148","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089490090&doi=10.1080%2f01431161.2020.1766148&partnerID=40&md5=1725c43d04cf2d1b2f285b2016fe9cd1","Rice is the most important food crop in Taiwan, directly feeding more than 23 million people in the country. Information on rice production is thus crucial for crop management and food policymaking. This study aims to develop a machine learning (ML) approach for predicting rice crop yields in Taiwan using time-series Moderate Resolution Imaging Spectroradiometer (MODIS) data. We processed the data for the period from 2000 to 2018, following three main steps: (1) data pre-processing to generate smooth time-series Normalized Difference Vegetation Index (NDVI) data, (2) establishment of models for yield predictions using the heading date (HD) NDVI value, and the accumulated NDVI value of the dates from heading to maturity (DHM). The data from 2000 to 2017 were used for building predictive models using the random forests (RF) and support vector machines (SVM), leaving the 2018 data for model assessment, and (3) evaluation of model performance. The results compared with the government’s yield statistics indicated good predictions, with the root mean square error (RMSE) and mean absolute error (MAE) values between 7.1% and 11.8%, and Willmott’s index of agreement (d) values between 0.81 and 0.84 for the first crop, and 5.6% and 11.3% and d values between 0.91 and 0.95 for the second crop, respectively. A slight underestimation of yield predictions was observed for both crops, with the relative error (RE) values of −6.5% to −8.2% and −3.8% to −6% for the first and second crops, respectively. The results of regression analysis also confirmed a close agreement between these two datasets, with the correlation coefficient (r) higher than 0.84 (p-value <0.05), in both cases. Although some factors, including mixed-pixel issues, boundary effects, and cloud cover potentially affected the modelling results, our study demonstrated the effectiveness of ML methods for regional rice yield predictions from MODIS NDVI data in Taiwan. © 2020 Informa UK Limited, trading as Taylor & Francis Group." "36657489600;7101984634;7404548584;7006423243;","Leveraging spatial textures, through machine learning, to identify aerosols and distinct cloud types from multispectral observations",2020,"10.5194/amt-13-5459-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093500613&doi=10.5194%2famt-13-5459-2020&partnerID=40&md5=4d56ed1362f65276c7a11bb1f418ea57","Current cloud and aerosol identification methods for multispectral radiometers, such as the Moderate Resolution Imaging Spectroradiometer (MODIS) and Visible Infrared Imaging Radiometer Suite (VIIRS), employ multichannel spectral tests on individual pixels (i.e., fields of view). The use of the spatial information in cloud and aerosol algorithms has been primarily through statistical parameters such as nonuniformity tests of surrounding pixels with cloud classification provided by the multispectral microphysical retrievals such as phase and cloud top height. With these methodologies there is uncertainty in identifying optically thick aerosols, since aerosols and clouds have similar spectral properties in coarse-spectral-resolution measurements. Furthermore, identifying clouds regimes (e.g., stratiform, cumuliform) from just spectral measurements is difficult, since low-altitude cloud regimes have similar spectral properties. Recent advances in computer vision using deep neural networks provide a new opportunity to better leverage the coherent spatial information in multispectral imagery. Using a combination of machine learning techniques combined with a new methodology to create the necessary training data, we demonstrate improvements in the discrimination between cloud and severe aerosols and an expanded capability to classify cloud types. The labeled training dataset was created from an adapted NASA Worldview platform that provides an efficient user interface to assemble a human-labeled database of cloud and aerosol types. The convolutional neural network (CNN) labeling accuracy of aerosols and cloud types was quantified using independent Cloud-Aerosol Lidar with Orthogonal Polarization (CALIOP) and MODIS cloud and aerosol products. By harnessing CNNs with a unique labeled dataset, we demonstrate the improvement of the identification of aerosols and distinct cloud types from MODIS and VIIRS images compared to a per-pixel spectral and standard deviation thresholding method. The paper concludes with case studies that compare the CNN methodology results with the MODIS cloud and aerosol products. © Author(s) 2020. This work is distributed under the Creative Commons Attribution 4.0 License." "55479265100;","A study on the transformation of accounting based on new technologies: Evidence from korea",2020,"10.3390/su12208669","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092911020&doi=10.3390%2fsu12208669&partnerID=40&md5=7d5754fe6afc2514064f0e3e8d6f75ec","This study identifies the new accounting technologies into Cloud, Artificial Intelligence, Big Data, and Blockchain, and introduces the case of Korean companies applying new technologies to their accounting process. The purpose of this study is to help understand accounting technologies and provide examples of the adoption of these technologies in actual practice. To achieve this aim of the study, a systematic review of the literature of the major academic publications and professional reports and websites was used as a research methodology. In order to select the cases, it performed the analytical process of reviewing Korean major business and economic newspaper articles. This study provides evidence from Korea to companies contemplating the transformation of their accounting process using technology. Such companies can consider the cases presented in this study as a benchmark. It also offers guidance for the application of technologies to accounting practices for businesses and related researchers. The technology transformation is expected to be accelerated, especially after COVID-19. Therefore, it is necessary to understand and explore ways to effectively apply them. Further, while new technologies offer many opportunities, associated risks and threats should be addressed. © 2020 by the author. Licensee MDPI, Basel, Switzerland." "57203415030;57193802017;7003272359;36544749400;7006307181;6505462735;6507765711;","National scale land cover classification for ecosystem services mapping and assessment, using multitemporal copernicus EO data and google earth engine",2020,"10.3390/rs12203303","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092889651&doi=10.3390%2frs12203303&partnerID=40&md5=5863f528b289eed99bbd6826bb99b12d","Land-Use/Land-Cover (LULC) products are a common source of information and a key input for spatially explicit models of ecosystem service (ES) supply and demand. Global, continental, and regional, readily available, and free land-cover products generated through Earth Observation (EO) data, can be potentially used as relevant to ES mapping and assessment processes from regional to national scales. However, several limitations exist in these products, highlighting the need for timely land-cover extraction on demand, that could replace or complement existing products. This study focuses on the development of a classification workflow for fine-scale, object-based land cover mapping, employed on terrestrial ES mapping, within the Greek terrestrial territory. The processing was implemented in the Google Earth Engine cloud computing environment using 10 m spatial resolution Sentinel-1 and Sentinel-2 data. Furthermore, the relevance of different training data extraction strategies and temporal EO information for increasing the classification accuracy was also evaluated. The different classification schemes demonstrated differences in overall accuracy ranging from 0.88% to 4.94% with the most accurate classification scheme being the manual sampling/monthly feature classification achieving a 79.55% overall accuracy. The classification results suggest that existing LULC data must be cautiously considered for automated extraction of training samples, in the case of new supervised land cover classifications aiming also to discern complex vegetation classes. The code used in this study is available on GitHub and runs on the Google Earth Engine web platform. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57219452272;37088138800;36816017700;7005865248;","Lowland rice mapping in Sédhiou region (Senegal) using sentinel 1 and sentinel 2 data and random forest",2020,"10.3390/rs12203403","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092765171&doi=10.3390%2frs12203403&partnerID=40&md5=45f2cc442d698ab383bb50dd55796c7a","In developing countries, information on the area and spatial distribution of paddy rice fields is an essential requirement for ensuring food security and facilitating targeted actions of both technical assistance and restoration of degraded production areas. In this study, Sentinel 1 (S1) and Sentinel 2 (S2) imagery was used to map lowland rice crop areas in the Sédhiou region (Senegal) for the 2017, 2018, and 2019 growing seasons using the Random Forest (RF) algorithm. Ground sample datasets were annually collected (416, 455, and 400 samples) for training and testing yearly RF classification. A procedure was preliminarily applied to process S2 scenes and yield a normalized difference vegetation index (NDVI) time series less affected by clouds. A total of 93 predictors were calculated from S2 NDVI time series and S1 vertical transmit–horizontal receive (VH) and vertical transmit–vertical receive (VV) backscatters. Guided regularized random forest (GRRF) was used to deal with the arising multicollinearity and identify the most important predictors. The RF classifier was then applied to the selected predictors. The algorithm predicted the five land cover types present in the test areas, with a maximum accuracy of 87% and kappa coefficient of 0.8 in 2019. The broad land cover maps identified around 12,500 (2017), 13,800 (2018), and 12,800 (2019) ha of lowland rice crops. The study highlighted a partial difficulty of the classifier to distinguish rice from natural herbaceous vegetation (NHV) due to similar temporal patterns and high intra-class variability. Moreover, the results of this investigation indicated that S2-derived predictors provided more valuable information compared to VV and VH backscatter-derived predictors, but a combination of radar and optical imagery always outperformed a classification based on single-sensor inputs. An example is finally provided that illustrates how the maps obtained can be combined with ground observations through a ratio estimator in order to yield a statistically sound prediction of rice area all over the study region. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57219293912;7405490236;57219054933;","Evaluation and enhancement of unmanned aircraft system photogrammetric data quality for coastal wetlands",2020,"10.1080/15481603.2020.1819720","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092107074&doi=10.1080%2f15481603.2020.1819720&partnerID=40&md5=7f1642edf786da2c4c621aad90075446","Understanding the impacts of flight configuration and post-mission data processing techniques on unmanned aircraft system (UAS) photogrammetric data quality is essential for employing this popular technique in coastal wetland ecosystems. In this study, we systematically evaluated the effects of flight configuration (flying altitude, image overlap, and lighting conditions) on UAS photogrammetric level 1 products: orthoimagery and point clouds, and level 2 products: digital terrain models (DTM) and canopy height models (CHM). We also developed an object-based machine learning approach to correct UAS DTMs to mitigate data uncertainties caused by flight configuration and dense vegetation. Flying altitude was identified as the leading parameter in the quality of level 1 products, while image overlap was the most influential determinant for the quality of level 2 products. The correction approach effectively reduced the vertical error of DTMs for two study sites. This study informs UAS photogrammetric survey design and data enhancement for applications in coastal wetlands. © 2020 Informa UK Limited, trading as Taylor & Francis Group." "57191287722;22433611700;57198063860;","Applications of remote sensing in precision agriculture: A review",2020,"10.3390/rs12193136","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093863787&doi=10.3390%2frs12193136&partnerID=40&md5=cc2c416926af0e18f7ecb09cb46b208a","Agriculture provides for the most basic needs of humankind: food and fiber. The introduction of new farming techniques in the past century (e.g., during the Green Revolution) has helped agriculture keep pace with growing demands for food and other agricultural products. However, further increases in food demand, a growing population, and rising income levels are likely to put additional strain on natural resources. With growing recognition of the negative impacts of agriculture on the environment, new techniques and approaches should be able to meet future food demands while maintaining or reducing the environmental footprint of agriculture. Emerging technologies, such as geospatial technologies, Internet of Things (IoT), Big Data analysis, and artificial intelligence (AI), could be utilized to make informed management decisions aimed to increase crop production. Precision agriculture (PA) entails the application of a suite of such technologies to optimize agricultural inputs to increase agricultural production and reduce input losses. Use of remote sensing technologies for PA has increased rapidly during the past few decades. The unprecedented availability of high resolution (spatial, spectral and temporal) satellite images has promoted the use of remote sensing in many PA applications, including crop monitoring, irrigation management, nutrient application, disease and pest management, and yield prediction. In this paper, we provide an overview of remote sensing systems, techniques, and vegetation indices along with their recent (2015–2020) applications in PA. Remote-sensing-based PA technologies such as variable fertilizer rate application technology in Green Seeker and Crop Circle have already been incorporated in commercial agriculture. Use of unmanned aerial vehicles (UAVs) has increased tremendously during the last decade due to their cost-effectiveness and flexibility in obtaining the high-resolution (cm-scale) images needed for PA applications. At the same time, the availability of a large amount of satellite data has prompted researchers to explore advanced data storage and processing techniques such as cloud computing and machine learning. Given the complexity of image processing and the amount of technical knowledge and expertise needed, it is critical to explore and develop a simple yet reliable workflow for the real-time application of remote sensing in PA. Development of accurate yet easy to use, user-friendly systems is likely to result in broader adoption of remote sensing technologies in commercial and non-commercial PA applications. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57202286610;36697084100;35762297500;57212672124;57219552485;","Retrieving the national main commodity maps in indonesia based on high-resolution remotely sensed data using cloud computing platform",2020,"10.3390/land9100377","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093847415&doi=10.3390%2fland9100377&partnerID=40&md5=a764bd528c14fc15187f9dfa5ab0da81","Indonesia has the most favorable climates for agriculture because of its location in the tropical climatic zones. The country has several commodities to support economics growth that are driven by key export commodities—e.g., oil palm, rubber, paddy, cacao, and coffee. Thus, identifying the main commodities in Indonesia using spatially-explicit tools is essential to understand the precise productivity derived from the agricultural sectors. Many previous studies have used predictions developed using binary maps of general crop cover. Here, we present national commodity maps for Indonesia based on remote sensing data using Google Earth Engine. We evaluated a machine learning algorithm—i.e., Random Forest to parameterize how the area in commodity varied in Indonesia. We used various predictors to estimate the productivity of various commodities based on multispectral satellite imageries (36 predictors) at 30-meters spatial resolution. The national commodity map has a relatively high accuracy, with an overall accuracy of about 95% and Kappa coefficient of about 0.90. The results suggest that the oil palm plantation was the highest commodity product that occupied the largest land of Indonesia. However, this study also showed that the land area in rubber, rice paddies, and cacao commodities was underestimated due to its lack of training samples. Improvement in training data collection for each commodity should be done to increase the accuracy of the commodity maps. The commodity data can be viewed online (website can be found in the end of conclusions). This data can further provide significant information related to the agricultural sectors to investigate food provisioning, particularly in Indonesia. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57184370500;57204759617;7201908774;8434627900;","Auroral Image Classification With Deep Neural Networks",2020,"10.1029/2020JA027808","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093817691&doi=10.1029%2f2020JA027808&partnerID=40&md5=98f502eafb7bb33fb8f6ae59ebb9f065","Results from a study of automatic aurora classification using machine learning techniques are presented. The aurora is the manifestation of physical phenomena in the ionosphere-magnetosphere environment. Automatic classification of millions of auroral images from the Arctic and Antarctic is therefore an attractive tool for developing auroral statistics and for supporting scientists to study auroral images in an objective, organized, and repeatable manner. Although previous studies have presented tools for detecting aurora, there has been a lack of tools for classifying aurora into subclasses with a high precision (>90%). This work considers seven auroral subclasses: breakup, colored, arcs, discrete, patchy, edge, and faint. Six different deep neural network architectures have been tested along with the well-known classification algorithms: k-nearest neighbor (KNN) and a support vector machine (SVM). A set of clean nighttime color auroral images, without clearly ambiguous auroral forms, moonlight, twilight, clouds, and so forth, were used for training and testing the classifiers. The deep neural networks generally outperformed the KNN and SVM methods, and the ResNet-50 architecture achieved the highest performance with an average classification precision of 92%. © 2020. The Authors." "57200570867;6602610108;","Development of a machine learning-based radiometric bias correction for noaa’s microwave integrated retrieval system (Mirs)",2020,"10.3390/rs12193160","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092904232&doi=10.3390%2frs12193160&partnerID=40&md5=0f38227dfb09f6353432189f9baaefab","We present the development of a dynamic over-ocean radiometric bias correction for the Microwave Integrated Retrieval System (MiRS) which accounts for spatial, temporal, spectral, and angular dependence of the systematic differences between observed and forward model-simulated radiances. The dynamic bias correction, which utilizes a deep neural network approach, is designed to incorporate dependence on the atmospheric and surface conditions that impact forward model biases. The approach utilizes collocations of observed Suomi National Polar-orbiting Partnership/Advanced Technology Microwave Sounder (SNPP/ATMS) radiances and European Centre for Medium-Range Weather Forecasts (ECMWF) model analyses which are used as input to the Community Radiative Transfer Model (CRTM) forward model to develop training data of radiometric biases. Analysis of the neural network performance indicates that in many channels, the dynamic bias is able to reproduce realistically both the spatial patterns of the original bias and its probability distribution function. Furthermore, retrieval impact experiments on independent data show that, compared with the baseline static bias correction, using the dynamic bias correction can improve temperature and water vapor profile retrievals, particularly in regions with higher Cloud Liquid Water (CLW) amounts. Ocean surface emissivity retrievals are also improved, for example at 23.8 GHz, showing an increase in correlation from 0.59 to 0.67 and a reduction of standard deviation from 0.035 to 0.026. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "6602533891;57219363922;6506175855;","Monitoring forest change in the amazon using multi-temporal remote sensing data and machine learning classification on Google Earth Engine",2020,"10.3390/ijgi9100580","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092420606&doi=10.3390%2fijgi9100580&partnerID=40&md5=2c791a7ef23271ffc999badad2441d34","Deforestation causes diverse and profound consequences for the environment and species. Direct or indirect effects can be related to climate change, biodiversity loss, soil erosion, floods, landslides, etc. As such a significant process, timely and continuous monitoring of forest dynamics is important, to constantly follow existing policies and develop new mitigation measures. The present work had the aim of mapping and monitoring the forest change from 2000 to 2019 and of simulating the future forest development of a rainforest region located in the Pará state, Brazil. The land cover dynamics were mapped at five-year intervals based on a supervised classification model deployed on the cloud processing platform Google Earth Engine. Besides the benefits of reduced computational time, the service is coupled with a vast data catalogue providing useful access to global products, such as multispectral images of the missions Landsat five, seven, eight and Sentinel-2. The validation procedures were done through photointerpretation of highresolution panchromatic images obtained from CBERS (China-Brazil Earth Resources Satellite). The more than satisfactory results allowed an estimation of peak deforestation rates for the period 2000- 2006; for the period 2006-2015, a significant decrease and stabilization, followed by a slight increase till 2019. Based on the derived trends a forest dynamics was simulated for the period 2019-2028, estimating a decrease in the deforestation rate. These results demonstrate that such a fusion of satellite observations, machine learning, and cloud processing, benefits the analysis of the forest dynamics and can provide useful information for the development of forest policies. © 2020 by the authors." "57197738532;26649033300;57192933085;23014409400;","Anomaly detection based on machine learning in IoT-based vertical plant wall for indoor climate control",2020,"10.1016/j.buildenv.2020.107212","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089804800&doi=10.1016%2fj.buildenv.2020.107212&partnerID=40&md5=6a773d9322021c8c563e894d00da1d10","Indoor climate is closely related to human health, comfort and productivity. Vertical plant wall systems, embedded with sensors and actuators, have become a promising application for indoor climate control. In this study, we explore the possibility of applying machine learning based anomaly detection methods to vertical plant wall systems so as to enhance the automation and improve the intelligence to realize predictive maintenance of the indoor climate. Two categories of anomalies, namely point anomalies and contextual anomalies are researched. Prediction-based and pattern recognition-based methods are investigated and applied to indoor climate anomaly detection. The results show that neural network-based models, specifically the autoencoder (AE) and the long short-term memory encoder decoder (LSTM-ED) model surpass the others in terms of detecting point anomalies and contextual anomalies, respectively, therefore can be deployed into vertical plant walls systems in industrial practice. Based on the results, a new data cleaning method is proposed and a prediction-based method is deployed to the cloud in practice as a proof-of-concept. This study showcases the advancements in machine learning and Internet of things can be fully utilized by researches on building environment to accelerate the solution development. © 2020 The Authors" "15840467900;57206259084;57218294888;57218293495;57218295171;57218290463;","Evaluation and comparison of a machine learning cloud identification algorithm for the SLSTR in polar regions",2020,"10.1016/j.rse.2020.111999","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088642244&doi=10.1016%2fj.rse.2020.111999&partnerID=40&md5=7fefd97fa53d3bd6c858104b9f535b8a","A Feed Forward Neural Net (NN) approach to distinguish between clouds and the surface has been applied to the Sea and Land Surface Temperature Radiometer in polar regions. The masking algorithm covers the Arctic, Antarctic and regions typically classified as the cryosphere such as northern hemisphere permafrost. The mask has been trained using collocations with the CALIOP active lidar, which in narrow strips provide more accurate detection of cloud, and was subsequently evaluated as a function of cloud type and surface type. The mask was compared with the existing operational Bayesian and Empirical cloud masks by eye and also statistically using CALIOP data. It was found to perform exceptionally well in the polar regions. The Kuiper skill score improved from 0.28, for the operational Bayesian and 0.17 for the Empirical masks to 0.77 for the NN. The NN algorithm also has a much more homogeneous performance over all surface types. The key improvement came from better identification of clear scenes; for the NN mask, the same performance in terms of contamination of cloudy pixels in the sample of identified clear pixels can be achieved while retaining 40% of the clear pixels compared with 10% for the operational cloud identification. The algorithm performed with almost the same skill over sea and land. The best performance was achieved for opaque clouds while transparent and broken clouds showed slightly reduced accuracy. © 2020 Elsevier Inc." "57219097296;55332214100;56893485300;6701640417;","GLM and ABI Characteristics of Severe and Convective Storms",2020,"10.1029/2020JD032858","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091204819&doi=10.1029%2f2020JD032858&partnerID=40&md5=27753df453ecfa35998ba2a1eb36953a","The recent deployment of the Geostationary Lightning Mapper (GLM) on board GOES-16 and GOES-17 provides a new perspective of total lightning production for the severe convective storms research and operational communities. While the GLM has met its performance targets, further understanding flash characteristics and the physical limitations of the GLM are required to increase the applicability of the data. Derived cloud-top height and infrared (IR) brightness temperature products from the Advanced Baseline Imager (ABI) are used to assess data quality and characteristics from gridded GLM imagery across 7 weeks of active severe weather: 13 April through 31 May 2019. Areas with cloud tops colder than 240 K typically produced lightning, though this becomes less certain near the edge of the field of view due to algorithm limitations. Increasing flash rates were observed to correlate with decreasing flash areas, increasing cloud-top heights, and colder cloud-top temperatures. However, flash rates and size were more strongly tied to convective intensity and proximity to convective hazards at the surface due to the ability to delineate between convective and stratiform precipitation. Results show that merging ABI and GLM data sets could add value to both machine learning and statistical-based algorithms and also forecast applications with each providing unique details, although parameters such as GOES-16 viewing angle should be considered. Lastly, two case studies (24 and 27 May 2019) are used to help interpret the results from the 7-week sampling period and identify GLM and ABI trends related to thunderstorm evolution. ©2020. American Geophysical Union. All Rights Reserved." "57200857362;7004563395;55962154500;56583139400;8284949000;56504563100;57094306300;","DSCOVR/EPIC-derived global hourly and daily downward shortwave and photosynthetically active radiation data at 0.1° × 0.1° resolution",2020,"10.5194/essd-12-2209-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093534894&doi=10.5194%2fessd-12-2209-2020&partnerID=40&md5=993605e47eededdce75d699fa19e18ea","Downward shortwave radiation (SW) and photosynthetically active radiation (PAR) play crucial roles in Earth system dynamics. Spaceborne remote sensing techniques provide a unique means for mapping accurate spatiotemporally continuous SW-PAR, globally. However, any individual polar-orbiting or geostationary satellite cannot satisfy the desired high temporal resolution (sub-daily) and global coverage simultaneously, while integrating and fusing multisource data from complementary satellites/sensors is challenging because of coregistration, intercalibration, near real-time data delivery and the effects of discrepancies in orbital geometry. The Earth Polychromatic Imaging Camera (EPIC) on board the Deep Space Climate Observatory (DSCOVR), launched in February 2015, offers an unprecedented possibility to bridge the gap between high temporal resolution and global coverage and characterize the diurnal cycles of SW-PAR globally. In this study, we adopted a suite of well-validated data-driven machine-learning models to generate the first global land products of SW-PAR, from June 2015 to June 2019, based on DSCOVR/EPIC data. The derived products have high temporal resolution (hourly) and medium spatial resolution (0.1° × 0.1°), and they include estimates of the direct and diffuse components of SW-PAR. We used independently widely distributed ground station data from the Baseline Surface Radiation Network (BSRN), the Surface Radiation Budget Network (SURFRAD), NOAA's Global Monitoring Division and the U.S. Department of Energy's Atmospheric System Research (ASR) program to evaluate the performance of our products, and we further analyzed and compared the spatiotemporal characteristics of the derived products with the benchmarking Clouds and the Earth's Radiant Energy System Synoptic (CERES) data. We found both the hourly and daily products to be consistent with ground-based observations (e.g., hourly and daily total SWs have low biases of-3.96 and-0.71 W m-2 and root-mean-square errors (RMSEs) of 103.50 and 35.40 W m-2, respectively). The developed products capture the complex spatiotemporal patterns well and accurately track substantial diurnal, monthly, and seasonal variations in SW-PAR when compared to CERES data. They provide a reliable and valuable alternative for solar photovoltaic applications worldwide and can be used to improve our understanding of the diurnal and seasonal variabilities of the terrestrial water, carbon and energy fluxes at various spatial scales. The products are freely available at https.//doi.org/10.25584/1595069 (Hao et al., 2020). © 2020 Author(s)." "57213686348;16642370700;23394040900;","Automatic extraction of high-voltage bundle subconductors using airborne LiDAR data",2020,"10.3390/RS12183078","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092271541&doi=10.3390%2fRS12183078&partnerID=40&md5=505baf5474bd509de08223e8b38085c3","Overhead high-voltage conductors are the chief components of power lines and their safety has a strong influence on social and daily life. In the recent decade, the airborne laser scanning (ALS) technique has been widely used to capture the three-dimensional (3D) information of power lines and surrounding objects. Most of the existing methods focused on extraction of single conductors or extracted all conductors as one object class by applying machine learning techniques. Nevertheless, power line corridors (PLCs) are built with multi-loop, multi-phase structures (bundle conductors) and exist in intricate environments (e.g., mountains and forests), and thus raise challenges to process ALS data for extraction of individual conductors. This paper proposes an automated method to extract individual subconductors in bundles from complex structure of PLCs using a combined image- and point-based approach. First, the input point cloud data are grouped into 3D voxel grid and PL points and separated from pylon and tree points using the fact that pylons and trees are vertical objects while power lines are non-vertical objects. These pylons are further separated from trees by employing a statistical analysis technique and used to extract span points between two consecutive pylons; then, by using the distribution properties of power lines in each individual span, the bundles located at different height levels are extracted using image-based processing; finally, subconductors in each bundle are detected and extracted by introducing a window that slides over the individual bundle. The orthogonal plane transformation and recursive clustering procedures are exploited in each window position and a point-based processing is conducted iteratively for extraction of complete individual subconductors in each bundle. The feasibility and validity of the proposed method are verified on two Australian sites having bundle conductors in high-voltage transmission lines. Our experiments show that the proposed method achieves a reliable result by extracting the real structure of bundle conductors in power lines with correctness of 100% and 90% in the two test sites, respectively. © 2020 by the authors." "57205522425;57212792894;57219311534;57219314338;57219312305;57189580728;6603095001;","Mapping crop types in southeast india with smartphone crowdsourcing and deep learning",2020,"10.3390/RS12182957","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092200075&doi=10.3390%2fRS12182957&partnerID=40&md5=a7146aa44ed68ae6098bf875aa811231","High resolution satellite imagery and modern machine learning methods hold the potential to fill existing data gaps in where crops are grown around the world at a sub-field level. However, high resolution crop type maps have remained challenging to create in developing regions due to a lack of ground truth labels for model development. In this work, we explore the use of crowdsourced data, Sentinel-2 and DigitalGlobe imagery, and convolutional neural networks (CNNs) for crop type mapping in India. Plantix, a free app that uses image recognition to help farmers diagnose crop diseases, logged 9 million geolocated photos from 2017-2019 in India, 2 million of which are in the states of Andhra Pradesh and Telangana in India. Crop type labels based on farmer-submitted images were added by domain experts and deep CNNs. The resulting dataset of crop type at coordinates is high in volume, but also high in noise due to location inaccuracies, submissions from out-of-field, and labeling errors. We employed a number of steps to clean the dataset, which included training a CNN on very high resolution DigitalGlobe imagery to filter for points that are within a crop field. With this cleaned dataset, we extracted Sentinel time series at each point and trained another CNN to predict the crop type at each pixel. When evaluated on the highest quality subset of crowdsourced data, the CNN distinguishes rice, cotton, and ""other"" crops with 74% accuracy in a 3-way classification and outperforms a random forest trained on harmonic regression features. Furthermore, model performance remains stable when low quality points are introduced into the training set. Our results illustrate the potential of non-traditional, high-volume/high-noise datasets for crop type mapping, some improvements that neural networks can achieve over random forests, and the robustness of such methods against moderate levels of training set noise. Lastly, we caution that obstacles like the lack of good Sentinel-2 cloud mask, imperfect mobile device location accuracy, and preservation of privacy while improving data access will need to be addressed before crowdsourcing can widely and reliably be used to map crops in smallholder systems. © 2020 by the authors." "57201284843;57203927097;47461334700;57186194700;57203928635;6506182755;57211096480;16551253800;57211096034;57211351934;55949126400;","Integrated geological and geophysical mapping of a carbonatite-hosting outcrop in siilinjärvi, finland, using unmanned aerial systems",2020,"10.3390/RS12182998","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092116770&doi=10.3390%2fRS12182998&partnerID=40&md5=53718447b7bf877f11556978c04f208e","Mapping geological outcrops is a crucial part of mineral exploration, mine planning and ore extraction. With the advent of unmanned aerial systems (UASs) for rapid spatial and spectral mapping, opportunities arise in fields where traditional ground-based approaches are established and trusted, but fail to cover sufficient area or compromise personal safety. Multi-sensor UAS are a technology that change geoscientific research, but they are still not routinely used for geological mapping in exploration and mining due to lack of trust in their added value and missing expertise and guidance in the selection and combination of drones and sensors. To address these limitations and highlight the potential of using UAS in exploration settings, we present an UAS multi-sensor mapping approach based on the integration of drone-borne photography, multi- and hyperspectral imaging and magnetics. Data are processed with conventional methods as well as innovative machine learning algorithms and validated by geological field mapping, yielding a comprehensive and geologically interpretable product. As a case study, we chose the northern extension of the Siilinjärvi apatite mine in Finland, in a brownfield exploration setting with plenty of ground truth data available and a survey area that is partly covered by vegetation. We conducted rapid UAS surveys from which we created a multi-layered data set to investigate properties of the ore-bearing carbonatite-glimmerite body. Our resulting geologic map discriminates between the principal lithologic units and distinguishes ore-bearing from waste rocks. Structural orientations and lithological units are deduced based on high-resolution, hyperspectral image-enhanced point clouds. UAS-based magnetic data allow an insight into their subsurface geometry through modeling based on magnetic interpretation. We validate our results via ground survey including rock specimen sampling, geochemical and mineralogical analysis and spectroscopic point measurements. We are convinced that the presented non-invasive, data-driven mapping approach can complement traditional workflows in mineral exploration as a flexible tool. Mapping products based on UAS data increase efficiency and maximize safety of the resource extraction process, and reduce expenses and incidental wastes. © 2020 by the authors." "7401911971;57219293007;55971004500;26322732800;56011074300;57206895864;57203474131;","Mapping paddy rice fields by combining multi-temporal vegetation index and synthetic aperture radar remote sensing data using Google Earth Engine machine learning platform",2020,"10.3390/RS12182992","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092076309&doi=10.3390%2fRS12182992&partnerID=40&md5=b56a53bd4f900888e392dfd019d29174","The knowledge of the area and spatial distribution of paddy rice fields is important for water resource management. However, accurate map of paddy rice is a long-term challenge because of its spatiotemporal discontinuity and short duration. To solve this problem, this study proposed a paddy rice area extraction approach by using the combination of optical vegetation indices and synthetic aperture radar (SAR) data. This method is designed to overcome the data-missing problem due to cloud contamination and spatiotemporal discontinuities of the traditional optical remote sensing method. More specifically, the Sentinel-1A SAR and the Sentinel-2 multispectral imager (MSI) Level-2A imagery are used to identify paddy rice with a high temporal and spatial resolution. Three vegetation indices, namely normalized difference vegetation index (NDVI), enhanced vegetation index (EVI), and land surface water index (LSWI), are estimated from optical bands. Two polarization bands (VH (vertical-horizontal) and VV (vertical-vertical)) are used to overcome the cloud contamination problem. This approach was applied with the random forest machine learning algorithm on the Google Earth Engine platform for the Jianghan Plain in China as an experimental area. The results of 39 experiments uncovered the effect of different factors. The results indicated that the combination of VV and VH band showed a better performance compared with other polarization bands; the average producer's accuracy of paddy rice (PA) is 72.79%, 1.58% higher than the second one VH. Secondly, the combination of three indices also showed a better result than others, with average PA 73.82%, 1.42% higher than using NDVI alone. The classification result presented the best combination is EVI, VV, and VH polarization band. The producer's accuracy of paddy rice was 76.67%, with the overall accuracy (OA) of 66.07%, and Kappa statistics of 0.45. However, NDVI, EVI, and VH showed better performance in mapping the morphology. The results demonstrated the method developed in this study can be successfully applied to the cloud-prone area for mapping paddy rice to overcome the data missing caused by cloud and rain during the paddy growing season. © 2020 by the authors." "16318937600;57196872978;7004385955;35477147000;7006056290;","Atmospheric parameters of Cepheids from flux ratios with ATHOS: I. The temperature scale",2020,"10.1051/0004-6361/202038277","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091840760&doi=10.1051%2f0004-6361%2f202038277&partnerID=40&md5=cb6ca943073030015930a37b04025a03","Context. The effective temperature is a key parameter governing the properties of a star. For stellar chemistry, it has the strongest impact on the accuracy of the abundances derived. Since Cepheids are pulsating stars, determining their effective temperature is more complicated than in the case of nonvariable stars. Aims. We want to provide a new temperature scale for classical Cepheids, with a high precision and full control of the systematics. Methods. Using a data-driven machine learning technique employing observed spectra, and in taking great care to accurately phase single-epoch observations, we tied flux ratios to (label) temperatures derived using the infrared surface brightness method. Results. We identified 143 flux ratios, which allow us to determine the effective temperature with a precision of a few Kelvin and an accuracy better than 150 K, which is in line with the most accurate temperature measures available to date. The method does not require a normalization of the input spectra and provides homogeneous temperatures for low-And high-resolution spectra, even at the lowest signal-To-noise ratios. Due to the lack of a dataset with a sufficient sample size for Small Magellanic Cloud Cepheids, the temperature scale does not extend to Cepheids with [Fe/H] <-0.6 dex. However, it nevertheless provides an exquisite, homogeneous means of characterizing Galactic and Large Magellanic Cloud Cepheids. Conclusions. The temperature scale will be extremely useful in the context of spectroscopic surveys for Milky Way archaeology with the WEAVE and 4MOST spectrographs. It paves the way for highly accurate and precise metallicity estimates, which will allow us to assess the possible metallicity dependence of Cepheids' period-luminosity relations and, in turn, to improve our measurement of the Hubble constant H0. © ESO 2020." "57219197101;55738957800;8905764300;56119479900;","A Moist Physics Parameterization Based on Deep Learning",2020,"10.1029/2020MS002076","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091654322&doi=10.1029%2f2020MS002076&partnerID=40&md5=9f3000e915b558d172c344c9d6ec9af9","Current moist physics parameterization schemes in general circulation models (GCMs) are the main source of biases in simulated precipitation and atmospheric circulation. Recent advances in machine learning make it possible to explore data-driven approaches to developing parameterization for moist physics processes such as convection and clouds. This study aims to develop a new moist physics parameterization scheme based on deep learning. We use a residual convolutional neural network (ResNet) for this purpose. It is trained with 1-year simulation from a superparameterized GCM, SPCAM. An independent year of SPCAM simulation is used for evaluation. In the design of the neural network, referred to as ResCu, the moist static energy conservation during moist processes is considered. In addition, the past history of the atmospheric states, convection, and clouds is also considered. The predicted variables from the neural network are GCM grid-scale heating and drying rates by convection and clouds, and cloud liquid and ice water contents. Precipitation is derived from predicted moisture tendency. In the independent data test, ResCu can accurately reproduce the SPCAM simulation in both time mean and temporal variance. Comparison with other neural networks demonstrates the superior performance of ResNet architecture. ResCu is further tested in a single-column model for both continental midlatitude warm season convection and tropical monsoonal convection. In both cases, it simulates the timing and intensity of convective events well. In the prognostic test of tropical convection case, the simulated temperature and moisture biases with ResCu are smaller than those using conventional convection and cloud parameterizations. © 2020. The Authors." "57219177671;57219172532;57219019454;57219180968;57212948551;","Ultra-short-term prediction method of photovoltaic electric field power based on ground-based cloud image segmentation",2020,"10.1051/e3sconf/202018501052","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091590959&doi=10.1051%2fe3sconf%2f202018501052&partnerID=40&md5=7b95ffe5105196bd92f2cba45cc18a15","As a large number of photovoltaic power stations are built and put into operation, the total amount of photovoltaic power generation accounts for an increasing proportion of the total electricity. The inability to accurately predict solar energy output has brought great uncertainty to the grid. Therefore, predicting the future power of photovoltaic fields is of great significance. According to different time scales, predictions are divided into long-term, medium-term and ultra-short-term predictions. The main difficulty of ultra-short-term forecasting lies in the power fluctuations caused by sudden and drastic changes in environmental factors. The shading of clouds is directly related to the irradiance received on the surface of the photovoltaic panel, which has become the main factor affecting the fluctuation of photovoltaic power generation. Therefore, sky images captured by conventional cameras installed near solar panels can be used to analyze cloud characteristics and improve the accuracy of ultra-short-term predictions. This paper uses historical power information of photovoltaic power plants and cloud image data, combined with machine learning methods, to provide ultra-short-term predictions of the power generation of photovoltaic power plants. First, the random forest method is used to use historical power generation data to establish a single time series prediction model to predict ultra-short-term power generation. Compared with the continuous model, the root mean square (RMSE) error of prediction is reduced by 28.38%. Secondly, the Unet network is used to segment the cloud image, and the cloud amount information is analyzed and input into the random forest prediction model to obtain the bivariate prediction model. The experimental results prove that, based on the cloud amount information contained in the cloud chart, the bivariate prediction model has an 11.56% increase in prediction accuracy compared with the single time series prediction model, and an increase of 36.66% compared with the continuous model. © The Authors, published by EDP Sciences, 2020." "53264602200;57219054187;57203388944;25625412100;17345922900;","Different Fates of Young Star Clusters after Gas Expulsion",2020,"10.3847/2041-8213/abad28","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091169604&doi=10.3847%2f2041-8213%2fabad28&partnerID=40&md5=cccd0c26927454cc45c36d5028efec4c","We identify structures of the young star cluster NGC 2232 in the solar neighborhood (323.0 pc) and a newly discovered star cluster, LP 2439 (289.1 pc). Member candidates are identified using the Gaia DR2 sky position, parallax, and proper-motion data by an unsupervised machine-learning method, StarGO. Member contamination from the Galactic disk is further removed using the color-magnitude diagram. The four identified groups (NGC 2232, LP 2439, and two filamentary structures) of stars are coeval with an age of 25 Myr and were likely formed in the same giant molecular cloud. We correct the distance asymmetry from the parallax error with a Bayesian method. The 3D morphology shows the two spherical distributions of clusters NGC 2232 and LP 2439. Two filamentary structures are spatially and kinematically connected to NGC 2232. Both NGC 2232 and LP 2439 are expanding. The expansion is more significant in LP 2439, generating a loose spatial distribution with shallow volume number and mass density profiles. The expansion is suggested to be mainly driven by gas expulsion. With 73% of the cluster mass bound, NGC 2232 is currently experiencing a process of revirialization, However, LP 2439, with 52% of the cluster mass unbound, may fully dissolve in the near future. The different survivability traces the different dynamical states of NGC 2232 and LP 2439 prior to the onset of gas expulsion. While NGC 2232 may have been substructured and subvirial, LP 2439 may have either been virial/supervirial or experienced a much faster rate of gas removal. © 2020. The American Astronomical Society. All rights reserved." "35239293700;8630380500;57217779740;56402112700;55939843000;","Spatial mapping of short-term solar radiation prediction incorporating geostationary satellite images coupled with deep convolutional LSTM networks for South Korea",2020,"10.1088/1748-9326/ab9467","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090892326&doi=10.1088%2f1748-9326%2fab9467&partnerID=40&md5=4373093ddef43c1cc22dceb27eb1df54","A practical approach to continuously monitor and provide real-time solar energy prediction can help support reliable renewable energy supply and relevant energy security systems. In this study on the Korean Peninsula, contemporaneous solar radiation images obtained from the Communication, Ocean and Meteorological Satellite (COMS) Meteorological Imager (MI) system, were used to design a convolutional neural network and a long short-term memory network predictive model, ConvLSTM. This model was applied to predict one-hour ahead solar radiation and spatially map solar energy potential. The newly designed ConvLSTM model enabled reliable prediction of solar radiation, incorporating spatial changes in atmospheric conditions and capturing the temporal sequence-to-sequence variations that are likely to influence solar driven power supply and its overall stability. Results showed that the proposed ConvLSTM model successfully captured cloud-induced variations in ground level solar radiation when compared with reference images from a physical model. A comparison with ground pyranometer measurements indicated that the short-term prediction of global solar radiation by the proposed ConvLSTM had the highest accuracy [root mean square error (RMSE) = 83.458 W • m-2, mean bias error (MBE) = 4.466 W • m-2, coefficient of determination (R2) = 0.874] when compared with results of conventional artificial neural network (ANN) [RMSE = 94.085 W • m-2, MBE =-6.039 W • m-2, R2 = 0.821] and random forest (RF) [RMSE = 95.262 W • m-2, MBE =-11.576 W • m-2, R2 = 0.839] models. In addition, ConvLSTM better captured the temporal variations in predicted solar radiation, mainly due to cloud attenuation effects when compared with two selected ground stations. The study showed that contemporaneous satellite images over short-term or near real-time intervals can successfully support solar energy exploration in areas without continuous environmental monitoring systems, where satellite footprints are available to model and monitor solar energy management systems supporting real-life power grid systems. © 2020 The Author(s). Published by IOP Publishing Ltd." "57195330111;57194694881;57209009649;57188734162;56460209100;6507122674;","Comparing machine and deep learning methods for large 3D heritage semantic segmentation",2020,"10.3390/ijgi9090535","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090876880&doi=10.3390%2fijgi9090535&partnerID=40&md5=c3a4164be648f13fce9dacf04fefec63","In recent years semantic segmentation of 3D point clouds has been an argument that involves different fields of application. Cultural heritage scenarios have become the subject of this study mainly thanks to the development of photogrammetry and laser scanning techniques. Classification algorithms based on machine and deep learning methods allow to process huge amounts of data as 3D point clouds. In this context, the aim of this paper is to make a comparison between machine and deep learning methods for large 3D cultural heritage classification. Then, considering the best performances of both techniques, it proposes an architecture named DGCNN-Mod+3Dfeat that combines the positive aspects and advantages of these two methodologies for semantic segmentation of cultural heritage point clouds. To demonstrate the validity of our idea, several experiments from the ArCH benchmark are reported and commented. © 2020 by the authors." "57218877342;57202676994;57195343832;57218880958;6603232744;","IM2ELEVATION: Building height estimation from single-view aerial imagery",2020,"10.3390/RS12172719","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090549355&doi=10.3390%2fRS12172719&partnerID=40&md5=9abed4b44067c3aa027cad63c4a307c1","Estimation of the Digital Surface Model (DSM) and building heights from single-view aerial imagery is a challenging inherently ill-posed problem that we address in this paper by resorting to machine learning. We propose an end-to-end trainable convolutional-deconvolutional deep neural network architecture that enables learning mapping from a single aerial imagery to a DSM for analysis of urban scenes. We perform multisensor fusion of aerial optical and aerial light detection and ranging (Lidar) data to prepare the training data for our pipeline. The dataset quality is key to successful estimation performance. Typically, a substantial amount of misregistration artifacts are present due to georeferencing/projection errors, sensor calibration inaccuracies, and scene changes between acquisitions. To overcome these issues, we propose a registration procedure to improve Lidar and optical data alignment that relies on Mutual Information, followed by Hough transform-based validation step to adjust misregistered image patches. We validate our building height estimation model on a high-resolution dataset captured over central Dublin, Ireland: Lidar point cloud of 2015 and optical aerial images from 2017. These data allow us to validate the proposed registration procedure and perform 3D model reconstruction from single-view aerial imagery. We also report state-of-the-art performance of our proposed architecture on several popular DSM estimation datasets. © 2020 by the authors." "57218363740;56226977100;54684821900;57210823248;55963151700;57200700310;6603888005;7401925341;57208121325;","Gaussian processes retrieval of LAI from Sentinel-2 top-of-atmosphere radiance data",2020,"10.1016/j.isprsjprs.2020.07.004","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088914980&doi=10.1016%2fj.isprsjprs.2020.07.004&partnerID=40&md5=e4fbed46a2e8b16a80a0efbdb8428a12","Retrieval of vegetation properties from satellite and airborne optical data usually takes place after atmospheric correction, yet it is also possible to develop retrieval algorithms directly from top-of-atmosphere (TOA) radiance data. One of the key vegetation variables that can be retrieved from at-sensor TOA radiance data is leaf area index (LAI) if algorithms account for variability in atmosphere. We demonstrate the feasibility of LAI retrieval from Sentinel-2 (S2) TOA radiance data (L1C product) in a hybrid machine learning framework. To achieve this, the coupled leaf-canopy-atmosphere radiative transfer models PROSAIL-6SV were used to simulate a look-up table (LUT) of TOA radiance data and associated input variables. This LUT was then used to train the Bayesian machine learning algorithms Gaussian processes regression (GPR) and variational heteroscedastic GPR (VHGPR). PROSAIL simulations were also used to train GPR and VHGPR models for LAI retrieval from S2 images at bottom-of-atmosphere (BOA) level (L2A product) for comparison purposes. The BOA and TOA LAI products were consistently validated against a field dataset with GPR (R2 of 0.78) and with VHGPR (R2 of 0.80) and for both cases a slightly lower RMSE for the TOA LAI product (about 10% reduction). Because of delivering superior accuracies and lower uncertainties, the VHGPR models were further applied for LAI mapping using S2 acquisitions over the agricultural sites Marchfeld (Austria) and Barrax (Spain). The models led to consistent LAI maps at BOA and TOA scale. The LAI maps were also compared against LAI maps as generated by the SNAP toolbox, which is based on a neural network (NN). Maps were again consistent, however the SNAP NN model tends to overestimate over dense vegetation cover. Overall, this study demonstrated that hybrid LAI retrieval algorithms can be developed from TOA radiance data given a cloud-free sky, thus without the need of atmospheric correction. To the benefit of the community, the development of such hybrid models for the retrieval vegetation properties from BOA or TOA images has been streamlined in the freely downloadable ALG-ARTMO software framework. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57202079855;55494663500;","Volcano video data characterized and classified using computer vision and machine learning algorithms",2020,"10.1016/j.gsf.2020.01.016","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081897711&doi=10.1016%2fj.gsf.2020.01.016&partnerID=40&md5=7ad1f0ad3d3a661782aedaa7a7550f10","Video cameras are common at volcano observatories, but their utility is often limited during periods of crisis due to the large data volume from continuous acquisition and time requirements for manual analysis. For cameras to serve as effective monitoring tools, video frames must be synthesized into relevant time series signals and further analyzed to classify and characterize observable activity. In this study, we use computer vision and machine learning algorithms to identify periods of volcanic activity and quantify plume rise velocities from video observations. Data were collected at Villarrica Volcano, Chile from two visible band cameras located ~17 km from the vent that recorded at 0.1 and 30 frames per second between February and April 2015. Over these two months, Villarrica exhibited a diverse range of eruptive activity, including a paroxysmal eruption on 3 March. Prior to and after the eruption, activity included nighttime incandescence, dark and light emissions, inactivity, and periods of cloud cover. We quantify the color and spatial extent of plume emissions using a blob detection algorithm, whose outputs are fed into a trained artificial neural network that categorizes the observable activity into five classes. Activity shifts from primarily nighttime incandescence to ash emissions following the 3 March paroxysm, which likely relates to the reemergence of the buried lava lake. Time periods exhibiting plume emissions are further analyzed using a row and column projection algorithm that identifies plume onsets and calculates apparent plume horizontal and vertical rise velocities. Plume onsets are episodic, occurring with an average period of ~50 s and suggests a puffing style of degassing, which is commonly observed at Villarrica. However, the lack of clear acoustic transients in the accompanying infrasound record suggests puffing may be controlled by atmospheric effects rather than a degassing regime at the vent. Methods presented here offer a generalized toolset for volcano monitors to classify and track emission statistics at a variety of volcanoes to better monitor periods of unrest and ultimately forecast major eruptions. © 2020 China University of Geosciences (Beijing) and Peking University" "56203601500;24072139200;","Utilization of Google Earth Engine (GEE) for land cover monitoring over Klang Valley, Malaysia",2020,"10.1088/1755-1315/540/1/012003","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090156914&doi=10.1088%2f1755-1315%2f540%2f1%2f012003&partnerID=40&md5=1e957d60bb508a88a751c952da0d5bf2","Geospatial Big Data is currently received overwhelming attention and are on highlight globally and Google Earth Engine (GEE) is currently the hot pot platform to cater big data processing for Remote Sensing and GIS. Currently few or no study regarding the usage of this platform to study land use/cover changes over years in Malaysia. The objective is to evaluate the feasibility of GEE as a free cloud-based platform by performing classification of Klang Valley area from Landsat composites of three different years (1988-2003-2018) using multiple Machine Learning Algorithms (MLA). The best classification results were then imported and further processed to quantify the changes over the years using commercial software. Although, the classification results are of high accuracy but CART shows the best accuracy with 94.71%, 97.72% and 96.57% in 1988, 2003 and 2018 in comparison with RF and SVM. Some misclassified pixels were encountered because the annual composited images were compiled without taken into considerations of crops phenological stages (paddy) which resulted to the misclassified agricultural land into urban and bare land. Hence, the selection and composition of data initially had to be structured and strategized prior to processing as they can affect the classification result and further analysis. Regardless, GEE has performed quite well and fast in term of time and processing complexity of multiple datasets with minimal human interaction and intervention. Generally, GEE has proven to be reliable in fulfilling the objectives of this study to evaluate the GEE feasibility by performing classification and quantifying the land use/cover of studied area and provide good base for further analysis using different platform. © Published under licence by IOP Publishing Ltd." "57219308076;57190759179;55053263400;","3D city models for urban mining: Point cloud based semantic enrichment for spectral variation identification in hyperspectral imagery",2020,"10.5194/isprs-Annals-V-4-2020-223-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092182432&doi=10.5194%2fisprs-Annals-V-4-2020-223-2020&partnerID=40&md5=1bebddcb9b0c91887a2c89db06843300","Urban mining aims at reusing building materials enclosed in our cities. Therefore, it requires accurate information on the availability of these materials for each separate building. While recent publications have demonstrated that such information can be obtained using machine learning and data fusion techniques applied to hyperspectral imagery, challenges still persist. One of these is the so-called 'salt-And-pepper noise', i.e.The oversensitivity to the presence of several materials within one pixel (e.g. chimneys, roof windows). For the specific case of identifying roof materials, this research demonstrates the potential of 3D city models to identify and filter out such unreliable pixels beforehand. As, from a geometrical point of view, most available 3D city models are too generalized for this purpose (e.g. in CityGML Level of Detail 2), semantic enrichment using a point cloud is proposed to compensate missing details. So-called deviations are mapped onto a 3D building model by comparing it with a point cloud. Seeded region growing approach based on distance and orientation features is used for the comparison. Further, the results of a validation carried out for parts of Rotterdam and resulting in KHAT values as high as 0.7 are discussed. © Copyright:" "57195695111;23005893600;57216759668;","Supervised Classification and Its Repeatability for Point Clouds from Dense Vhr Tri-Stereo Satellite Image Matching Using Machine Learning",2020,"10.5194/isprs-annals-V-2-2020-525-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091102393&doi=10.5194%2fisprs-annals-V-2-2020-525-2020&partnerID=40&md5=21c722c8bc08a99d4e0b266991fa510b","Image matching of aerial or satellite images and Airborne Laser Scanning (ALS) are the two main techniques for the acquisition of geospatial information (3D point clouds), used for mapping and 3D modelling of large surface areas. While ALS point cloud classification is a widely investigated topic, there are fewer studies related to the image-derived point clouds, even less for point clouds derived from stereo satellite imagery. Therefore, the main focus of this contribution is a comparative analysis and evaluation of a supervised machine learning classification method that exploits the full 3D content of point clouds generated by dense image matching of tri-stereo Very High Resolution (VHR) satellite imagery. The images were collected with two different sensors (Pleíades and WorldView-3) at different timestamps for a study area covering a surface of 24 km2, located in Waldviertel, Lower Austria. In particular, we evaluate the performance and precision of the classifier by analysing the variation of the results obtained after multiple scenarios using different training and test data sets. The temporal difference of the two Pleíades acquisitions (7 days) allowed us to calculate the repeatability of the adopted machine learning algorithm for the classification. Additionally, we investigate how the different acquisition geometries (ground sample distance, viewing and convergence angles) influence the performance of classifying the satellite image-derived point clouds into five object classes: ground, trees, roads, buildings, and vehicles. Our experimental results indicate that, in overall the classifier performs very similar in all situations, with values for the F1-score between 0.63 and 0.65 and overall accuracies beyond 93%. As a measure of repeatability, stable classes such as buildings and roads show a variation below 3% for the F1-score between the two Pleíades acquisitions, proving the stability of the model. © 2020 Copernicus GmbH. All rights reserved." "57202445503;23767232700;6602209960;","Classification of Tree Species and Standing Dead Trees by Fusing Uav-Based Lidar Data and Multispectral Imagery in the 3D Deep Neural Network Pointnet++",2020,"10.5194/isprs-annals-V-2-2020-203-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091088085&doi=10.5194%2fisprs-annals-V-2-2020-203-2020&partnerID=40&md5=f57a7ce628cb5ba3acfbe2d7a609ebe0","Knowledge of tree species mapping and of dead wood in particular is fundamental to managing our forests. Although individual tree-based approaches using lidar can successfully distinguish between deciduous and coniferous trees, the classification of multiple tree species is still limited in accuracy. Moreover, the combined mapping of standing dead trees after pest infestation is becoming increasingly important. New deep learning methods outperform baseline machine learning approaches and promise a significant accuracy gain for tree mapping. In this study, we performed a classification of multiple tree species (pine, birch, alder) and standing dead trees with crowns using the 3D deep neural network (DNN) PointNet++ along with UAV-based lidar data and multispectral (MS) imagery. Aside from 3D geometry, we also integrated laser echo pulse width values and MS features into the classification process. In a preprocessing step, we generated the 3D segments of single trees using a 3D detection method. Our approach achieved an overall accuracy (OA) of 90.2% and was clearly superior to a baseline method using a random forest classifier and handcrafted features (OA Combining double low line 85.3%). All in all, we demonstrate that the performance of the 3D DNN is highly promising for the classification of multiple tree species and standing dead trees in practice. © 2020 Copernicus GmbH. All rights reserved." "26221613300;57218827452;55586607900;24536694300;","Automated updating of forest cover maps from cloud-free sentinel-2 mosaic images using object-based image analysis and machine learning methods",2020,"10.5194/isprs-Annals-V-3-2020-803-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090343402&doi=10.5194%2fisprs-Annals-V-3-2020-803-2020&partnerID=40&md5=dad63df872c4d4a0b24f30bfd4bcd860","Planning sustainable use of land resources and environmental monitoring benefit from accurate and detailed forest information. The basis of accurate forest information is data on the spatial extent of forests. In Norway land resource maps have been carefully created by field visits and aerial image interpretation for over four decades with periodic updating. However, due to prioritization of agricultural and built-up areas, and high requirements with respect to the map accuracy, forest areas and outfields have not been frequently updated. Consequently, in some part of the country, the map has not been updated since its first creation in the 1960s. The Sentinel-2 satellite acquires images with high spatial and temporal resolution which provides opportunities for creating cloud-free mosaic images over areas that are often covered with clouds. Here, we combine object-based image analysis with machine learning methods in an automated framework to map forest area in Sentinel-2 mosaic images. The images are segmented using the eCogntion™ software. Training data are collected automatically from the existing land resource map and filtered using height and greenness information so that the training samples certainly represent their respective classes. Two machine learning algorithms, namely Random Forest (RF) and the Multilayer Perceptron Neural Network (MLP), are then trained and validated before mapping forest area. The effects of including and excluding some features on the classification accuracy is investigated. The results show that the method produces forest cover map at very high accuracy (up to 97%). The MLP performs better than the RF algorithm both in classification accuracy and in robustness against inclusion and exclusion of features. © Authors 2020. All rights reserved." "57218827879;57218822575;6507146779;57218829464;57210999953;35203565800;57218822623;57192830024;57198996571;","Machine learning for classification of an eroding scarp surface using terrestrial photogrammetry with nir and rgb imagery",2020,"10.5194/isprs-Annals-V-3-2020-431-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090337515&doi=10.5194%2fisprs-Annals-V-3-2020-431-2020&partnerID=40&md5=d0043d3d0246cf9f82296bf8f7bdb9a0","Increasingly advanced and affordable close-range sensing techniques are employed by an ever-broadening range of users, with varying competence and experience. In this context a method was tested that uses photogrammetry and classification by machine learning to divide a point cloud into different surface type classes. The study site is a peat scarp 20 metres long in the actively eroding river bank of the Rotmoos valley near Obergurgl, Austria. Imagery from near-infra red (NIR) and conventional (RGB) sensors, georeferenced with coordinates of targets surveyed with a total station, was used to create a point cloud using structure from motion and dense image matching. NIR and RGB information were merged into a single point cloud and 18 geometric features were extracted using three different radii (0.02 m, 0.05 m and 0.1 m) totalling 58 variables on which to apply the machine learning classification. Segments representing six classes, dry grass, green grass, peat, rock, snow and target, were extracted from the point cloud and split into a training set and a testing set. A Random Forest machine learning model was trained using machine learning packages in the R-CRAN environment. The overall classification accuracy and Kappa Index were 98% and 97% respectively. Rock, snow and target classes had the highest producer and user accuracies. Dry and green grass had the highest omission (1.9% and 5.6% respectively) and commission errors (3.3% and 3.4% respectively). Analysis of feature importance revealed that the spectral descriptors (NIR, R, G, B) were by far the most important determinants followed by verticality at 0.1 m radius. © Authors 2020. All rights reserved." "57188763259;57196055256;","Virtual infrastructure provisioning virtual machine with machine learning prediction in green cloud computing",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091597411&partnerID=40&md5=49996417f1b5d648a644662201da215f","Cloud computing is one of the emerging technology in the world it refer different technologies, services concepts and architectures. Cloud has provided everything as a service to end users, medium and large scale enterprises across globally. The datacenters contribute major role in the cloud, it has different architecture, service provided by cloud providers. The datacenters level required lots of optimization, it will helps to gain the less energy consumption, less pollution and carbon emission, provide more powerful performance to consumer with less cost, QoS within SLA, innovation. To achieve these datacenter levels required continuously monitoring, analysis and take decisions on the log level, monitoring metrics, thresholds. The datacenters resource objects using training and prediction on usage during provisioning and server consolidation. The application workflow request and response experiment analyses. There selections of energy selection on datacenter. The agent based monitoring, agent less script based monitoring, automation, artificial intelligence, neural network. This paper mainly concentrated analysis metrics, thresholds on cloud, capacity analysis, application workflow and Resource such as CPU, RAM utilization and prediction using Machine Learning techniques such as LR, RantomTree, RandomForest and 10 folds cross validation in cloud. The data center power source, architecture selection and usage of resource in usage optimum will helps us to reduce carbon emission and green cloud computing. © 2020 Alpha Publishers. All rights reserved." "55764719600;57217651295;55415880000;6506685277;","Optimized energy cost and carbon emission-aware virtual machine allocation in sustainable data centers",2020,"10.3390/SU12166383","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090083510&doi=10.3390%2fSU12166383&partnerID=40&md5=f66ddaee4b3bf7b872c8e80cd15eb3d1","Cloud data center's total operating cost is conquered by electricity cost and carbon tax incurred due to energy consumption from the grid and its associated carbon emission. In this work, we consider geo-distributed sustainable datacenter's with varying on-site green energy generation, electricity prices, carbon intensity and carbon tax. The objective function is devised to reduce the operating cost including electricity cost and carbon cost incurred on the power consumption of servers and cooling devices. We propose renewable-aware algorithms to schedule the workload to the data centers with an aim to maximize the green energy usage. Due to the uncertainty and time variant nature of renewable energy availability, an investigation is performed to identify the impact of carbon footprint, carbon tax and electricity cost in data center selection on total operating cost reduction. In addition, on-demand dynamic optimal frequency-based load distribution within the cluster nodes is performed to eliminate hot spots due to high processor utilization. The work suggests optimal virtual machine placement decision to maximize green energy usage with reduced operating cost and carbon emission. © 2020 by the authors." "57190034210;57194694881;55444846100;54787529900;6507122674;","A hierarchical machine learning approach for multi-level and multi-resolution 3d point cloud classification",2020,"10.3390/RS12162598","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090017553&doi=10.3390%2fRS12162598&partnerID=40&md5=1d252318efd2db65b905fc1cde7c5ed9","The recent years saw an extensive use of 3D point cloud data for heritage documentation, valorisation and visualisation. Although rich in metric quality, these 3D data lack structured information such as semantics and hierarchy between parts. In this context, the introduction of point cloud classification methods can play an essential role for better data usage, model definition, analysis and conservation. The paper aims to extend a machine learning (ML) classification method with a multi-level and multi-resolution (MLMR) approach. The proposed MLMR approach improves the learning process and optimises 3D classification results through a hierarchical concept. The MLMR procedure is tested and evaluated on two large-scale and complex datasets: the Pomposa Abbey (Italy) and the Milan Cathedral (Italy). Classification results show the reliability and replicability of the developed method, allowing the identification of the necessary architectural classes at each geometric resolution. © 2020 by the authors." "57211519015;8600097300;57213346004;57192512150;57205863098;","A novel classification extension-based cloud detection method for medium-resolution optical images",2020,"10.3390/RS12152365","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089689487&doi=10.3390%2fRS12152365&partnerID=40&md5=147cee8296c6f2bf6a93c639729a2d6d","Accurate cloud detection using medium-resolution multispectral satellite imagery (such as Landsat and Sentinel data) is always difficult due to the complex land surfaces, diverse cloud types, and limited number of available spectral bands, especially in the case of images without thermal bands. In this paper, a novel classification extension-based cloud detection (CECD) method was proposed for masking clouds in the medium-resolution images. The new method does not rely on thermal bands and can be used for masking clouds in different types of medium-resolution satellite imagery. First, with the support of low-resolution satellite imagery with short revisit periods, cloud and non-cloud pixels were identified in the resampled low-resolution version of the medium-resolution cloudy image. Then, based on the identified cloud and non-cloud pixels and the resampled cloudy image, training samples were automatically collected to develop a random forest (RF) classifier. Finally, the developed RF classifier was extended to the corresponding medium-resolution cloudy image to generate an accurate cloud mask. The CECD method was applied to Landsat-8 and Sentinel-2 imagery to test the performance for different satellite images, and the well-known function of mask (FMASK) method was employed for comparison with our method. The results indicate that CECD is more accurate at detecting clouds in Landsat-8 and Sentinel-2 imagery, giving an average F-measure value of 97.65% and 97.11% for Landsat-8 and Sentinel-2 imagery, respectively, as against corresponding results of 90.80% and 88.47% for FMASK. It is concluded, therefore, that the proposed CECD algorithm is an effective cloud-classification algorithm that can be applied to the medium-resolution optical satellite imagery. © 2020 by the authors." "57218454517;24822286200;","Lightning Distance Estimation Using LF Lightning Radio Signals via Analytical and Machine-Learned Models",2020,"10.1109/TGRS.2020.2972153","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089239621&doi=10.1109%2fTGRS.2020.2972153&partnerID=40&md5=9f5269d976911f2ce5480140d766b9f5","Lightning geolocation is useful in a variety of applications, ranging from weather nowcasting to a better understanding of thunderstorm evolution processes and remote sensing of the ionosphere. Lightning-generated radio signals can be used in range estimation of lightning return strokes, for which the most commonly employed technique is the time difference of arrival in lightning detection networks. Though these instrument networks provide the most reliability and best accuracy, users without access to them can instead benefit from lightning geolocation using a standalone instrument. In this article, we present the framework for training fast models capable of estimating negative cloud-to-ground lightning location from single-instrument observations of very low frequency/low frequency (VLF/LF, 3-300 kHz) radio pulses or 'sferics,' without knowledge of the ionosphere's D-region state. The models are generated using an analytical method, based on the delay between ground and skywave, and a machine learning method. The training framework is applied to three different data sets to assess model accuracy. Validation of the machine-learned models for these data sets, which include both simulated and observed sferics, confirms this technique as a promising solution for lightning distance estimation using a single receiver. Distance estimates using a machine-learned model for observed sferics in Kansas yield an RMSE of 53 km with 68% of them being within 9.8 km. Estimates using the analytical method are found to have an RMSE of 54 km with 68% of them being within 32 km. Limitations of our methodology and potential improvements to be investigated are also discussed. © 1980-2012 IEEE." "57218474008;15023850900;55630272400;57217680171;57217687864;57217675347;22234853700;","Thin cloud removal in optical remote sensing images based on generative adversarial networks and physical model of cloud distortion",2020,"10.1016/j.isprsjprs.2020.06.021","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087526039&doi=10.1016%2fj.isprsjprs.2020.06.021&partnerID=40&md5=94e3dce17405ccf1cf7703204fe5079b","Cloud contamination is an inevitable problem in optical remote sensing images. Unlike thick clouds, thin clouds do not completely block out background which makes it possible to restore background information. In this paper, we propose a semi-supervised method based on generative adversarial networks (GANs) and a physical model of cloud distortion (CR-GAN-PM) for thin cloud removal with unpaired images from different regions. A physical model of cloud distortion which takes the absorption of cloud into consideration was also defined in this paper. It is worth noting that many state-of-the-art methods based on deep learning require paired cloud and cloud-free images from the same region, which is often unavailable or time-consuming to collect. CR-GAN-PM has two main steps: first, the cloud-free background and cloud distortion layers were decomposed from an input cloudy image based on GANs and the principles of image decomposition; then, the input cloudy image was reconstructed by putting those layers into the redefined physical model of cloud distortion. The decomposition process ensured that the decomposed background layer was cloud-free and the reconstruction process ensured that generated background layer was correlated with the input cloudy image. Experiments were conducted on Sentinel-2A imagery to validate the proposed CR-GAN-PM. Averaged over all testing images, the SSIMs values (structural similarity index measurement) of CR-GAN-PM were 0.72, 0.77, 0.81 and 0.83 for visible and NIR bands respectively. Those results were similar to the end-to-end deep learning-based methods and better than traditional methods. The number of input bands and values of hyper-parameters affected little on the performance of CR-GAN-PM. Experimental results show that CR-GAN-PM is effective and robust for thin cloud removal in different bands. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57194328647;55807752200;7004635295;","Machine learning techniques for regional scale estimation of high-resolution cloud-free daily sea surface temperatures from MODIS data",2020,"10.1016/j.isprsjprs.2020.06.008","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086632052&doi=10.1016%2fj.isprsjprs.2020.06.008&partnerID=40&md5=4cebe15a5dbe418a6b996de59fe95052","High-resolution sea surface temperature (SST) estimates are dependent on satellite-based infrared radiometers, which are proven to be highly accurate in the past decades. However, the presence of clouds is a big stumbling block when physical approaches are used to derive SST. This problem is more prominent across tropical regions such as Arabian Sea(AS) and Bay of Bengal(BoB), restricting the availability of high-resolution SST data for ocean applications. The previous studies for developing daily high-resolution cloud-free SST products mainly focus on fusion of multiple satellites and in-situ data products that are computationally expensive and often time consuming. At the same time, it was observed that the capabilities of data-driven approaches are not yet fully explored in the estimation of cloud-free high-resolution SST data. Hence, in this study an attempt has been made for the first time to estimate daily cloud free SST from a single sensor (MODIS Aqua) dataset using advanced machine learning techniques. Here, three distinct machine learning techniques such as Artificial Neural Networks (ANN), Support Vector Regression (SVR) and Random Forest (RF)-based algorithms were developed and evaluated over two different study areas within the AS and BoB using 10 years of MODIS data and in-situ reference data. Among the developed algorithms, the SVR-based algorithm performs consistently better. In AS region, while testing, the SVR-based SST estimates was able to achieve an adjusted coefficient of determination (Radj2) of 0.82 and root mean square error (RMSE) of 0.71 °C with respect to the in situ data. Similarly, in BoB too, the SVR algorithm outperforms the other algorithms with Radj2 of 0.78 with RMSE of 0.88 °C. Further, a spatio-temporal and visual analysis of the results as well as an inter-comparision with NOAA AVHRR daily optimally interpolated global SST (a standard SST product available in practice) the suggest that the proposed SVR-based algorithm has huge potential to produce operational high-resolution cloud-free SST estimates, even if there is cloud cover in the image. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "57218292971;7409462943;57190281074;57215817129;57196043593;57196040405;57218393669;56809128600;37041512600;","Upward Expansion of Supra-Glacial Debris Cover in the Hunza Valley, Karakoram, During 1990 ∼ 2019",2020,"10.3389/feart.2020.00308","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089099686&doi=10.3389%2ffeart.2020.00308&partnerID=40&md5=4c2f8c9b7afcb4de435632ddc3fd0e5c","Supra-glacial debris cover is key to glacier ablation through increasing (thin debris layer) or decreasing (thick debris layer) melt rates, thereby regulating the mass balance of a glacier and its meltwater runoff. The thickening or lateral expansion of supra-glacial debris cover correlates with a reduction of glacier ablation and, consequently, runoff generation, which is also considered to be an influential factor on the rheology and dynamics of a glacierized system. Studies on supra-glacial debris cover have recently attracted wide attention especially for glaciers in the Himalayas and Karakoram, where the glaciers have heterogeneously responded to climate change. In this study, we used 32 images from the Landsat Thematic Mapper, Enhanced Thematic Mapper Plus, and Operational Land Imager archive, going back to 1990, which are available on the Google Earth Engine cloud-computing platform, to map the supra-glacial debris cover in the Hunza Valley, Karakoram, Pakistan, based on a band ratio segmentation method (normalized difference snow index [NDSI] < 0.4), Otsu thresholding, and machine learning algorithms. Compared with manual digitization, the random forest (RF) model was found to have the greatest accuracy in identifying supra-glacial debris, with a Kappa coefficient of 0.94 ± 0.01 and an overall accuracy of 95.5 ± 0.9%. Overall, the supra-glacial debris cover in the study area showed an increasing trend, and the total area expanded by 8.1–21.3% for various glaciers from 1990 to 2019. The other two methods (Otsu thresholding and NDSI < 0.4) generally overestimated the supra-glacial debris covered area, by 36.3 and 18.8%, respectively, compared to that of the RF model. The supra-glacial debris cover has migrated upward on the glaciers, with intensive variation near the equilibrium-line altitude zone (4,500–5,500 m a.s.l.). The increase in ice or snow avalanche activity at high altitudes may be responsible for this upward expansion of supra-glacial debris cover in the Hunza Valley, which is attributed to the combined effect of temperature decrease and precipitation increase in the study area. © Copyright © 2020 Xie, Liu, Wu, Zhu, Gao, Qi, Duan, Saifullah and Tahir." "57213598861;36921601500;7003341789;55636444100;55479763800;8791306500;7003817690;15770244400;","Exploration of machine learning methods for the classification of infrared limb spectra of polar stratospheric clouds",2020,"10.5194/amt-13-3661-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088032027&doi=10.5194%2famt-13-3661-2020&partnerID=40&md5=48014d96d020cc98e0b2b7432f66887b","Polar stratospheric clouds (PSCs) play a key role in polar ozone depletion in the stratosphere. Improved observations and continuous monitoring of PSCs can help to validate and improve chemistry-climate models that are used to predict the evolution of the polar ozone hole. In this paper, we explore the potential of applying machine learning (ML) methods to classify PSC observations of infrared limb sounders. Two datasets were considered in this study. The first dataset is a collection of infrared spectra captured in Northern Hemisphere winter 2006/2007 and Southern Hemisphere winter 2009 by the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS) instrument on board the European Space Agency's (ESA) Envisat satellite. The second dataset is the cloud scenario database (CSDB) of simulated MIPAS spectra. We first performed an initial analysis to assess the basic characteristics of the CSDB and to decide which features to extract from it. Here, we focused on an approach using brightness temperature differences (BTDs). From both the measured and the simulated infrared spectra, more than 10 000 BTD features were generated. Next, we assessed the use of ML methods for the reduction of the dimensionality of this large feature space using principal component analysis (PCA) and kernel principal component analysis (KPCA) followed by a classification with the support vector machine (SVM). The random forest (RF) technique, which embeds the feature selection step, has also been used as a classifier. All methods were found to be suitable to retrieve information on the composition of PSCs. Of these, RF seems to be the most promising method, being less prone to overfitting and producing results that agree well with established results based on conventional classification methods. © Author(s) 2020. This work is distributed under the Creative Commons Attribution 4.0 License." "55266523600;57188869324;57215429741;57215414947;6603393238;35762297500;6508239454;","A preliminary study on machine learning and google earth engine for mangrove mapping",2020,"10.1088/1755-1315/500/1/012038","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087755156&doi=10.1088%2f1755-1315%2f500%2f1%2f012038&partnerID=40&md5=2c43b7e9b8788ad2f2a69f47d87ae9d8","The alarming rate of global mangrove forest degradation corroborates the need for providing fast, up-to-date and accurate mangrove maps. Conventional scene by scene image classification approach is inefficient and time consuming. The development of Google Earth Engine (GEE) provides a cloud platform to access and seamlessly process large amount of freely available satellite imagery. The GEE also provides a set of the state-of-the-art classifiers for pixel-based classification that can be used for mangrove mapping. This study is an initial effort which is aimed to combine machine learning and GEE for mapping mangrove extent. We used two Landsat 8 scenes over Agats and Timika Papua area as pilot images for this study; path 102 row 64 (2014/10/19) and path 103 row 63 (2013/05/16). The first image was used to develop local training areas for the machine learning classification, while the second one was used as a test image for GEE on the cloud. A total of 838 points samples were collected representing mangroves (244), non-mangroves (161), water bodies (311), and cloud (122) class. These training areas were used by support vector machine classifier in GEE to classify the first image. The classification result show mangrove objects could be efficiently delineated by this algorithm as confirmed by visual checking. This algorithm was then applied to the second image in GEE to check the consistency of the result. A simultaneous view of both classified images shows a corresponding pattern of mangrove forest, which mean the mangrove object has been consistently delineated by the algorithm. © 2020 IOP Publishing Ltd. All rights reserved." "57209638348;57217094757;57208372273;57188746436;","Snow and cloud detection using a convolutional neural network and low-resolution data from the Electro-L No. 2 Satellite",2020,"10.1117/1.JRS.14.034506","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092645221&doi=10.1117%2f1.JRS.14.034506&partnerID=40&md5=6c3b3d056a9a651aeb99b4efec2b2893","We describe an algorithm based on a convolutional neural network that detects cloud formations and snow cover in satellite images using textures. Herein, multispectral satellite images, received from a multizone scanning instrument used for hydrometeorological support and installed on the Russian satellite Electro-L No. 2, are used as input data. The problem of snow and cloud classification in the absence of a spectral channel in the range of 1.4 to 1.8 μm, which is necessary for their accurate separation, is considered. The developed algorithm can produce cloud and snow cover masks for an area limited by the values of the solar zenith angle in the range of 0 deg to 80 deg for daytime. Algorithm accuracy was evaluated using machine learning metrics and comparing its results with ground truth masks segmented manually by an experienced interpreter. In addition, we compared the resulting masks with a similar cloud mask product from the European Organisation for the Exploitation of Meteorological Satellites based on the data of the Spinning Enhanced Visible and Infrared Imager (SEVIRI) instrument installed on the Meteosat-8 satellite. According to the results of this comparison, we conclude that the cloud masks produced by the proposed convolutional neural network-based algorithm have a lower probability of false detection than products based on the SEVIRI data. The proposed algorithm is fully automatic, and it works in any season of the year during the daytime. © 2020 Society of Photo-Optical Instrumentation Engineers (SPIE)." "57211970265;57211855484;57210936702;57218799260;","A novel intelligent diagnosis and disease prediction algorithm in green cloud using machine learning approach",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090686536&partnerID=40&md5=e8b122ffbd1395c2d96bc43218ac29ba","Public health threats and epidemics are affecting the human life day-to-day. These include obesity, diabetes, cardiovascular diseases, cancer, osteoporosis and dental diseases. The trouble of chronic diseases may be cured during initial stages if it is properly predicted and it requires a lot of training using the medical data. Existing models can support disease diagnosis to certain extent based on the training. This can predict unique diseases and separate system is required for different diagnosis. A generic disease prediction model can reduce the burden of physician while making clinical decisions and has not been evolved yet. Machine learning and artificial intelligence offer one such a generic model with principled approach for intelligent disease diagnosis. A novel hybrid algorithm and a prediction model are proposed in this paper based on disease symptoms of the patient Collected from hospital and are stored in cloud.The features of KNN and CNN are combined to provide high speed prediction analysis. The enormous amount of data growth in the field of medical diagnosis helps the proposed system to find hidden patterns with respect to individual diseases. © 2020 Alpha Publishers. All rights reserved." "57201727508;55740373500;35323728000;","Combining forecasts of day-ahead solar power",2020,"10.1016/j.energy.2020.117743","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089285975&doi=10.1016%2fj.energy.2020.117743&partnerID=40&md5=e7edef1d4b83edb6e19fa50afdc5bcc0","Solar power forecasting is important for the reliable and economic operation of power systems with high penetration of solar energy. The solar power forecasts for the day-ahead time horizon are more erroneous than the hour-ahead time horizon. Numerical weather prediction (NWP) variables such as irradiance, cloud cover, precipitation etc. are used as input to day-ahead forecasting models. The uncertainty in NWP varies with weather conditions. Different forecasting algorithms based on a single method are available in the literature. Combination of individual forecasting algorithms increases the accuracy of the forecasts. However, the combined-forecast has yet not been analysed much in the area of day-ahead solar power forecasting. This paper thus explores different combined-forecast methods such as mean, median, linear regression and non-linear regressions using supervised machine learning algorithms. The number of models required for day-ahead solar power forecasts is studied. One for all hour (same) or separate models for each hour of the day are possible. The effects of retraining frequency on the performance of the forecasting models, which is important for the computational burden of the system, are also studied. Forecasting algorithms are applied to three solar plants in Australia. © 2020 Elsevier Ltd" "57216154309;35332559300;57217729762;21933438300;","Hydrometeor identification using multiple-frequency microwave links: A numerical simulation",2020,"10.3390/rs12132158","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087547745&doi=10.3390%2frs12132158&partnerID=40&md5=e88a1272b2a0ca67620ab317fefbdcd0","A method for identifying hydrometeor types (rain, graupel, and wet snow) based on a microwave link is proposed in this paper. The measured hydrometeor size distribution (HSD) data from the winters of 2014 to 2019 in Nanjing, China, were used to carry out simulation experiments to verify the performance of the model. Single-, dual-, and tri-frequency models (combinations of 15 GHz, 18 GHz, 25 GHz, 38 GHz, 50 GHz, 60 GHz, 70 GHz, and 80 GHz) were established with the extreme learning machine (ELM) algorithm. The results showed that the performance of the tri-frequency models was overall better than that of the dual-frequency models, for which the performance was better than that of the single-frequency models. The mean (maximum) test set accuracies of the single-frequency, dual-frequency, and tri-frequency models reached 75.8%, 80.7%, and 83.2% (83.0%, 84.4%, and 85.6%), respectively. For the dual-frequency and tri-frequency models, it was found that the accuracy increased with the overall frequency or the frequency difference. In addition, the influences of different noise levels on the model performance were also analyzed. Finally, the effects of position and length of link relative to precipitation cell were analyzed and are also discussed. © 2020 by the authors." "57193788172;57209398568;57206924573;6506051565;7004171611;36622868000;55914904100;55466977400;7102689523;57217728673;8525147900;","Cirrus cloud identification from airborne far-infrared and mid-infrared spectra",2020,"10.3390/rs12132097","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087543275&doi=10.3390%2frs12132097&partnerID=40&md5=676968cfa40c35c7e83f55bda62a3c94","Airborne interferometric data, obtained from the Cirrus Coupled Cloud-Radiation Experiment (CIRCCREX) and from the PiknMix-F field campaign, are used to test the ability of a machine learning cloud identification and classification algorithm (CIC). Data comprise a set of spectral radiances measured by the Tropospheric Airborne Fourier Transform Spectrometer (TAFTS) and the Airborne Research Interferometer Evaluation System (ARIES). Co-located measurements of the two sensors allow observations of the upwelling radiance for clear and cloudy conditions across the far-and mid-infrared part of the spectrum. Theoretical sensitivity studies show that the performance of the CIC algorithm improves with cloud altitude. These tests also suggest that, for conditions encompassing those sampled by the flight campaigns, the additional information contained within the far-infrared improves the algorithm's performance compared to using mid-infrared data only. When the CIC is applied to the airborne radiance measurements, the classification performance of the algorithm is very high. However, in this case, the limited temporal and spatial variability in the measured spectra results in a less obvious advantage being apparent when using both mid-and far-infrared radiances compared to using mid-infrared information only. These results suggest that the CIC algorithm will be a useful addition to existing cloud classification tools but that further analyses of nadir radiance observations spanning the infrared and sampling a wider range of atmospheric and cloud conditions are required to fully probe its capabilities. This will be realised with the launch of the Far-infrared Outgoing Radiation Understanding and Monitoring (FORUM) mission, ESA's 9th Earth Explorer. © 2020 by the authors. Licensee MDPI, Basel, Switzerland." "57204396783;57190372479;57188719498;57216602841;12785706100;57214208469;","GRNet: Geometric relation network for 3D object detection from point clouds",2020,"10.1016/j.isprsjprs.2020.05.008","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085217744&doi=10.1016%2fj.isprsjprs.2020.05.008&partnerID=40&md5=cbfa85069ee2beaa1de323270b6b1b68","Rapid detection of 3D objects in indoor environments is essential for indoor mapping and modeling, robotic perception and localization, and building reconstruction. 3D point clouds acquired by a low-cost RGB-D camera have become one of the most commonly used data sources for 3D indoor mapping. However, due to the sparse surface, empty object center, and various scales of point cloud objects, 3D bounding boxes are challenging to be estimated and located accurately. To address this, geometric shape, topological structure, and object relation are commonly employed to extract box reasoning information. In this paper, we describe the geometric feature among object points as an intra-object feature and the relation feature between different objects as an inter-object feature. Based on these two features, we propose an end-to-end point cloud geometric relation network focusing on 3D object detection, which is termed as geometric relation network (GRNet). GRNet first extracts intra-object and inter-object features for each representative point using our proposed backbone network. Then, a centralization module with a scalable loss function is proposed to centralize each representative object point to its center. Next, proposal points are sampled from these shifted points, following a proposal feature pooling operation. Finally, an object-relation learning module is applied to predict bounding box parameters. Such parameters are the additive sum of prediction results from the relation-based inter-object feature and the aggregated intra-object feature. Our model achieves state-of-the-art 3D detection results with 59.1% mAP@0.25 and 39.1% mAP@0.5 on ScanNetV2 dataset, 58.4% mAP@0.25 and 34.9% mAP@0.5 on SUN RGB-D dataset. © 2020" "57197867114;24450860900;","Linking large-scale circulation patterns to low-cloud properties",2020,"10.5194/acp-20-7125-2020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086723013&doi=10.5194%2facp-20-7125-2020&partnerID=40&md5=fb995243ca9423be7244438cf5e1f936","The North Pacific High (NPH) is a fundamental meteorological feature present during the boreal warm season. Marine boundary layer (MBL) clouds, which are persistent in this oceanic region, are influenced directly by the NPH. In this study, we combine 11 years of reanalysis and an unsupervised machine learning technique to examine the gamut of 850 hPa synoptic-scale circulation patterns. This approach reveals two distinguishable regimes-a dominant NPH setup and a land-falling cyclone-and in between a spectrum of large-scale patterns. We then use satellite retrievals to elucidate for the first time the explicit dependence of MBL cloud properties (namely cloud droplet number concentration, liquid water path, and shortwave cloud radiative effect-CRESW) on 850 hPa circulation patterns over the northeast Pacific Ocean. We find that CRESW spans from-146.8 to-115.5Wm-2, indicating that the range of observed MBL cloud properties must be accounted for in global and regional climate models. Our results demonstrate the value of combining reanalysis and satellite retrievals to help clarify the relationship between synoptic-scale dynamics and cloud physics. © 2020 Author(s)." "14035135900;57216397984;57219482763;","A digital twin model of a pasteurization system for food beverages: Tools and architecture",2020,"10.1109/ICE/ITMC49519.2020.9198625","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093120907&doi=10.1109%2fICE%2fITMC49519.2020.9198625&partnerID=40&md5=ea9fe5677b66bb4e2467f46e8abe2278","Many enabling technologies of Industry 4.0 (Internet of Things 'IoT', Cloud systems, Big Data Analytics) contribute to the creation of what is the Digital Twin or virtual twin of a physical process, that is a mathematical model capable of describing the process, product or service in a precise way in order to carry out analyses and apply strategies. Digital Twin models integrate artificial intelligence, machine learning and analytics software with the data collected from the production plants to create digital simulation models that update when the parameters of the production processes or the working conditions change. This is a self-learning mechanism, which makes use of data collected from various sources (sensors that transmit operating conditions; experts, such as engineers with deep knowledge of the industrial domain; other similar machines or fleets of similar machines) and integrates also historical data relating to the past use of the machine. Starting from the virtual twin vision, simulation plays a key role within the Industry 4.0 transformation. Creating a virtual prototype has become necessary and strategic to raise the safety levels of the operators engaged in the maintenance phases, but above all the integration of the digital model with the IoT has become particularly effective, as the advent of software platforms offers the possibility of integrating real-time data with all the digital information that a company owns on a given process, ensuring the realization of the Digital Twin. In this context, this work aims at developing optimized solutions for application in a beverage pasteurization system using the Digital Twin approach, capable of creating a virtual modelling of the process and preventing high-risk events for operators. © 2020 IEEE." "57213701812;14027120600;","Green IoT: Advancements and Sustainability with Environment by 2050",2020,"10.1109/ICRITO48877.2020.9197796","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093115224&doi=10.1109%2fICRITO48877.2020.9197796&partnerID=40&md5=25f144593f3fd246251cf0ccc3e1c9fc","Green IoT is eco-friendly technology evolves from IoT, in today's scenario each and every device connected over internet and store on cloud and these devices are known as 'Smart Devices'. At present 31 billion devices are connected as IoT devices and by 2050 it surge pass 170 billion limit so on average Iot devices are increases 12% annually as per the reports in proportionally the carbon footprint percentage and GHG(green-house gas) emission percentage increases and enhance the overall pollution percentage on earth, so to look after this major environmental issues and for proper initiatives against these issues technically as well as environmental prospective first objective of this paper is analyze the carbon footprints of smart devices with IoT for sustainability with environment and provide major steps to minimize it and emphasize the use of 'Green IoT' instead of IoT for environment safety perspectives and second objective is improve the LCA assessment model with Deep Learning and Data Mining techniques with various impact factors for better and efficient results.The proposed method cut down the production and manufacturing cycle rate and promotes the recyclability and usability of devices from environmental point of view and from technical point of view it cut down the overall cost and save energy and time during installing phase of smart devices with updating service packs and in overall estimation of GHG(green-house gas)emission rate and carbon footprint rate. © 2020 IEEE." "57219486486;7004522901;15730011900;","Potential Bottleneck and Measuring Performance of Serverless Computing: A Literature Study",2020,"10.1109/ICRITO48877.2020.9197837","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093103522&doi=10.1109%2fICRITO48877.2020.9197837&partnerID=40&md5=48c6321f04c6fcb41cb3ef2aa12106cd","Trending form of cloud computing is Serverless computing, where developer just needs to focus on his code rather than worrying about server management. In serverless computing, application is nothing but collection of one or more functions, written for specific business functionality, which triggers on an event. There are various cloud service providers, i.e. Amazon, Microsoft, Google, IBM, etc. who provide serverless services, on pay as you use and auto scalable solution to execute the application code as a function. The developer just needs to upload the code for execution. The performance of the serverless computing may vary due to dynamic configuration of the solution, technologies and different technology used by the service provider.This paper reviews various past and recent work in the serverless computing to identify possible bottlenecks and the scope of measuring performance of serverless computing. It will also put some light to leverage machine learning in various possible ways to do performance engineering for future research. © 2020 IEEE." "57193812309;","Towards developing a classification model for water potability in philippine rural areas",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090490193&partnerID=40&md5=e2966654b973e6810a5402f32e6c59ff","In the Philippines, access to safe and sustainable water source is a major problem especially in rural areas. Thus, water monitoring in different water resources has been practiced to ensure safe drinking water. However, manual monitoring of safe drinking water is known to be inconvenient since it requires high operational and transportation costs, and time consuming. This study develops a data-driven water classification model for rural household areas using sensor nodes and machine learning algorithm. Sensor nodes are installed in several water sources in different rural areas to collect water parameters such as pH, turbidity, total dissolved solids, and temperature which are wirelessly transmitted to a base station. The collected sensor data is used to build and train the model to classify water potability using a hard-voting method in ensemble learning. The ensemble learning combined three machine learning algorithms namely k-nearest Neighbor, Naive Bayes, and Classification and Regression Tree. Finally, data are sent to a cloud for data storage and remote monitoring. Results show that the voting classifier model achieves an accuracy of 97% compared with other stand-alone classification algorithms. Furthermore, the model achieves 90% match with conventional industrial laboratory test. © 2020 ASEAN University Network/Southeast Asia Engineering Education Development Network. All right reserved." "36675430500;57164754800;","Enhanced intrusion detection system for cloud environment using machine learning techniques",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088987770&partnerID=40&md5=ff147155d3fcb7f0accc61de20dfcc07","Tremendous growth in Cloud computing environment brings great happiness to the IT sector. Easy access and pay per use are the attractive glimpse towards the cloud environment. Users can’t put out the cloud though there may be many threatening security problems. This is because they tasted the cloud easy deployment anywhere. But the security compliance is not yet attained its goal. Cloud has endless boundaries in the field so there is much crucial vulnerability arising every time.IDS being a very good solution to switch off the vulnerabilities. This paper is going to discuss about the cloud IDS which is tested using the feature selection method and eliminates unwanted attributes to gain the effective ids. The feature selection takes only the essential attributes from the total 42 features to bring enhanced model to suit the Cloud computing. The effective ids is found using NSL KDD dataset applying on the methods that is MLP and J48 algorithm to bring higher accuracy rate to improve the intrusion detection in the Cloud environment. © 2020 Alpha Publishers. All rights reserved." "57212030072;55772930800;","Sentiment analysis of green product using cloud system",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088973077&partnerID=40&md5=aca7e32eca427e242650f17221432c54","A typical manner in which valuable information can be obtained by extracting the sentiment or opinion from any message is called sentiment analysis. The sentiment classification exploits the technologies in machine learning owing to their ability to learn from training data set to predict and support decision making with high accuracy level. Some algorithms do not maintain proper scalability for large datasets. Today, there is a need to deal with some big datasets for involving features in high numbers. The methods of feature selection have been aiming at the elimination of the noisy, the irrelevant or the redundant features that can bring down the performance of classification. Most of the traditional methods lack the capability to be able to cope with the results within a given time. In this work, Term Frequency (TF) is used for feature extraction. The focus is on the Green product opinion mining done using the Information Gain (IG) in feature selection and it is compared with the Group Search Optimization (GSO). This method of feature selection has reduced the original feature sets by means of removing the irrelevant features that enhance the accuracy of classification and bring down the run time of the learning algorithms. The proposed method has been evaluated using the Support Vector Machine (SVM) classifier. The experimental results have proved that the proposed method had achieved better performance. © 2020 Alpha Publishers. All rights reserved." "57204919704;55636320055;56509029800;57216081524;57214837520;","Time-series prediction of sea level change in the east coast of Peninsular Malaysia from the supervised learning approach",2020,"10.18280/ijdne.150314","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087835571&doi=10.18280%2fijdne.150314&partnerID=40&md5=ac5524177d180a9214fb459c723884ef","Analyzing and predicting the rises in sea level are vital elements in oceanography and marine management especially in managing low-lying coastal areas. The present study aims to analyze the ability of machine learning algorithm viz. regression support vector machine (RSVM) in predicting the changes in the sea level on the east coast of Peninsular Malaysia. The selected inputs for the proposed model are monthly mean sea level (MMSL), monthly sea surface temperature (SST), rainfall and mean cloud cover (MCC) for the period from January 2007 to December 2017. A total of 132 data points for each meteorological parameter were used, where 92 (70%) data points from January 2007 to December 2015 were used for training and 40 (30%) data points from January 2016 to December 2017 were used for validating and testing. Results showed based on the correlation coefficient that the model predicts the sea level rises accurately (R= 0.861, 0.825 and 0.857) for Kerteh, Tanjung Sedili, and Tioman Island, respectively. Moreover, the predicted values were similar to the historical tide-gauge data with very low error, which indicates that the proposed RSVM model can be a promising tool for decision-makers and can be reliable to predict monthly mean sea level rises in Malaysia. © 2020 WITPress. All rights reserved." "57211816997;6602927210;","Fine-tuning self-organizing maps for Sentinel-2 imagery: Separating clouds from bright surfaces",2020,"10.3390/rs12121923","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086987137&doi=10.3390%2frs12121923&partnerID=40&md5=0c00f87308837b0a83a14de070f2b572","Removal of cloud interference is a crucial step for the exploitation of the spectral information stored in optical satellite images. Several cloud masking approaches have been developed through time, based on direct interpretation of the spectral and temporal properties of clouds through thresholds. The problem has also been tackled by machine learning methods with artificial neural networks being among the most recent ones. Detection of bright non-cloud objects is one of the most difficult tasks in cloud masking applications since spectral information alone often proves inadequate for their separation from clouds. Scientific attention has recently been redrawn on self-organizing maps (SOMs) because of their unique ability to preserve topologic relations, added to the advantage of faster training time and more interpretative behavior compared to other types of artificial neural networks. This study evaluated a SOM for cloud masking Sentinel-2 images and proposed a fine-tuning methodology to separate clouds from bright land areas. The fine-tuning process which is based on the output of the non-fine-tuned network, at first directly locates the neurons that correspond to the misclassified pixels. Then, the incorrect labels of the neurons are altered without applying further training. The fine-tuning method follows a general procedure, thus its applicability is broad and not confined only in the field of cloud-masking. The network was trained on the largest publicly available spectral database for Sentinel-2 cloud masking applications and was tested on a truly independent database of Sentinel-2 cloud masks. It was evaluated both qualitatively and quantitatively with the interpretation of its behavior through multiple visualization techniques being a main part of the evaluation. It was shown that the fine-tuned SOM successfully recognized the bright non-cloud areas and outperformed the state-of-the-art algorithms: Sen2Cor and Fmask, as well as the version that was not fine-tuned. © 2020 by the authors." "56326408200;55739545700;57209326451;7102410570;","Data reconstruction for remotely sensed Chlorophyll-a concentration in the ross sea using ensemble-based machine learning",2020,"10.3390/rs12111898","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086444937&doi=10.3390%2frs12111898&partnerID=40&md5=7d084e43b8f9402c226913c1b6962746","Polar regions are too harsh to be continuously observed using ocean color (OC) sensors because of various limitations due to low solar elevations, ice effects, peculiar phytoplankton photosynthetic parameters, optical complexity of seawater and persistence of clouds and fog. Therefore, the OC data undergo a quality-control process, eventually accompanied by considerable data loss. We attempted to reconstruct these missing values for chlorophyll-a concentration (CHL) data using a machine-learning technique based on multiple datasets (satellite and reanalysis datasets) in the Ross Sea, Antarctica. This technique-based on an ensemble tree called random forest (RF)-was used for the reconstruction. The performance of the RF model was robust, and the reconstructed CHL data were consistent with satellite measurements. The reconstructed CHL data allowed a high intrinsic resolution of OC to be used without specific techniques (e.g., spatial average). Therefore, we believe that it is possible to study multiple characteristics of phytoplankton dynamics more quantitatively, such as bloom initiation/termination timings and peaks, as well as the variability in time scales of phytoplankton growth. In addition, because the reconstructed CHL showed relatively higher accuracy than satellite observations compared with the in situ data, our product may enable more accurate planktonic research. © 2020 by the authors." "26032307400;57199127093;6603002503;","Identifying corporate sustainability issues by analyzing shareholder resolutions: A machine-learning text analytics approach",2020,"10.3390/su12114753","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086368657&doi=10.3390%2fsu12114753&partnerID=40&md5=bc75d52091851fe8011c32d5ab9f9fa8","Corporations have embraced the idea of corporate environmental, social, and governance (ESG) under the general framework of sustainability. Studies have measured and analyzed the impact of internal sustainability efforts on the performance of individual companies, policies, and projects. This exploratory study attempts to extract useful insight from shareholder sustainability resolutions using machine learning-based text analytics. Prior research has studied corporate sustainability disclosures from public reports. By studying shareholder resolutions, we gain insight into the shareholders' perspectives and objectives. The primary source for this study is the Ceres sustainability shareholder resolution database, with 1737 records spanning 2009-2019. The study utilizes a combination of text analytic approaches (i.e., word cloud, co-occurrence, row-similarities, clustering, classification, etc.) to extract insights. These are novel methods of transforming textual data into useful knowledge about corporate sustainability endeavors. This study demonstrates that stakeholders, such as shareholders, can influence corporate sustainability via resolutions. The incorporation of text analytic techniques offers insight to researchers who study vast collections of unstructured bodies of text, improving the understanding of shareholder resolutions and reaching a wider audience. © 2020 by the authors." "35098748100;57205299839;57216524741;55576700800;56095856700;51461664100;36815873200;8284949000;57214786637;8437626600;","A review of the estimation of downward surface shortwave radiation based on satellite data: Methods, progress and problems",2020,"10.1007/s11430-019-9589-0","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083806719&doi=10.1007%2fs11430-019-9589-0&partnerID=40&md5=922d29aa22cdaad3079ccfa7a6122e9d","The estimation of downward surface shortwave radiation (DSSR) is important for the Earth’s energy budget and climate change studies. This review was organised from the perspectives of satellite sensors, algorithms and future trends, retrospects and summaries of the satellite-based retrieval methods of DSSR that have been developed over the past 10 years. The shortwave radiation reaching the Earth’s surface is affected by both atmospheric and land surface parameters. In recent years, studies have given detailed considerations to the factors which affect DSSR. It is important to improve the retrieval accuracy of cloud microphysical parameters and aerosols and to reduce the uncertainties caused by complex topographies and high-albedo surfaces (such as snow-covered areas) on DSSR estimation. This review classified DSSR retrieval methods into four categories: empirical, parameterisation, look-up table and machine-learning methods, and evaluated their advantages, disadvantages and accuracy. Further efforts are needed to improve the calculation accuracy of atmospheric parameters such as cloud, haze, water vapor and other land surface parameters such as albedo of complex terrain and bright surface, organically combine machine learning and other methods, use the new-generation geostationary satellite and polar orbit satellite data to produce high-resolution DSSR products, and promote the application of radiation products in hydrological and climate models. © 2020, Science China Press and Springer-Verlag GmbH Germany, part of Springer Nature." "56313481400;57146989100;7102653996;36799026400;","Hyper-local, efficient extreme heat projection and analysis using machine learning to augment a hybrid dynamical-statistical downscaling technique",2020,"10.1016/j.uclim.2020.100606","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079842964&doi=10.1016%2fj.uclim.2020.100606&partnerID=40&md5=00c814a5ce2b487119121e119c9ca6f6","This paper describes a scalable system for quantifying hyper-local heat stress in urban environments and its expected response within the changing climate. A hybrid dynamical-statistical downscaling approach links Global Climate Models (GCMs) with dynamically downscaled extreme heat events using the Weather Research and Forecasting model (WRF). Downscaled historical simulations in WRF incorporate urban canopy physics to better describe localized land surface details in the urban environment relevant to extreme heat. This downscaled library is then used in an analog-based approach. This contribution reports a series of enhancements to existing analog-based methods which can efficiently produce more detailed results. The system here uses advanced statistical methods and simple machine learning (ML) techniques to optimize analog selection, perform spatially-consistent bias correction, and decompose patterns of extreme heat into dynamic components such as the land-sea contrast and inland sea-breeze penetration. Hindcast projections are validated against observational data from in-situ weather observing stations. The results demonstrate the scalability and efficiency of this system as it is deployed in cloud-based architectures with parallelized code. Downscaled predictions are equally applicable to heat stress at weather and climate time scales, supporting infrastructure resilience and adaptation, and emergency response. © 2020 Elsevier B.V." "55174951200;36537144100;57210959173;57216249671;57219418376;","Development of an Assistive Device via Smart Glasses",2020,"10.1109/ECBIOS50299.2020.9203629","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85092632906&doi=10.1109%2fECBIOS50299.2020.9203629&partnerID=40&md5=a5f71bbe1bba1a16e01b6c46760d7b7b","This study used smart glasses as the carrier to develop three applications aimed at the elderly. Taiwan entered an elderly society in 2018, with the proportion of the elderly population exceeding 14%. It is estimated that the elderly population will exceed 20% in 2026, and Taiwan will enter an ultra-elderly society. To improve the quality of life of the elderly and attach importance to the health care of the elderly, more and more scientific and technological products are used to improve the life and consumption patterns of the elderly, thereby promoting the popularization of smart wearable devices and sensing technologies. In view of the purpose, this research develops related functional applications with smart glasses. First, the memory recall mechanism is carried out based on face recognition. When the user uses the function, an embedded camera in glasses is turned on to capture the target personnel. Afterwards, a KNN (K Nearest Neighbor) machine learning approach is applied to identify the target, and the personal data is performed in the glasses through searching on the developed cloud database. Besides, the user can add, delete and modify the personal data through the computer, which the entrance is carried out by scanning the two-dimension code shown in the computer program with glasses for security purpose. Secondly, the image enhancement function shows that the captured texts can be expanded and enlarged in the projection monitor of glasses. Finally, the construction of voice commands is performed through the microphone within the suite of smart glasses. The results show that the accuracy rate of KNN face recognition is 93.3%, which can be applied to general life situations. © 2020 IEEE." "57218116828;57218283904;57218120506;","Mobile mapping, machine learning and digital twin for road infrastructure monitoring and maintenance: Case study of mohammed VI bridge in Morocco",2020,"10.1109/Morgeo49228.2020.9121882","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087967317&doi=10.1109%2fMorgeo49228.2020.9121882&partnerID=40&md5=f64c9686dd6e9bcc47bedf0562162dde","The concepts of Digital Twin has been recently introduced, it refers to functional connections between a complex physical system and its high-fidelity digital replica. Digital Twin process workflow is proposed in case of Mohammed VI Bridge modeling in Morocco. The current maintenance of a road infrastructure is based on a manual inspection and a system based on traditional tools. Aging infrastructures require a new approach to maintenance in terms of inspection, bridge maintenance system, simulation and systematic evaluation. This system now exists and is called the Digital Twin. Digital Twin can be thought of as a virtual prototype in service that changes dynamically in near real time as its physical twin changes. An urban infrastructure digital twin is a virtual instance of his physical twin that is continuously updated with multisource, multisensor and multitemporal data that can be used for monitoring, simulating and forecasting any potential problem that may appear in the structure and proposing planning for repair and maintenance of health status throughout the life cycle of this infrastructure. This work presents a general vision and a justification for integrating DT technology with geospatial data. The paper examines the benefits of integrating 3D GIS data acquired by automated mobile mapping (MMS) workflows for modeling the reality of a major bridge infrastructure in Morocco. This allowed to study the future performance of this bridge structure on virtual twin structures under different environmental conditions. Cloud point data are acquired by a Mobile Mapping System on Mohammed VI Bridge and converted in BIM model by a scan to BIM process and is integrated in a GIS and BIM virtual environment and shows the efficiency of volumetric auscultation in terms of surface flatness and distortion inspection. This project provides a new bridge maintenance system using the concept of a Digital Twin. This digital model is a platform that allows to collect, organize and share the maintenance history of this important road infrastructure in Morocco. © 2020 IEEE." "36615943200;57218539795;35180845800;15031611500;37046365700;56477626600;56372294100;57196301839;14060679800;","Searching for Molecular Outflows with Support Vector Machines: The Dark Cloud Complex in Cygnus",2020,"10.3847/1538-4365/ab879a","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85087208421&doi=10.3847%2f1538-4365%2fab879a&partnerID=40&md5=2afa08c0212e32f3e08e2e163b0697d7","We present a survey of molecular outflows across the dark cloud complex in the Cygnus region, based on a 46.75 deg2 field of CO isotopologue data from the Milky Way Imaging Scroll Painting survey. A supervised machine-learning algorithm, the support vector machine, is introduced to accelerate our visual assessment of outflow features in the data cube of 12CO and 13CO J = 1-0 emission. A total of 130 outflow candidates are identified, 77 of which show bipolar structures and 118 are new detections. Spatially, these outflows are located inside dense molecular clouds, and some of them are found in clusters or in elongated linear structures tracing the underlying gas filament morphology. Along the line of sight, 97, 31, and 2 candidates reside in the Local, Perseus, and Outer Arms, respectively. Young stellar objects as outflow drivers are found near most outflows, while 36 candidates show no associated source. The clusters of outflows that we detect are inhomogeneous in their properties; nevertheless, we show that the outflows cannot inject turbulent energy on cloud scales. Instead, at best, they are restricted to affecting the so-called ""clump""and ""core""scales, and only on short (∼0.3 Myr) estimated timescales. Combined with outflow samples in the literature, our work shows a tight outflow mass-size correlation. © 2020. The American Astronomical Society. All rights reserved." "56094876500;14821958500;35581315600;57208478400;57216672036;7006392180;","Maximum fraction images derived from year-based project for on-board autonomy-vegetation (PROBA-V) data for the rapid assessment of land use and land cover areas in Mato Grosso state, Brazil",2020,"10.3390/LAND9050139","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085590974&doi=10.3390%2fLAND9050139&partnerID=40&md5=a298842ae08e04076da51a961ace5866","This paper presents a new approach for rapidly assessing the extent of land use and land cover (LULC) areas in Mato Grosso state, Brazil. The novel idea is the use of an annual time series of fraction images derived from the linear spectral mixing model (LSMM) instead of original bands. The LSMM was applied to the Project for On-Board Autonomy-Vegetation (PROBA-V) 100-m data composites from 2015 (~73 scenes/year, cloud-free images, in theory), generating vegetation, soil, and shade fraction images. These fraction images highlight the LULC components inside the pixels. The other new idea is to reduce these time series to only six single bands representing the maximum and standard deviation values of these fraction images in an annual composite, reducing the volume of data to classify the main LULC classes. The whole image classification process was conducted in the Google Earth Engine platform using the pixel-based random forest algorithm. A set of 622 samples of each LULC class was collected by visual inspection of PROBA-V and Landsat-8 Operational Land Imager (OLI) images and divided into training and validation datasets. The performance of the method was evaluated by the overall accuracy and confusion matrix. The overall accuracy was 92.4%, with the lowest misclassification found for cropland and forestland (<9% error). The same validation data set showed 88% agreement with the LULC map made available by the Landsat-based MapBiomas project. This proposed method has the potential to be used operationally to accurately map the main LULC areas and to rapidly use the PROBA-V dataset at regional or national levels. © 2020 by the authors." "57006249500;36803116800;57211112377;57201257574;8593424100;57200294885;","Integration of multi-sensor data to estimate plot-level stem volume using machine learning algorithms-case study of evergreen conifer planted forests in Japan",2020,"10.3390/rs12101649","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085577942&doi=10.3390%2frs12101649&partnerID=40&md5=b6e6c0ada259f8e19c630b9cd63def67","The development of new methods for estimating precise forest structure parameters is essential for the quantitative evaluation of forest resources. Conventional use of satellite image data, increasing use of terrestrial laser scanning (TLS), and emerging trends in the use of unmanned aerial systems (UASs) highlight the importance of modern technologies in the realm of forest observation. Each technology has different advantages, and this work seeks to incorporate multiple satellite, TLS-and UAS-based remote sensing data sets to improve the ability to estimate forest structure parameters. In this paper, two regression analysis approaches are considered for the estimation: random forest regression (RFR) and support vector regression (SVR). To collect the dependent variable, in situ measurements of individual tree parameters (tree height and diameter at breast height (DBH) were taken in a Japanese cypress forest using the nondestructive TLS method, which scans the forest to obtain dense and accurate point clouds under the tree canopy. Based on the TLS data, the stem volume was then computed and treated as ground truth information. Topographic and UAS information was then used to calculate various remotely sensed explanatory variables, such as canopy size, canopy cover, and tree height. Canopy cover and canopy shapes were computed via the orthoimages derived from the UAS and watershed segmentation method, respectively. Tree height was computed by combining the digital surface model (DSM) from the UAS and the digital terrain model (DTM) from the TLS data. Topographic variables were computed from the DTM. The backscattering intensity in the satellite imagery was obtained based on L-band (Advanced Land Observing Satellite-2 (ALOS-2) Phased Array type L-band Synthetic Aperture Radar-2 (PALSAR-2) and C-band (Sentinel-1) synthetic aperture radar (SAR). All satellite (10-25 m resolution), TLS (3.4 mm resolution) and UAS (2.3-4.6 cm resolution) data were then combined, and RFR and SVR were trained; the resulting predictive powers were then compared. The RFR method yielded fitting R2 up to 0.665 and RMSE up to 66.87 m3/ha (rRMSE = 11.95%) depending on the input variables (best result with canopy height, canopy size, canopy cover, and Sentinel-1 data), and the SVR method showed fitting R2 up to 0.519 and RMSE up to 80.12 m3/ha (rRMSE = 12.67%). The RFR outperformed the SVR method, which could delineate the relationship between the variables for better model accuracy. This work has demonstrated that incorporating various remote sensing data to satellite data, especially adding finer resolution data, can provide good estimates of forest parameters at a plot level (10 by 10 m), potentially allowing advancements in precision forestry. © 2020 by the authors." "57194068533;7006499360;6603916527;7202643131;","Sentinel-1 observation frequency significantly increases burnt area detectability in tropical SE Asia",2020,"10.1088/1748-9326/ab7765","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085275425&doi=10.1088%2f1748-9326%2fab7765&partnerID=40&md5=78c6cfc1ee28121ba469faa080162086","Frequent cloud cover in the tropics significantly affects the observation of the surface by satellites. This has enormous implications for current approaches that estimate greenhouse gas (GHG) emissions from fires or map fire scars. These mainly employ data acquired in the visible to middle infrared bands to map fire scars or thermal data to estimate fire radiative power and consequently derive emissions. The analysis here instead explores the use of microwave data from the operational Sentinel-1A (S-1A) in dual-polarisation mode (VV and VH) acquired over Central Kalimantan during the 2015 fire season. Burnt areas were mapped in three consecutive periods between August and October 2015 using the random forests machine learning algorithm. In each mapping period, the omission and commission errors of the unburnt class were always below 3%, while the omission and commission errors of the burnt class were below 20% and 5% respectively. Summing the detections from the three periods gave a total burnt area of ∼1.6 million ha, but this dropped to ∼1.2 million ha if using only a pair of pre- and post-fire season S-1A images. Hence the ability of Sentinel-1 to make frequent observations significantly increases fire scar detection. Comparison with burnt area estimates from the Moderate Resolution Imaging Spectroradiometer (MODIS) burnt area product at 5 km scale showed poor agreement, with consistently much lower estimates produced by the MODIS data-on average 14%-51% of those obtained in this study. The method presented in this study offers a way to reduce the substantial errors likely to occur in optical-based estimates of GHG emissions from fires in tropical areas affected by substantial cloud cover. © 2020 The Author(s). Published by IOP Publishing Ltd." "57215057713;37089364100;57215056447;47962102500;37088003400;55626446400;","Improvement of the Rapid-Development Thunderstorm (RDT) Algorithm for Use with the GK2A Satellite",2020,"10.1007/s13143-020-00182-6","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079799174&doi=10.1007%2fs13143-020-00182-6&partnerID=40&md5=b6dabc24aafc3aaef3f706875d78a457","New technologies for the classification of convective cloud lifecycles and the prediction of their movements are needed to detect severe convective weather and to support objective cloud guidance. Satellites enable earlier detection of severe weather over larger coverage areas than ground-based observations or radar. The use of satellite observations for nowcasting is thus likely. In this study, convective initiation (CI) data are paired with a modified rapid-development thunderstorm (RDT) algorithm for the analysis of new data from the Geostationary Korea Multi-Purpose Satellite-2A (GEO-KOMPSAT-2A, GK2A). The RDT algorithm is further modified to accommodate the additional GK2A satellite channels, and new satellite data are used to continuously analyze thunderstorms associated with severe weather in Korea. The logistic regression (LR) machine learning approach is used to optimize the criteria of interest fields and weighting coefficients of the RDT algorithm for convective detection. In addition, auxiliary data (cloud type, convective rainfall rate, and cloud top temperature/height) calculated from RDT sub-module is replaced with GK2A derived products. The fully modified RDT algorithm (K-RDT) is quantitatively verified using lightning data from summer convection cases. The probability of detection (POD) for convective clouds is increased by 30–40%, and the threat score (TS) for average lightning activity is improved by 10–30%. The channel properties of Japan Himawari-8 satellite are similar to those of the GK2A satellite. Due to the lack of GK2A satellite data during the development period, CI data from the Himawari-8 satellite are used as proxies. © 2020, Korean Meteorological Society and Springer Nature B.V." "57216206627;","Reaction: Can We Grow a Quantum Processor?",2020,"10.1016/j.chempr.2020.03.018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082831288&doi=10.1016%2fj.chempr.2020.03.018&partnerID=40&md5=9b352f6a55423ef6b2a37d05df410c13","Alexei Marchenkov is an experimental physicist who has devoted most of his research career to studying quantum phenomena in materials and structures at cryogenic temperatures, including graphene, atomic-sized wires and contacts, and superconducting quantum bits and circuits. He is a former associate professor at the Georgia Institute of Technology and founder and chief executive officer of the Bleximo Corp., a startup developing application-specific superconducting quantum computers. © 2020 Elsevier Inc.Today's quantum computers contain tens of qubits and are readily available for experimentation in the cloud. However, building quantum computers capable of outperforming digital supercomputers in practical simulation, optimization, and machine-learning tasks requires more coherent qubits assembled in larger processors. Although molecular electron-spin qubits have shown early promise, validating whether they can provide a viable platform for practical quantum processors and compete with the leading superconducting and trapped-ion modalities requires a significant experimental effort. © 2020 Elsevier Inc." "57117012400;57216413349;57193791654;57211840394;","The potential of integrating blockchain technology into smart sustainable city development",2020,"10.1088/1755-1315/463/1/012020","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083428106&doi=10.1088%2f1755-1315%2f463%2f1%2f012020&partnerID=40&md5=51493834ae51217e05553b3da09f1314","The rise of global urbanisation has led to massive pressures on resources such as food, water, infrastructure, and energy demand to support growing populations. It brings adverse impacts on the liveable condition and economic growth of a country if this problem remains unsolved. Smart city is a potential solution to address the challenges of urbanisation by leveraging the technological breakthrough such as internet of things (IoT), Artificial Intelligence (AI), machine learning, big data, and cloud computing to facilitate scarce resources planning and management. With numerous connected devices and vast communication networks, it poses a challenges of security threat which cannot be addressed by the conventional cybersecurity solutions. Blockchain offers a solution in securing the huge numbers of connected devices in smart city network. The application of blockchain technology is leading in the banking and financial industry. However, the uses and implementations in smart city have emerged in recent years. The combination of blockchain technology and smart city has offered a great potential for sustainable development. Thus, it is imperative to discuss the potential of these two elements in making the city safer and sustainable. This paper explores how the blockchain technology application can help in managing smart city and achieve sustainability. The findings revealed that there are five key areas of blockchain application in smart city which are smart governance, smart mobility, smart asset, smart utility and smart logistic. A framework for smart sustainable city with blockchain technology is presented as an outcome of this study. It gives a clear overview for the policy makers and regulators of how blockchain supports within smart city framework. It facilitates the transition towards smart and sustainable cities through the use of blockchain. © 2020 Institute of Physics Publishing. All rights reserved." "55811505000;56589932000;57148539400;57211159673;","An optimal machine learning classifier for cloud based web URL phishing detection",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088855154&partnerID=40&md5=201849c9b90764b0fba7965940c306d6","Cloud based Web URL phishing is a crucial threat in the cyber world a decade before and even till now. An enormous web server instance is created in the cloud platform and hosted online. Phishing is an act of impersonating to achieve sensitive information from the intended recipient by deceiving them. The users access information through various mediums such as websites, emails, mobile applications, SMS, and social media. Phishing is demonstrated using a malicious URL that the attacker replicates from the original website and sends it to the target or by using a mail that injects malware in the target’s computer to obtain sensitive information. Malicious URL is detected using machine learning techniques that efficiently identifies a malicious website or a new website. Random Forest's anti-phishing classifier has outperformed to accomplish a genuine positive rate as 98.42% and false positive rate as 1.58% compared to all other ML classifiers. © 2020 Alpha Publishers." "57203397955;57216490159;57218081287;","An Efficient Solution for Semantic Segmentation of Three Ground-based Cloud Datasets",2020,"10.1029/2019EA001040","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083636404&doi=10.1029%2f2019EA001040&partnerID=40&md5=96ab039a3967ec6569507b3fe7d657e5","The machine learning approach has shown its state-of-the-art ability to handle segmentation and detection tasks. It is increasingly employed to extract patterns and spatiotemporal features from the ever-increasing stream of Earth system data. However, there is still a significant challenge, which is the generalization capability of the model on cloud images in different types and weather conditions. After studying several popular methods, we propose a semantic segmentation neural network for cloud segmentation. It extracts features learned by source and target domains in an end-to-end behavior, which can address the problem of significant lack of labels in the observed cloud image data. It is further evaluated on the Singapore Whole Sky Image Segmentation (SWIMSEG) dataset by using Mean Intersection-over-Union, recall, F-score, and accuracy matrices. The scores of these matrices are 86%, 97%, 92%, and 96%, which prove that it has excellent efficiency and robustness. Most importantly, a new benchmark based on the SWIMSEG dataset for the task of cloud segmentation is introduced. The others, BENCHMARK, Cirrus Cumulus Stratus Nimbus are evaluated through the model trained from the SWIMSEG dataset by way of visualization. ©2020. The Authors." "55941476900;57014496500;57197784493;55612256200;","Snow cover estimation from MODIS and Sentinel-1 SAR data using machine learning algorithms in the western part of the Tianshan Mountains",2020,"10.1007/s11629-019-5723-1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083324480&doi=10.1007%2fs11629-019-5723-1&partnerID=40&md5=de77ff3430d69bad6df3d3b4017216cc","Obtaining the spatial distribution of snow cover in mountainous areas using the optical image of remote sensing technology is difficult because of cloud and fog. In this study, the object-based principle component analysis-support vector machine (PCA-SVM) method is proposed for snow cover mapping through the integration of moderateresolution imaging spectroradiometer (MODIS) snow cover products and the Sentinel-1 synthetic aperture radar (SAR) scattering characteristics. First, derived from the Sentinel-1A SAR images, the feature parameters, including VV/VH backscatter, scattering entropy, and scattering alpha, were used to describe the variations of snow and non-snow covers. Second, the optimum feature combinations of snow cover were formed from the feature parameters using the principle component analysis (PCA) algorithm. Finally, using the optimum feature combinations, a snow cover map with a 20 m spatial resolution was extracted by means of an object-based SVM classifier. This method was applied in the study area of the Xinyuan County, which is located in the western part of the Tianshan Mountains in Xinjiang, China. The accuracies in this method were analyzed according to the data observed at different experimental sites. Results showed that the snow cover pixels of the extraction were less than those in the actual situation (FB1=93.86, FB2=59.78). The evaluation of the threat score (TS), probability of detection (POD), and false alarm ratio (FAR) for the snow-covered pixels obtained from the two-stage SAR images were different (TS1=86.84, POD1=90.10, FAR1=4.01; TS2=56.40, POD2=57.62, FAR2=3.62). False and misclassifications of the snow cover and non-snow cover pixels were found. Although the classifications were not highly accurate, the approach showed potential for integrating different sources to retrieve the spatial distribution of snow covers during a stable period. © 2020, Science Press, Institute of Mountain Hazards and Environment, CAS and Springer-Verlag GmbH Germany, part of Springer Nature." "57202642788;7202994044;56333692000;57192701006;57202917893;","A Three-Dimensional Deep Learning Framework for Human Behavior Analysis Using Range-Doppler Time Points",2020,"10.1109/LGRS.2019.2930636","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082883574&doi=10.1109%2fLGRS.2019.2930636&partnerID=40&md5=d598b8eec8031fdc8f2c1d3119c3ec99","Deep neural networks have shown promise in the radar-based human activity analysis application. Different from existing deep learning models that take either micro-Doppler spectrograms or range profiles as their input, the proposed method can process micromotion signatures in a 3-D way. In this letter, we first transform radar echoes into range-Doppler (RD) time points and then directly process the point sets via a designed 3-D network called the RD PointNet. In fact, our point model is a discrete representation of the motion trajectory. Through this quantitative model, we can use the 3-D network to simultaneously capture human motion profiles and temporal variations. The motion capture simulations and ultrawideband radar measurements show that the proposed framework can achieve superior classification accuracy and noise robustness when compared with image-based methods. © 2004-2012 IEEE." "57215037375;36443824600;49661020300;57212631853;55738672900;56026541800;56168282700;","Climate-based approach for modeling the distribution of montane forest vegetation in Taiwan",2020,"10.1111/avsc.12485","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85081245150&doi=10.1111%2favsc.12485&partnerID=40&md5=091dcef2a352bcb71f80ce5eabdd3024","Aims: Climate shapes forest types on our planet and also drives the differentiation of zonal vegetation at regional scale. A climate-based ecological model may provide an effective alternative to the traditional approach for assessing limitations, thresholds, and the potential distribution of forests. The main objective of this study is to develop such a model, with a machine-learning approach based on scale-free climate variable estimates and classified vegetation plots, to generate a fine-scale predicted vegetation map of Taiwan, a subtropical mountainous island. Location: Taiwan. Methods: A total of 3,824 plots from 13 climate-related forest types and 57 climatic variable estimates for each plot were used to build an individual ecological niche model for each forest type with random forest (RF). A predicted vegetation map was developed through the assemblage of RF predictions for each forest type at the spatial resolution of 100 m. The accuracy of the ensemble RF model was evaluated by comparing the predicted forest type with its original classification by plot. Results: The climate environment of regions higher than 100 m above sea level in Taiwan was classified into potential habitats of 13 forest types by using model predictions. The predicted vegetation map displays a distinct altitudinal zonation from subalpine to montane cloud forests, followed by the latitudinal differentiation of subtropical mountain forests in the north and tropical montane forests in the south, with an average mismatch rate of 6.59%. An elevational profile and 3D visualization demonstrate the excellence of the model in estimating a fine, precise, and topographically corresponding potential distribution of forests. Conclusions: The machine-learning approach is effective for handling a large number of variables and to provide accurate predictions. This study provides a statistical procedure integrating two sources of training data: (a) the locations of field sampling plots; and (b) their corresponding climate variable estimates, to predict the potential distribution of climate-related forests. © 2020 International Association for Vegetation Science" "57216035656;57189929871;","Impact of Hurricane Maria on Beach Erosion in Puerto Rico: Remote Sensing and Causal Inference",2020,"10.1029/2020GL087306","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082532550&doi=10.1029%2f2020GL087306&partnerID=40&md5=398ac8ebf167235c5d24785bf37fbbdb","Beach erosion due to large storms critically affects coastal vulnerability, but is challenging to monitor and quantify. Attributing erosion to a specific storm requires a reliable counterfactual scenario: hypothetical beach conditions, absent the storm. Calibrating models to construct counterfactuals requires numerous observations that are rarely available. Storm paths are unpredictable, making long-term instrumentation of specific beaches costly. Optical remote sensing is hampered by persistent cloud cover. We use Sentinel-1 satellite radar imagery to monitor shoreline changes through clouds and propose regression discontinuity as a strategy to estimate the causal effect of large storms on beach erosion. Applied to 75 beaches across Puerto Rico, the approach detects shoreline changes with a root-mean-square error comparable to the resolution of the imagery. Hurricane Maria caused an erosion of 3 to 5 m along its path, up to 40 m at particular beaches. Results reveal strong local disparities that are consistent with simulated nearshore hydrodynamic conditions. ©2020. American Geophysical Union. All Rights Reserved." "57216129885;","Research on the Application of Artificial Intelligence in Energy Science and Engineering Monitoring Software Engineering Technology under the Background of Big Data",2020,"10.1088/1755-1315/440/3/032058","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082627416&doi=10.1088%2f1755-1315%2f440%2f3%2f032058&partnerID=40&md5=a1f26bac185a5867810e292641832fa4","Artificial intelligence belongs to the world's leading technology and technology, which can make people's work and life become more intelligent and faster. The arrival of the era of big data accelerates the development of artificial intelligence technology, and the artificial intelligence big data platform begins to appear. Through the analysis of the relationship between big data and artificial intelligence, the AI big data platform and its application are comprehensively discussed, aiming to promote the development of artificial intelligence big data platform and improve the overall level of domestic artificial intelligence. The smart grid is artificial. One of the important application areas of intelligence, the advancement and breakthrough of the new generation of artificial intelligence technology, which is mainly represented by advanced machine learning theory, big data and cloud computing, will promote the development of smart grid. Therefore, an energy science and engineering data monitoring software is needed to realize the monitoring and analysis and unified scheduling of energy science. The article details the design and implementation of data monitoring software for power and energy intelligence. © 2020 Published under licence by IOP Publishing Ltd." "57212588620;57005349200;15846943000;57212587521;57212585560;","Point-cloud detection of buildings based on a latent Dirichlet allocation model with waveform data",2020,"10.1080/2150704X.2019.1706006","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077157679&doi=10.1080%2f2150704X.2019.1706006&partnerID=40&md5=0ede6552b6aa57c4b5daf7c97aa1ecd6","Light detection and ranging are important methods for acquiring digital surface models and can be used to extract building data. Point-cloud detection of buildings is a prerequisite for the model-based expression of buildings. Existing methods are insufficient because of their abstractness of feature extraction and poor accuracy of the detection results. This paper proposes a method for the point-cloud detection of buildings based on a latent Dirichlet allocation (LDA) model with waveform data. This method can extract waveform data via the global convergence Levenberg Marquard algorithm, convert discrete point clouds into point-cluster objects via super voxel segmentation, and detect the point clouds of buildings via the LDA model. Moreover, it supports vector machine classification. Experimental results demonstrate that waveform features and the LDA model both improve the accuracy of building detection. In addition, this method is less susceptible to variations in feature dimensions and is robust in terms of the number of topics and words. © 2019, © 2019 Informa UK Limited, trading as Taylor & Francis Group." "57219195975;57193749143;","Estimating Crop Yields with Remote Sensing and Deep Learning",2020,"10.1109/LAGIRS48042.2020.9165608","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091676099&doi=10.1109%2fLAGIRS48042.2020.9165608&partnerID=40&md5=bc587f96d7c10752b3d83087d268f4ba","Increasing the accuracy of crop yield estimates may allow improvements in the whole crop production chain, allowing farmers to better plan for harvest, and for insurers to better understand risks of production, to name a few advantages. To perform their predictions, most current machine learning models use NDVI data, which can be hard to use, due to the presence of clouds and their shadows in acquired images, and due to the absence of reliable crop masks for large areas, especially in developing countries. In this paper, we present a deep learning model able to perform pre-season and in-season predictions for five different crops. Our model uses crop calendars, easy-to-obtain remote sensing data and weather forecast information to provide accurate yield estimates. © 2020 IEEE." "56494144900;57219200521;6602269604;36623278200;36136174500;","Brazildam: A Benchmark Dataset for Tailings Dam Detection",2020,"10.1109/LAGIRS48042.2020.9165620","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091634482&doi=10.1109%2fLAGIRS48042.2020.9165620&partnerID=40&md5=97f530e672ba28f5f148d1939522ae9d","In this work we present BrazilDAM, a novel public dataset based on Sentinel-2 and Landsat-8 satellite images covering all tailings dams cataloged by the Brazilian National Mining Agency (ANM). The dataset was built using georeferenced images from 769 dams, recorded between 2016 and 2019. The time series were processed in order to produce cloud free images. The dams contain mining waste from different ore categories and have highly varying shapes, areas and volumes, making BrazilDAM particularly interesting and challenging to be used in machine learning benchmarks. The original catalog contains, besides the dam coordinates, information about: the main ore, constructive method, risk category, and associated potential damage. To evaluate BrazilDAM's predictive potential we performed classification essays using state-of-the-art deep Convolutional Neural Network (CNNs). In the experiments, we achieved an average classification accuracy of 94.11% in tailing dam binary classification task. In addition, others four setups of experiments were made using the complementary information from the original catalog, exhaustively exploiting the capacity of the proposed dataset. © 2020 IEEE." "57198861721;7006011851;37023334200;","Characterizing lognormal fractional-Brownian-motion density fields with a convolutional neural network",2020,"10.1093/mnras/staa122","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091418077&doi=10.1093%2fmnras%2fstaa122&partnerID=40&md5=8e2e9b830de56de95eda55fca52d2044","In attempting to quantify statistically the density structure of the interstellar medium, astronomers have considered a variety of fractal models. Here, we argue that, to properly characterize a fractal model, one needs to define precisely the algorithm used to generate the density field, and to specify – at least – three parameters: one parameter constrains the spatial structure of the field, one parameter constrains the density contrast between structures on different scales, and one parameter constrains the dynamic range of spatial scales over which self-similarity is expected (either due to physical considerations, or due to the limitations of the observational or numerical technique generating the input data). A realistic fractal field must also be noisy and non-periodic. We illustrate this with the exponentiated fractional Brownian motion (xfBm) algorithm, which is popular because it delivers an approximately lognormal density field, and for which the three parameters are, respectively, the power spectrum exponent, β, the exponentiating factor, S, and the dynamic range, R. We then explore and compare two approaches that might be used to estimate these parameters: machine learning and the established ∆-Variance procedure. We show that for 2 ≤ β ≤ 4 and 0 ≤ S ≤ 3, a suitably trained Convolutional Neural Network is able to estimate objectively both β (with root-mean-square error ∊β ∼ 0.12) and S (with ∊S ∼ 0.29). ∆-variance is also able to estimate β, albeit with a somewhat larger error (∊β ∼ 0.17) and with some human intervention, but is not able to estimate S. © 2020 The Author(s)" "6506328135;34881780600;7005446873;6701754792;","A Machine Learning Assisted Development of a Model for the Populations of Convective and Stratiform Clouds",2020,"10.1029/2019MS001798","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083242186&doi=10.1029%2f2019MS001798&partnerID=40&md5=29d46bf7c828c63bc3cf2fea84e4afd2","Traditional parameterizations of the interaction between convection and the environment have relied on an assumption that the slowly varying large-scale environment is in statistical equilibrium with a large number of small and short-lived convective clouds. They fail to capture nonequilibrium transitions such as the diurnal cycle and the formation of mesoscale convective systems as well as observed precipitation statistics and extremes. Informed by analysis of radar observations, cloud-permitting model simulation, theory, and machine learning, this work presents a new stochastic cloud population dynamics model for characterizing the interactions between convective and stratiform clouds, with the goal of informing the representation of these interactions in global climate models. Fifteen wet seasons of precipitating cloud observations by a C-band radar at Darwin, Australia are fed into a machine learning algorithm to obtain transition functions that close a set of coupled equations relating large-scale forcing, mass flux, the convective cell size distribution, and the stratiform area. Under realistic large-scale forcing, the derived transition functions show that, on the one hand, interactions with stratiform clouds act to dampen the variability in the size and number of convective cells and therefore in the convective mass flux. On the other, for a given convective area fraction, a larger number of smaller cells is more favorable for the growth of stratiform area than a smaller number of larger cells. The combination of these two factors gives rise to solutions with a few convective cells embedded in a large stratiform area, reminiscent of mesoscale convective systems. ©2020. The Authors." "57200602927;57191057256;15034793900;56233739200;56579557400;55640064500;57214086717;57203541311;","Learning sequential slice representation with an attention-embedding network for 3D shape recognition and retrieval in MLS point clouds",2020,"10.1016/j.isprsjprs.2020.01.003","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078134393&doi=10.1016%2fj.isprsjprs.2020.01.003&partnerID=40&md5=5f832be3a7282de12eccc8d6b31782c2","The representation of 3D data is the key issue for shape analysis. However, most of the existing representations suffer from high computational cost and structure information loss. This paper presents a novel sequential slice representation with an attention-embedding network, named RSSNet, for 3D point cloud recognition and retrieval in road environments. RSSNet has two main branches. Firstly, a sequential slice module is designed to map disordered 3D point clouds to ordered sequence of shallow feature vectors. A gated recurrent unit (GRU) module is applied to encode the spatial and content information of these sequential vectors. The second branch consists of a key-point based graph convolution network (GCN) with an embedding attention strategy to fuse the sequential and global features to refine the structure discriminability. Three datasets were used to evaluate the proposed method, one acquired by our mobile laser scanning (MLS) system and two public datasets (KITTI and Sydney Urban Objects). Experimental results indicated that the proposed method achieved better performance than recognition and retrieval state-of-the-art methods. RSSNet provided recognition rates of 98.08%, 95.77% and 70.83% for the above three datasets, respectively. For the retrieval task, RSSNet obtained excellent mAP values of 95.56%, 87.16% and 69.99% on three datasets, respectively. © 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS)" "6602981892;","Geospatial Artificial Intelligence: Potentials of Machine Learning for 3D Point Clouds and Geospatial Digital Twins",2020,"10.1007/s41064-020-00102-3","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85080980476&doi=10.1007%2fs41064-020-00102-3&partnerID=40&md5=423f399f054e2449ed47dcf3b3c110c1","Artificial intelligence (AI) is changing fundamentally the way how IT solutions are implemented and operated across all application domains, including the geospatial domain. This contribution outlines AI-based techniques for 3D point clouds and geospatial digital twins as generic components of geospatial AI. First, we briefly reflect on the term “AI” and outline technology developments needed to apply AI to IT solutions, seen from a software engineering perspective. Next, we characterize 3D point clouds as key category of geodata and their role for creating the basis for geospatial digital twins; we explain the feasibility of machine learning (ML) and deep learning (DL) approaches for 3D point clouds. In particular, we argue that 3D point clouds can be seen as a corpus with similar properties as natural language corpora and formulate a “Naturalness Hypothesis” for 3D point clouds. In the main part, we introduce a workflow for interpreting 3D point clouds based on ML/DL approaches that derive domain-specific and application-specific semantics for 3D point clouds without having to create explicit spatial 3D models or explicit rule sets. Finally, examples are shown how ML/DL enables us to efficiently build and maintain base data for geospatial digital twins such as virtual 3D city models, indoor models, or building information models. © 2020, The Author(s)." "56099899900;56943040200;57219173203;57219185242;7404338756;","Data-driven fault diagnosis of industrial robots with a cloud computing framework",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091572258&partnerID=40&md5=6d54aee9433a5ea1e1e2d844806ea760","In modern manufacturing industry, industrial robots have been widely used. Health status of the robots should be always monitored to prevent a sudden shutdown of manufacturing lines. Supervising the signals measured from the industrial robot and diagnosing the status of machines in real time are essential tasks for us to manage the manufacturing lines. In this work, we developed a system for data-driven fault diagnosis of industrial robots, which includes a data acquisition and mining process, a machining learning process, and a cloud computing framework. The signals gathered from attached sensors on the robot are stored in a database within the framework. Structured data are extracted from the stored raw signals. The most important features are selected from the structured data for preventing the overfitting problems in the machine learning process. The fault diagnosis models are trained based on several machine learning algorithms and selected features. Finally, the fault diagnosis results are monitored by operators using mobile devices in real time. All monitoring and diagnosing processes including signal processing, feature extraction, feature selection, and diagnosis operate in the server of our cloud computing framework. Copyright © Proceedings of the 20th International Conference of the European Society for Precision Engineering and Nanotechnology, EUSPEN 2020. All rights reserved." "57211095441;57217271715;57213186458;57218919599;57193959504;55817600100;","Impact of Satellite Sounding Data on Virtual Visible Imagery Generation Using Conditional Generative Adversarial Network",2020,"10.1109/JSTARS.2020.3013598","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090798696&doi=10.1109%2fJSTARS.2020.3013598&partnerID=40&md5=bea9a3e9113e0d6aa18712e58991c4ab","The visible band of satellite sensors is of limited use during the night due to a lack of solar reflection. This study presents an improved conditional generative adversarial networks (CGANs) model to generate virtual nighttime visible imagery using infrared (IR) multiband satellite observations and the brightness temperature difference between the two IR bands in the communication, ocean, and meteorological satellite. For the summer daytime case study with visible band imagery, our multiband CGAN model showed better statistical results [correlation coefficient (CC) = 0.952, bias = -1.752 (in a digital number (DN) unit from 0 to 255, converted from reflectance from 0 to 1), and root-mean-square-error (RMSE) = 26.851 DN] than the single-band CGAN model using a pair of visible and IR bands (CC = 0.916, bias = -4.073 DN, and RMSE = 35.349 DN). The proposed multiband CGAN model performed better than the single-band CGAN model, particularly, in convective clouds and typhoons, because of the sounding effects from the water vapor band. In addition, our multiband CGAN model provided detailed patterns for clouds and typhoons at twilight. Therefore, our results could be used for visible-based nighttime weather analysis of convective clouds and typhoons, using data from next-generation geostationary meteorological satellites. © 2008-2012 IEEE." "57211499439;35849722200;55914518400;57198885698;","CDL: A Cloud Detection Algorithm over Land for MWHS-2 Based on the Gradient Boosting Decision Tree",2020,"10.1109/JSTARS.2020.3014136","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090762033&doi=10.1109%2fJSTARS.2020.3014136&partnerID=40&md5=e7a77ccbba5ff99cee7251d3e72a49d5","This article presents a standalone cloud detection algorithm over the land (CDL) for microwave humidity sounder-2 (MWHS-2), which is characterized by the first operational satellite sensor measuring 118.75 GHz. The CDL is based on the advanced machine learning algorithm gradient boosting decision tree, which achieves the state-of-the-art performance on tabular data with high accuracy, fast training speed, great generalization ability, and weight factor ranking of predictors (or features). Given that the new-generation weather radar of China (CINRAD) provides improved cloud information with extensive temporal-spatial coverage, the observations from CINRAD are used to train the algorithm in this study. There are four groups of radiometric information employed to evaluate the CDL: all frequency ranges from MWHS-2 (all-algorithm), the humidity channels near 183.31 GHz (hum-algorithm), the temperature channels near 118.75 GHz (tem-algorithm), and the window channels at 89 and 150 GHz (win-algorithm). It is revealed that the tem-algorithm (around 118.75 GHz) has a superior performance for CDL along with the optimal values of most evaluation metrics. Although the all-algorithm uses all available frequencies, it shows inferior ability for CDL. Followed are the win-algorithm and hum-algorithm, and the win-algorithm performs better. The analysis also indicates that the latitude, zenith angle, and the azimuth are the top-ranking features for all four algorithms. The presented algorithm CDL can be applied in the quality control processes of assimilating microwave radiances or in the retrieval of atmospheric and surface parameters for cloud filtering. © 2008-2012 IEEE." "56119224700;57208747547;57202362638;24502843200;6602345158;57191895403;","Road infrastructure heritage: From scan to infrabim",2020,"10.37394/232015.2020.16.65","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090685608&doi=10.37394%2f232015.2020.16.65&partnerID=40&md5=dfe11c30022e85640748b806687a30cb","-BIM is a methodology applied to the realization of design models applied to new buildings. To date, however, most of the building interventions as it happens in the field of cultural heritage are developed in the existing. For this reason, scan to BIM procedures are improved and improved every day to make the use of BIM easier. This document will describe the combination of different geomatics techniques used by the Geomatics Laboratory of the University of the Mediterranean in Reggio Calabria to create a Building Information Model of a highway viaduct (infraBIM). In particular, we paid more attention to the scan to BIM phase through the segmentation of the point cloud using machine learning techniques that allow to obtain the constitutive parametric elements of the 3D model. The model containing the geometric and physical data made available by the ANAS management body in order to use the potential of infraBIM. This methodology today is of particular importance for the control, monitoring, intervention, and maintenance of road infrastructures, optimizing the procedures existing up to now. The advantages would be even more evident considering that we are living in a particular historical moment, in which a large number of bridges and viaducts in our nation are subject to advanced forms of degradation. © 2020, World Scientific and Engineering Academy and Society. All rights reserved." "57214044022;","An Optimized Approach-Based Machine Learning to Mitigate DDoS Attack in Cloud Computing",2020,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090649698&partnerID=40&md5=e4cccfbbfa4c2a779dbb7fe78b120ca4","With the emerging growth of cloud computing technology and on-demand services, users can access cloud services and software freely and applications based on the ""pay-as-you go"" concept. This innovation reduced service costs and made them cheaper with high reliability. One of the most significant characteristics of the cloud concept is on-demand services. One can access the applications of cloud computing at any time at a much lower cost. In addition to providing cloud users with much-needed services, the cloud also gets rid of security concerns which are not tolerated by the cloud. One of the most security problems in the cloud environment is Distributed Denial of Service (DDoS) attack that are responsible for overloading the cloud servers. This paper highlights a prevention technique (CS-ANN) which detect the DDoS attack and makes the server side more sensitive by integrating a Cuckoo Search (CS) approach with the Artificial Neural Network (ANN) approach. The cloud user features, along with the attacker features, are optimized using CS as a nature-inspired approach. Later on, these optimized features are passed to the ANN structure. The trained features are stored in the database and used during testing process to match the test features with the trained features and hence provide results in terms of attacker and normal cloud users. The test results of CS-ANN show a True Positive Rate (TPR), False Positive Rate (FPR) and detection accuracy of 0.99, 0.0105 and 0.9865% respectively. The proposed approach outperforms in contrast to the other two state-of-the-art techniques. © International Research Publication House." "57192543260;57204349215;57191667621;56249968800;","Tree detection and health monitoring in multispectral aerial imagery and photogrammetric pointclouds using machine learning",2020,"10.1109/JSTARS.2020.2995391","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85090504711&doi=10.1109%2fJSTARS.2020.2995391&partnerID=40&md5=ca5486ecadcff129dc5cc2bbb6bf3b02","A machine learning methodology is developed for the detection of individual trees, classification of health, and detection of dead/dying trees in 125 mm resolution aerial multispectral or-thoimagery and photogrammetric pointcloud data. The novelty of the proposed method lies in its flexible utilization of features from different georegistered data sources. The methodology is evaluated on a commercial Pinus radiata plantation with a known outbreak of Sirex noctilio. An analysis was carried out to determine the value of the different sensors/data sources and the influence of the spatial resolution of the data on performance. Using the proposed methodology, trees were detected with approximate commission and omission errors of 5% and 22%, respectively, whilst dead or dying trees could be detected with commission and omission errors of approximately 26% and 9%. It was found that the multispectral imagery was the most informative sourceofdata for the given tasks. Tree detection in general was found to be sensitive to the spatial resolution of the data, whilst diseased tree detection was found to be more robust. © 2020 Institute of Electrical and Electronics Engineers. All rights reserved." "57218501642;55904509900;","Research on a Single-Tree Point Cloud Segmentation Method Based on UAV Tilt Photography and Deep Learning Algorithm",2020,"10.1109/JSTARS.2020.3008918","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85089341117&doi=10.1109%2fJSTARS.2020.3008918&partnerID=40&md5=e7089d787531c317c0a1ab8650df4465","Developing a robust point cloud segmentation algorithm for individual trees from an amount of point cloud data has great significance for tracking tree changes. This method can measure the size, growth, and mortality of individual trees to track and understand forest carbon storage and variation. Traditional measurement methods are not only slow but also tardy. In order to obtain forest information better and faster, this article focuses on two aspects: The first is using UAVs to obtain multiview remote sensing images of the forest, and then using the structure from motion algorithm to construct the forest sparse point cloud and patch-based MVS algorithm to construct the dense point cloud. The second is that a targeted point cloud deep learning method is proposed to extract the point cloud of a single tree. The research results show that the accuracy of single-tree point cloud segmentation of deep learning methods is more than 90%, and the accuracy is far better than traditional planar image segmentation and point cloud segmentation. The combination of point cloud data acquisition with UAV remote sensing and point cloud deep learning algorithms can meet the needs of forestry surveys. It is undeniable that this method, as a forestry survey tool, has a large space for promotion and possible future development. © 2008-2012 IEEE." "57212311690;16023225600;55339456900;57208524615;36052693100;","Deep convolutional neural networks for forensic age estimation: A review",2020,"10.1007/978-3-030-35746-7_17","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85085219163&doi=10.1007%2f978-3-030-35746-7_17&partnerID=40&md5=ad7f1f3d077118448629a22db4548bc8","Forensic age estimation is usually requested by courts, but applications can go beyond the legal requirement to enforce policies or offer age-sensitive services. Various biological features such as the face, bones, skeletal and dental structures can be utilised to estimate age. This article will cover how modern technology has developed to provide new methods and algorithms to digitalise this process for the medical community and beyond. The scientific study of Machine Learning (ML) have introduced statistical models without relying on explicit instructions, instead, these models rely on patterns and inference. Furthermore, the large-scale availability of relevant data (medical images) and computational power facilitated by the availability of powerful Graphics Processing Units (GPUs) and Cloud Computing services have accelerated this transformation in age estimation. Magnetic Resonant Imaging (MRI) and X-ray are examples of imaging techniques used to document bones and dental structures with attention to detail making them suitable for age estimation. We discuss how Convolutional Neural Network (CNN) can be used for this purpose and the advantage of using deep CNNs over traditional methods. The article also aims to evaluate various databases and algorithms used for age estimation using facial images and dental images. © 2020, Springer Nature Switzerland AG." "57208754607;6701824774;12753037900;6506881892;57209202081;57214155278;55905144900;57215408871;","Assessing the changes in the moisture/dryness of water cavity surfaces in imlili sebkha in southwestern morocco by using machine learning classification in google earth engine",2020,"10.3390/rs12010131","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078350488&doi=10.3390%2frs12010131&partnerID=40&md5=cfe4979da56272e8541c6ff05961c792","Imlili Sebkha is a stable and flat depression in southern Morocco that is more than 10 km long and almost 3 km wide. This region is mainly sandy, but its northern part holds permanent water pockets that contain fauna and flora despite their hypersaline water. Google Earth Engine (GEE) has revolutionized land monitoring analysis by allowing the use of satellite imagery and other datasets via cloud computing technology and server-side JavaScript programming. This work highlights the potential application of GEE in processing large amounts of satellite Earth Observation (EO) Big Data for the free, long-term, and wide spatio-temporal wet/dry permanent salt water cavities and moisture monitoring of Imlili Sebkha. Optical and radar images were used to understand the functions of Imlili Sebkha in discovering underground hydrological networks. The main objective of this work was to investigate and evaluate the complementarity of optical Landsat, Sentinel-2 data, and Sentinel-1 radar data in such a desert environment. Results show that radar images are not only well suited in studying desertic areas but also in mapping the water cavities in desert wetland zones. The sensitivity of these images to the variations in the slope of the topographic surface facilitated the geological and geomorphological analyses of desert zones and helped reveal the hydrological functions of Imlili Sebkha in discovering buried underground networks. © 2020 by the authors." "7004111660;56606212400;6602954401;7004167798;","Improvement of the 24 hr forecast of surface UV radiation using an ensemble approach",2020,"10.1002/met.1865","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076739442&doi=10.1002%2fmet.1865&partnerID=40&md5=5d5025679b0f9f5e9130e2a9cd3a91d8","A methodology is proposed to improve the 24 hr forecast of the ultraviolet (UV) index and the duration of exposure to obtain the minimal erythemal dose (MED). A forecast ensemble consisting of 10 members (differing in initial and boundary conditions) is examined to search for the best performed ensemble member. Routine UV measurements are used for the forecast validation. These are carried out at Belsk (20.8 ° E, 51.8 ° N) and in Racibórz (18.2 ° E, 50.1 ° N) representing a rural and an urban site in Poland, respectively. Each ensemble member is built using the clear-sky simulations by a radiative transfer model. The clear-sky irradiance is attenuated using the cloud modification factor (CMF) depending on the cloud cover by low- and mid-level clouds. The 24 hr forecast of cloudiness is obtained by the Weather Research and Forecasting (WRF) model. Every day, for each ensemble member, the optimal CMF values are built by the offline bootstrapping of the original CMF matrix. The performance of all ensemble members is evaluated for the day preceding the forecast. The best one is subsequently used for the next-day forecast. This procedure provides a more accurate forecast than that based on a single member of the ensemble. For both sites, the root mean square percentage error for the duration of the MED exposure changes from about 30% to about 15%, and mean absolute percentage error from about 20–25% to about 10%. © 2019 The Authors. Meteorological Applications published by John Wiley & Sons Ltd on behalf of the Royal Meteorological Society." "13204812100;57217112643;","Information technology for sustainable agriculture",2019,"10.1007/978-3-030-23169-9_19","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086296686&doi=10.1007%2f978-3-030-23169-9_19&partnerID=40&md5=3f0dd5e12c09862fd7e165f5d856f1a9","Feeding global population by 2050 requires 60% increase in the agricultural production. Agricultural transformation has a role to play for food security, poverty reduction and economic growth. However, sustainability and climate risk management are the challenges. The recent advancements in information technology (I. T) delivered smart devices, computing and sensor technologies. Application of those smart technologies have the potential to enable agricultural industry meets its productivity and sustainability challenge as well as solving indigenous agricultural problems of the developing nations. The geospatial data archives and real-time data from satellites, UAVs, RFIDs in combination with weather data, digitized soil data, and other real-time data streams coming from in-situ smart sensors can now give us a better understanding of the interaction between crops, weather and soils than ever before. Further the analytics of that big data assisted by machine learning can provide decision support in this regard. The customized I. T packages are required where, e-farm production system based on precision agriculture techniques, crops and livestock management, precision irrigation applications, crop water and pest/disease management, wireless moisture sensing networks, wireless communication in UAVs used for vegetation health detection, rainfall monitoring system based on mobile communication data, cloud services for knowledgebase on soils, nutrients, yields by making soil, nutrient and yield maps and disseminating through mobile networks and variable rate application based on GPS and GIS systems. Every passing day, the use of internet and smartphones is enhancing rapidly. The cloud-based services for big data analytics in agriculture and data sharing apps with linkages to integrated platforms and models are the future of farming in both modern and developing world. © Springer Nature Switzerland AG 2019." "57216256526;7404843459;57216256773;","Super-Resolution Algorithm of Satellite Cloud Image Based on WGAN-GP",2019,"10.1109/ICMO49322.2019.9026112","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082963009&doi=10.1109%2fICMO49322.2019.9026112&partnerID=40&md5=50fe0e227441937edf12f5ac4aa39e44","The resolution of an image is an important indicator for measuring image quality. The higher the resolution, the more detailed information is contained in the image, which is more conducive to subsequent image analysis and other tasks. Improving the resolution of images has always been the unremitting pursuit of industry and academia. In the past, people used hardware devices to increase the resolution, which is a practical solution. However, there are many limitations in the method of improving the image resolution by hardware devices. We use software-based image super-resolution technology, which transforms low-resolution images into high-resolution images through a series of machine learning algorithms. The classic GAN algorithm is difficult to train a model, and the improved Wasserstein GAN algorithm can make the model training more stable. Based on SRGAN model, this algorithm replaces the classical GAN algorithm with the improved WGAN algorithm. We will use the FY-3D satellite's Medium Resolution Spectral Imager Type II (MERSI-II) data, using super-resolution algorithms to make the reconstructed image significantly better visually. We conducted four sets of controlled experiments using four different improved methods. We will evaluate the image from three aspects: peak signal to noise ratio value, structural similarity value and visual effect. We applied the WGAN-GP algorithm to super-resolution tasks and achieved the desired results. © 2019 IEEE." "57216260024;57214130629;57216259010;55688898000;57213519323;","The Cloud Images Classification Based on Convolutional Neural Network",2019,"10.1109/ICMO49322.2019.9026121","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082932298&doi=10.1109%2fICMO49322.2019.9026121&partnerID=40&md5=17324f5a5ed39740b08a4e67012cd8cd","In order to achieve the goal of clouds images classification, a deep-learning method based on Convolutional Neural Network (CNN) is adopted.2300 raw true-color images of eight types of clouds are collected as the datasets from the website www.baidu.com, along with the corresponding labels. Before feeding to machine learning, some preprocessing steps are conducted. Firstly, every photo is preprocessed using the method of the sliding window that can crop the initial photo into a series of sub-images with 200∗200 pixels and RGB channels. Not only can this step increase the size of datasets, but it also can make more features reserved from inputs. Next, the image histogram equalization is adopted. Besides, In the CNN network construction, the ReLU activation function is placed directly after the convolutional layer when there isn't a BN layer behind it, while the ReLU function is placed after the BN layer if the BN layer exists. This kind of structure is designed to avoid the overfitting problem. The result obviously performs better when adopts more large and comprehensive datasets than that when datasets are unhandled. Meanwhile, owing to the MBSD gradient algorithm, the adjustment of arguments and layers can influence the final results as well. The classification accuracy has reached a comparatively high value, which can nearly meet our requirements. © 2019 IEEE." "57212174713;","Large-scale Ship Fault Data Retrieval Algorithm Supporting Complex Query in Cloud Computing",2019,"10.2112/SI97-034.1","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85076112150&doi=10.2112%2fSI97-034.1&partnerID=40&md5=a24e6b408b5b717495918fc637c758cf","In the cloud computing environment, the mass ship fault data retrieval is easy to be interfered by the association rule items, the fuzzy clustering of the data retrieval is not good, the fault diagnosis efficiency of the ship is reduced, and in order to improve the fault diagnosis capability of the ship, The invention provides a mass ship fault data retrieval technology based on complex query support in a cloud computing environment. The distributed storage structure analysis of mass ship fault data is carried out by adopting a vector quantization characteristic coding technology, the spectral characteristic analysis of the mass ship fault data is carried out by adopting a subsection adaptive regression analysis method, the quantitative recursive analysis model is used for extracting the mass ship fault data, the method comprises the following steps of: extracting an association rule feature set reflecting the attribute of a mass ship fault data category, carrying out data classification retrieval on the extracted mass ship fault data feature quantity by using a BP neural network classifier, introducing a machine learning factor to perform convergence control on a support vector machine, And the global stability of the mass ship fault data retrieval is improved. The simulation results show that the accuracy of the data retrieval is high, the error rate is small, and the fault diagnosis ability of the ship is improved. © Coastal Education and Research Foundation, Inc. 2019." "57216288692;57216289852;57204629821;15060531500;","Non-invasive blood glucose monitoring using near-infrared spectroscopy based on internet of things using machine learning",2019,"10.1109/R10-HTC47129.2019.9042479","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083028663&doi=10.1109%2fR10-HTC47129.2019.9042479&partnerID=40&md5=d73eb77ca11d08344de0b59b254038c8","According to International Diabetes Federation, Indonesia ranked 6th in country with highest number of diabetes patients in the world. Now, the measurement of blood glucose in Indonesia mostly done in invasive manner, which is expensive, painful, and impractical. Moreover, according to Indonesia's Ministry of Health, there are more diabetes patient in rural areas, who have economic shortages and limited access to health care. So, a cheap and easy to use non-invasive blood glucose measurement system is needed to solve this problem. One of the current trend non-invasive blood glucose is the use of Near-Infrared Spectroscopy (NIR). In this paper, we discuss the development of a NIR based blood glucose monitor. The device's sensor consists of a pair of LED and photodiode which transmit and receive light with wavelength of 940 nm. The light intensity reading from the sensor will be amplified and filtered to reduce noise, then transmitted to the smart-phone. In the smart-phone application, the reading result will be converted to blood glucose level using a machine learning model embedded in the application. The model used in this final project is a sequential, layer-based neural network model provided by Keras. The model was built and trained on top of TensorFlow, and then converted for mobile use with the help of TensorFlow Lite, The model achieved an acceptable result with Mean Absolute Error (MAE) of 5.855 mg/dL. The result then could be stored in the online database using Cloud Firestore. © 2019 IEEE." "57203835698;57203835732;34969310900;","Short-Term Prediction of Electricity Outages Caused by Convective Storms",2019,"10.1109/TGRS.2019.2921809","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074498250&doi=10.1109%2fTGRS.2019.2921809&partnerID=40&md5=7bfda0ba98fbb093ee6fe885c45832d8","Prediction of power outages caused by convective storms, which are highly localized in space and time, is of crucial importance to power grid operators. We propose a new machine learning approach to predict the damage caused by storms. This approach hinges identifying and tracking of storm cells using weather radar images on the application of machine learning techniques. Overall prediction process consists of identifying storm cells from CAPPI weather radar images by contouring them with a solid 35-dBZ threshold, predicting a track of storm cells, and classifying them based on their damage potential to power grid operators. Tracked storm cells are then classified by combining data obtained from weather radar, ground weather observations, and lightning detectors. We compare random forest classifiers and deep neural networks as alternative methods to classify storm cells. The main challenge is that the training data are heavily imbalanced, as extreme weather events are rare. © 1980-2012 IEEE." "57197462965;56825747200;54784958200;55646959300;","Design and Implementation of Digital Twin for Predicting Failures in Automobiles Using Machine Learning Algorithms",2019,"10.4271/2019-28-0159","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074758484&doi=10.4271%2f2019-28-0159&partnerID=40&md5=328fbd7149521ccfb1133a2ed3345c64","The drastic technological advancements in the field of autonomous vehicles and connected cars lead to substantial progression in the commercial values of automobile industries. However, these advancements force the Original Equipment Manufacturers (OEMs) to shift from feedback-based reactive business analysis to operational-data based predictive analysis thereby enhancing both the customer satisfaction as well as business opportunities. The operational data is nothing but the parameters obtained from several parts of an automobile during its operation such as, temperature in radiator, viscosity of the engine oil and force applied over the brake disk. These operational data are gathered using several sensors implanted in different parts of an automobile and are continuously transmitted to backend computers to develop Digital Twin, which is a virtual model of the physical automobile. Later, gathered operational data are analyzed using data mining algorithms to predict the failures of an automobile well in advance, better insights into performance of an automobile thereby recommending alternative design choices and remote service management of failures by a professional technician. Firstly, this research work illustrates the platform for the creation of digital twin using Eclipse Hono, Eclipse Kura and Eclipse Ditto. Secondly, it explains about the operational data gathering and processing at the nearby edge devices as well as the remote cloud. Finally, the prediction of failures is demonstrated using Turbofan Engine Degradation Simulation Dataset by means of several machine learning regression algorithms and compare their accuracies. Finally, it is concluded that Gradient Boost Regressor provides better accuracy in predicting future failures. © 2019 SAE International. All Rights Reserved." "57217665158;56041758500;","Machine learning strategies for enhancing bathymetry extraction from imbalanced lidar point clouds",2019,"10.23919/OCEANS40490.2019.8962400","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079077294&doi=10.23919%2fOCEANS40490.2019.8962400&partnerID=40&md5=d6271c8cca635611b3559ba5b13d9f98","Density-based approaches to extract bathymetry from airborne lidar point clouds generally rely on histogram/frequency-based disambiguation rules to separate noise from signal. The present work targets the improvement of such disambiguation rules by enhancing each pulse with a machine learning-based estimate of its p(Bathy) - i.e., its probability of truly being bathymetry. Extreme gradient boosting (XGB) is used to assess the strength of bathymetric signal in pulse return metadata. Because lidar point clouds can be highly imbalanced between Bathymetry and NotBathymetry, three strategies for mitigating the effects of imbalanced samples were examined. Impacts of an imbalanced lidar point cloud were successfully mitigated by: •Applying an 'optimal' decision threshold (ODT) that equalizes accuracy for Bathymetry and NotBathymetry to p(Bathy) rather than using a conventional probability decision threshold (PDT) of 0.50, and •Using proportional class weighting to fit the XGB model. However, decomposing a confusion matrix by iteratively discarding misclassified points and re-fitting an XGB model was not successful in improving the strength or detectability of the bathymetric signal in the metadata. The same was true for iteratively discarding correctly classified points. The bathymetric signal in the metadata was found to be sufficiently strong to explore the operational incorporation of results into the disambiguation rules of density-based bathymertric extraction methods. © 2019 Marine Technology Society." "57211910620;57211908810;","AI & CoE",2019,"10.35940/ijeat.A9846.109119","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075295961&doi=10.35940%2fijeat.A9846.109119&partnerID=40&md5=314d5856d171ad0701babe1765f8b3fc","Artificial Intelligence or AI is being positioned as the panacea for all organizational problems; while Centre of Excellence or CoE, which is distinctly different from Research and Development activities helps organizations in their pursuit of higher revenue and profit. In this paper, the researchers have analysed the growth and importance of each of these two concepts–AI and the CoE; and have worked towards putting them together for creating a unique combination which shall benefit the organizations and hence the economy at large. In the process, a framework is provided for companies to improve, innovate, optimize, and eventually automate their management systems while making the core-competencies of their business AI proof. It is hoped that through this framework, organizations will be able to create a substantial impact by improving existing capabilities and actively creating new strategic resources in the interest of all the stakeholders. © BEIESP." "57211621229;56910124600;","PM2.5 prediction using machine learning hybrid model for smart health",2019,"10.35940/ijeat.A1187.109119","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074658732&doi=10.35940%2fijeat.A1187.109119&partnerID=40&md5=175dcd98b3f99d08082557214863bda2","Air Pollution is one of the current serious issue attributable to people's health causing cardiopulmonary deaths, lung cancer and several respiratory problems. Air is polluted by numerous air pollutants, among which Particulate Matter (PM2.5) is considered harmful consists of suspended particles with a diameter less than 2.5 micrometers.This paper aims to acquire PM2.5 data through IoT devices,store it in Cloud and propose an improved hybrid model that predicts the PM2.5 concentration in the air. Finally through forecasting system we alert the public in case of an undesired condition. The experimental result shows that our proposed hybrid model achieve better performance than other regression models. © BEIESP." "57215597635;57203397955;57216355489;56722821200;57201399639;7101627008;","A hybrid algorithm for mineral dust detection using satellite data",2019,"10.1109/eScience.2019.00012","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083233958&doi=10.1109%2feScience.2019.00012&partnerID=40&md5=bbf101231a6b3d1365a70fcf2a134609","Mineral dust, defined as aerosol originating from the soil, can have various harmful effects to the environment and human health. The detection of dust, and particularly incoming dust storms, may help prevent some of these negative impacts. In this paper, using satellite observations from Moderate Resolution Imaging Spectroradiometer (MODIS) and the Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observation Observation (CALIPSO), we compared several machine learning algorithms to traditional physical models and evaluated their performance regarding mineral dust detection. Based on the comparison results, we proposed a hybrid algorithm to integrate physical model with the data mining model, which achieved the best accuracy result among all the methods. Further, we identified the ranking of different channels of MODIS data based on the importance of the band wavelengths in dust detection. Our model also showed the quantitative relationships between the dust and the different band wavelengths. © 2019 IEEE." "57188690965;57203079997;57194625045;56046460900;","Estimation of -CG lightning distances using single-station E-field measurements and machine learning techniques",2019,"10.1109/SIPDA47030.2019.9004484","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85080936194&doi=10.1109%2fSIPDA47030.2019.9004484&partnerID=40&md5=f763a6eb28b3220890ac10c75c7b34eb","Machine learning (ML) techniques have been used around the world to solve different problems. In this work, we applied ML techniques to estimate lightning distance for return strokes (RS) in negative cloud-to-ground (-CG) flashes. The approach uses E-field records from a single-station system. A lightning electric field waveform dataset containing more than 1500 waveforms of negative RS recorded at LOG, Florida, US was used to train and validate the ML classifiers. The dataset was split into day time and night time records. For day-time records, the quadratic Support Vector Machine (SVM) classifier was the one with the best accuracy (80%) and for night-time, the best one was the linear SVM with an accuracy of 88%. The ML classifiers were applied to estimate lightning distance in thunderstorms in the Amazon region of Brazil, and the results were compared against GOES-16 images and STARNET lightning location data. The main application of such methodology is for regions with no lightning location systems or no communication links to obtain lightning location data. © 2019 IEEE." "57195197915;56271680000;","Machine learning algorithm for early detection of heart diseases using 3-tier IOT architecture",2019,"10.35940/ijeat.F1360.0986S319","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075458510&doi=10.35940%2fijeat.F1360.0986S319&partnerID=40&md5=83d384750db497fdc187418d9cf3c375","Among the applications empowered by the Internet of Things (IoT), regular health monitoring framework is an important one. Wearable sensor gadgets utilized in IoT health monitoring framework have been producing huge amount of data on regular basis. The speed of data generation by IoT sensor gadgets is very high. Henceforth, the volume of data generated from the IoT-based health monitoring framework is also very high. So as to overcome this problem, this paper proposes adaptable three-tier architecture to store and process such immense volume of wearable sensor data. Tier 1 focuses on gathering of data from IoT wearable sensor gadgets. Tier 2 employs Apache HBase for storing substantial volume of wearable IoT sensor data in cloud computing. Likewise, Tier-3 utilizes Apache Mahout for building up logistic regression-based prediction model for heart related issues. At long last, ROC examination is performed to identify the most significant clinical parameters to get heart diseases. © BEIESP." "57202077645;55663177800;37012907500;36988104100;","Cloud-based non-intrusive load monitoring system (NILM)",2019,"10.35940/ijeat.F1021.0986S319","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075453130&doi=10.35940%2fijeat.F1021.0986S319&partnerID=40&md5=d0c920bccedb92ef7ad74ca5c73cdbc0","Design and development of a cloud-based non-intrusive load monitoring System (NILM) is presented. It serves for monitoring and disaggregating the aggregated data such as smart metering into appliance-level load information by using cloud computing and machine learning algorithms implemented in cloud. The existing NILM systems are lack of scalability and limited in computing resources (computation and data storage) due to dedicated, closed and proprietary-based characteristics. They are inaccessible to variety of heterogeneity data (electrical and non-electrical data) openly for improving NILM performance. Therefore, this paper proposed a novel cloud-based NILM system to enable collection of these open data for load monitoring and other energy-related services. The collected data such as smart meter or data acquisition unit (DAQ), is pre-processed and uploaded to the cloud platform. A classifier algorithm based on Artificial Neural Network (ANN) is implemented in Azure ML Studio (AzureML), followed by the classifier testing with different combinations of feature set for the performance comparison. Furthermore, a web service is deployed for web APIs (Application Programming Interfaces) of applications such as smart grid and smart cities. The results shows that the ANN classifier for multiclass classification has improved performance with additional features of harmonics, apart from active and reactive powers used. It also demonstrates the feasibility of proposed cloud-based classifier model for load monitoring. Therefore, the proposed solution offers a convenient and cost-effective way of load monitoring via cloud computing technology for smart grid and smart home applications. Further work includes the use of other ML algorithms for classifier, performance analysis, development of cloud-based universal appliance data and use cases. © BEIESP." "57211938438;","5G-smart diabetes: Personalized diabetes diagnosis with human services big data clouds",2019,"10.35940/ijeat.F1335.0986S319","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85075445463&doi=10.35940%2fijeat.F1335.0986S319&partnerID=40&md5=5f900f2f6c5db1acecc2965507dbc6aa","Ongoing advances in remote systems administration and huge information innovations, for example, 5G systems, medicinal huge information investigation, and the Internet of Things, alongside ongoing improvements in wearable figuring and man-made brainpower, are empowering the advancement and usage of imaginative diabetes checking frameworks and applications. Because of the deep rooted and efficient damage endured by diabetes patients, it is basic to plan powerful strategies for the determination and treatment of diabetes. In light of our far reaching examination, this paper characterizes those strategies into Diabetes 1.0 and Diabetes 2.0, which show insufficiencies as far as systems administration and insight. Consequently, our objective is to structure a manageable, financially savvy, and insightful diabetes finding arrangement with customized treatment. In this paper, we initially propose the 5G-Smart Diabetes framework, which joins the best in class advancements, for example, wearable 2.0, machine learning, and huge information to create complete detecting and investigation for patients experiencing diabetes. At that point we present the information sharing system and customized information examination display for 5G-Smart Diabetes. At last, we construct a 5G-Smart Diabetes testbed that incorporates savvy dress, cell phone, and huge information mists. The trial results demonstrate that our framework can successfully give customized analysis and treatment proposals to patients. © BEIESP." "55982189400;55639123200;","Research on inefficiency analysis method of building energy utilizing time series data",2019,"10.1088/1755-1315/294/1/012052","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071868044&doi=10.1088%2f1755-1315%2f294%2f1%2f012052&partnerID=40&md5=c347553e2fd7a298028bccc18c067908","Purpose; Our research purpose is to utilize time series data related to building energy and perform an inefficiency analysis of building energy efficiency. Development; We developed a cloud building energy and environment monitoring system (cloud Building Management System (BEMS)) and an optimized control system (an artificial intelligence [AI] control system). Methodology; First, we designed an inefficiency analysis model and the analysis steps. Next, this model was tested in several actual buildings, and the method was confirmed to be adaptable. Finally, through iterative adaptations based on the results, we significantly improved our method. Findings; (1) The developed cloud BEMS and AI control system were confirmed to be useful for a) improvement of building management, b) improvement of equipment and system, and c) renovation. (2) Our analytical method and its steps can quantify inefficiency. (3) When using this methodology, we can quantitatively predict before implementation the effect of a) building improvement, b) equipment and system improvement, and c) renovation. Originality/value; Cloud BEMS collects time series data related to building energy consumption from sensors and building automation systems and accumulates data in a cloud server. The AI control system and analytical method finds energy inefficiencies and improves the operation of the building. Moreover, this improvement in operation is unmanned and carried out automatically. Keywords; Sustainable, Internet of things (IoT), machine learning, artificial intelligence (AI), energy management system (EMS), energy conservation, environment efficiency, optimized control, CO2 reduction. Paper type; Academic paper. © Published under licence by IOP Publishing Ltd." "57211289630;57203689843;","A systematic access through machine learning methods for expectation in malady related qualities",2019,"10.35940/ijeat.F1193.0886S19","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073348095&doi=10.35940%2fijeat.F1193.0886S19&partnerID=40&md5=7177f48073f74d611d246edbf05383ae","There are Many learning strategies that are been identified with distinguish infection based related qualities. At the early, they as a rule moved toward this issue as a parallel arrangement, where preparing set is involved examples. Examples developed sickness qualities, whereas negative examples are there mining which are not known to be connected with contaminations. This is the essential of the twofold deals based diagrams, since the negative arranging set ought to be true non-infection qualities; regardless, advancement of this set is on a very basic level unfeasible in biomedical inspects. Therefore, to reduce this delicacy, insightfully sensible social gathering based techniques have been proposed. For example, unary outline strategy subject to one-class SVM framework was proposed by grabbing from fundamentally positive models. Also, there mining set may contain cloud torment qualities; as such, semi-formed methodologies, for example, twofold semi-controlled & positive & unlabeled (PU) learning blueprints have been proposed. Specifically, PU learning frameworks, which increase from both known suffering qualities & there mining attributes, were appeared to beat others. In these examinations, information sources are commonly tended to by vectorial plan for cemented classifiers, while they are in bit frameworks for unary & PU learning ones. The bit based information blend is reasonable for information with various sorts & it has the majority of the stores of being uncalled for or the relationship subject to various information diagrams. In like manner, in this examination, we looked accumulating structures for the ailment quality figure dependent on vectorial delineation of tests. The spread outcome demonstrated that the unary strategy structure, which joins both thickness & class likelihood estimation approachs, accomplished the best execution, where as it is most recognizably stunning for the one-class SVM-based technique. fascinatingly, execution of a best twofold outline framework is in each rational sense misty with that of uneven SVM-based PU learning & twofold semidirected hoarding strategy they are altogether improved. © BEIESP." "57205578378;57211300508;","IOT and data research in industrial power management",2019,"10.35940/ijeat.F1011.0886S19","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85073334195&doi=10.35940%2fijeat.F1011.0886S19&partnerID=40&md5=27fc7500d246f55dc33e1006c34f4f30","IOT plays an important role in collecting data and machine learning for prediction in variety of applications like homecare, healthcare and energy management. In energy management there are various variables such as future power demands, generation status weather conditions and current battery status hard to expect high efficiency. Here, in this proposed idea, for higher efficiency of renewable energy, an IOT system is needed to monitor and collect these Statuses and provide energy management services. Energy will be consumed of passive operation according to hourly variation in price and battery status will be predicted by using machine learning algorithms like Logistic regression, SVM, and k-NN. We trained the system by considering five random variables in datasheet such as Current time, Current cost, predicted time, predicted cost and Solar battery status from the device. This integrated system is used for uploading power related details of Grid and Solar to IBM cloud. Depending on previous datasheet, analytics will be done by resulting which source has to be triggered to drive the load either Solar or Grid. APIs and Node-Red Tool were used for wiring sensor data and Model predicted output. In future power demands, this design will help to predict the price according to hourly variation based on the units and to trigger the source. © BEIESP." "57210971634;56771498800;57210972983;","Detection of intrusion using hybrid feature selection and flexible rule based machine learning",2019,"10.35940/ijeat.F8783.088619","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072062467&doi=10.35940%2fijeat.F8783.088619&partnerID=40&md5=11c21ef35407df6c7b020615abdab1d6","With the rapid growth in the data processing and data sharing, the application owners and the consumers of the applications are more influenced to use the remote storage on cloud-based data centre and the application generated data is also growing ups and bounds. Nevertheless, the adaptation of the data sharing, and data processing applications were not easy for the consumers. The application owners and the service providers have struggled with the sensitive data of the consumers and the consumers were also faced trust issues with the complete framework. The standard legacy applications were designed for the traditional centralized scenarios, where the intrusion detection can be performed only using the network status analysis and the application characteristics analysis. Moreover, most of the parallel calculations initially enhance the hybrid likelihood and change likelihood of GA as indicated by the populace advancement variable-based math and wellness esteem. Nevertheless, the population of data and the attacks on the data is high and the correct population size is highly difficult to determine. Regardless to mention, that the use of fitness functions will restrict the attack detection to certain types and these algorithms are bound to fail in case of a newer attack. However, with the migration of application to the data processing framework, the consumers have started demanding more security against the intrusions. A good number of research attempts were made to map the traditional security algorithms into the data processing space, nonetheless, the attempts were highly criticized due to the lack of proper analysis of security attacks on data processing applications. Hence, this work proposes a novel framework to detect the intrusions on data processing framework with justifying attack characteristics. This work proposes a novel algorithm to reduce the features of attack characteristics to justify the gaps on data processing frameworks with significant reduction in time for processing and further, proposes an algorithm to derive a strong rule engine to analyse the attack characteristics for detecting newer attacks. The complete proposed framework demonstrates nearly 93% and higher accuracy, which is much higher than the existing parallel research outcomes with least time complexity. © BEIESP." "57190491796;56252036100;","Decision tree: A predictive modeling tool used in cloud trust prediction",2019,"10.35940/ijeat.F8763.088619","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072045101&doi=10.35940%2fijeat.F8763.088619&partnerID=40&md5=5fe4f597066479a6aee3adf32caa475f","Trust is one of the important challenges faced by the cloud industry. Ever increasing data theft cases are contributing in worsening this issue. Regarding trust, author has a perception that this challenge can be handled to some extend if consumer can evaluate “Trust Value “ of the provider or can predict the same on some reliable basis. Current research is using predictive modeling for predicting trustworthiness of cloud provider. This paper is an attempt to utilize the data mining algorithm for predictive modeling. Decision Tree, a supervised data mining algorithm has been used in the current work for making predictions. Certification attainment criteria as prime basis for trust evaluation. In current scenario, data mining algorithm will classify providers in category of low, medium and high category of trust on the basis of information displayed on the public domain. ©BEIESP." "57210957409;56040721700;57210959142;","Upgrade-data security in cloud by machine learning and cryptography techniques",2019,"10.35940/ijeat.F8395.088619","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072019154&doi=10.35940%2fijeat.F8395.088619&partnerID=40&md5=8cd96d23d16ce5f9c3a71b563cff0441","The term cloud computing is referred as the shared pool of customizable computer resources and high quantity services which can easily be provisioned with less management endeavours via internet. It transfigured the mode associations reach IT, which enables them to be more perceptive, launch new business models, and minimise the IT costs. These technologies are to be administrated in an interdisciplinary collection of architectures, characterized into various deployment and service models, and can synchronize with other related technologies. The widespread issues with cloud computing are security, reliability, data privacy and anonymity. Cloud computing provides a way to share distributed sources and services that are owned by different organizations or sites. Since it shares distributed resources via network in open environment that results in security issues. In this paper, our aim is to upgrade the security of data in the cloud and also to annihilate the difficulties related to the data security with encipher algorithm. In our proposed plan, some key services of security like authentication and cryptographic techniques are assigned in cloud computing environment. © BEIESP." "55430538200;8529553300;55908790400;36170340400;36806957800;57210434487;57210436461;","Microservices and machine learning algorithms for adaptive green buildings",2019,"10.3390/su11164320","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070712595&doi=10.3390%2fsu11164320&partnerID=40&md5=547ebabe27f2f4e288d0a8b65e03a6b6","In recent years, the use of services for Open Systems development has consolidated and strengthened. Advances in the Service Science and Engineering (SSE) community, promoted by the reinforcement ofWeb Services and SemanticWeb technologies and the presence of new Cloud computing techniques, such as the proliferation of microservices solutions, have allowed software architects to experiment and develop new ways of building open and adaptable computer systems at runtime. Home automation, intelligent buildings, robotics, graphical user interfaces are some of the social atmosphere environments suitable in which to apply certain innovative trends. This paper presents a schema for the adaptation of Dynamic Computer Systems (DCS) using interdisciplinary techniques on model-driven engineering, service engineering and soft computing. The proposal manages an orchestrated microservices schema for adapting component-based software architectural systems at runtime. This schema has been developed as a three-layer adaptive transformation process that is supported on a rule-based decision-making service implemented by means of Machine Learning (ML) algorithms. The experimental development was implemented in the Solar Energy Research Center (CIESOL) applying the proposed microservices schema for adapting home architectural atmosphere systems on Green Buildings. © 2019 by the authors." "57205651985;57193516410;8502218200;57195735997;6507631512;55802422900;35595209900;","Evaluation of Cloud Type Classification Based on Split Window Algorithm Using Himawari-8 Satellite Data",2019,"10.1109/IGARSS.2019.8898451","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077721443&doi=10.1109%2fIGARSS.2019.8898451&partnerID=40&md5=f04c62e5f9884e5bdc96b3080796574f","Precise evaluation of cloud types is indispensable for the detailed analysis of the Earth's radiation budget. The split window algorithm (SWA) is an algorithm that has been widely employed for cloud type classification from meteorological satellite imagery. In this study, we apply the SWA to analyze the clouds that appear in the Japan area using the imagery of Himawari-8 meteorological satellite. The brightness temperature (BT) information from band 13 (BT13, 10 μm) and band 15 (BT15, 12 μm) are employed with the BT difference (BTD) between these two bands (BTD13-15). For daytime analysis, the albedo of band 1 (0.47 μm) is also used to discriminate the cloudy and cloud-free areas. The validation of the resulting cloud type (SWA13-15), which includes ten classes including cloud-free condition, is carried out using the space-borne lidar data concurrent with the satellite observations. In addition, two different classifiers, namely, the sequential minimal optimization (SMO) and Naïve Bayes (NB) classifiers are tested with the results of SWA. When about 10% of 2 million data points are used for training the classifiers, the test results reveal that the correctly classified points are 97.0% and 89.5% for the first dataset (observed in July 2015) and 97.4%, and 92.1% for the second dataset (July 2016) for SMO and NB, respectively. © 2019 IEEE." "55918225100;57189026667;55706555700;","Cloud Detection and Classification for S-NPP FSR CRIS Data Using Supervised Machine Learning",2019,"10.1109/IGARSS.2019.8898876","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077694896&doi=10.1109%2fIGARSS.2019.8898876&partnerID=40&md5=d85a9995d1c16c9c0fb6f069d14350b3","Cloud detection and classification are important and challenging in satellite infrared remote sensing. This study proposes an end-to-end cloud detection and classification method for the Cross-track Infrared Sounder (CrIS) based on big data analysis and relevant machine learning schemes. Using the full spectral resolution (FSR) CrIS data, a set of FSR cloud detection indexes (FCDIs) is derived from the brightness temperatures of the selected CrIS LWIR-SWIR channel pairs. Linear and cubic regressions are compared and discussed when deriving FCDI. It's shown that FCDIs can capture clouds and structures well by comparing with the Visible Infrared Imaging Radiometer Suite (VIIRS) cloud cover/layer product. Selected from several supervised machine learning schemes, the extreme learning machine (ELM) is applied to train FCDI for cloud detection. Overall the ELM classification accuracy is about 80%. Results and discussion are provided. © 2019 IEEE." "57213189957;56049862900;57191270639;57213199230;57207365988;57213192204;","KORE Application: Potatoes Yield Assessment",2019,"10.1109/IGARSS.2019.8898996","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077673883&doi=10.1109%2fIGARSS.2019.8898996&partnerID=40&md5=6fab62b0e05f347ddab595b7c4adf558","KORE is a cloud-based precision agriculture platform, based on subscriptions, that combines satellites, drones and ground sensor data. This paper focuses on one of its applications: yield assessment in potato fields using UAV data and deep learning. © 2019 IEEE." "35176903900;57211186337;8976516300;7006684098;56913686000;55946457200;","Study on temporal and spatial adaptability of crop classification models",2019,"10.1109/Agro-Geoinformatics.2019.8820233","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85072912373&doi=10.1109%2fAgro-Geoinformatics.2019.8820233&partnerID=40&md5=b1383142e1c93c480ed146be76746c97","Crop classification is an important part of national agricultural management, and accurate crop classification is conducive to crop growth monitoring and yield assessment. However, due to the different growing years and regions, even the same crop has different growth processes and different phenological characteristics. Therefore, improving the spatial and temporal adaptability of the classification model is an important research content for large-scale crop classification. In this paper, several adjacent agricultural production areas are studied. Based on the stable time-series remote sensing image dataset, the adaptive changes of several machine learning classification methods with higher classification accuracy in spatial and temporal are studied. The paper selected Sentinel 1 satellite data with good anti-cloud interference and a short return visit cycle for experiments. Firstly, the training of each classification model in the same area is completed, and then the spatial adaptability of the model is studied in different adjacent ranges. Finally, the adaptability of different classification models to the change of the growth cycle of the same type of crop is also compared. The paper finds that the models such as CNN+LSTM and BinConvLSTM perform better in temporal and spatial. © 2019 IEEE." "57211353806;57206774247;56669065400;57211351208;35750657400;57211347521;57211353913;","Sharing learnings: The methodology, optimisation and benefits of moving subsurface data to the public cloud",2019,"10.3997/2214-4609.201900773","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088067716&doi=10.3997%2f2214-4609.201900773&partnerID=40&md5=1c5035aa0abfba12108487d076eeda22","We detail the application of cloud technology to Chevron's seismic data repository, share the applicable learnings and highlight areas of workflow evolution that delivered value. The aim of the project, run as a piloted field trial, was to couple the transformative capability of public cloud services with subsurface data. The learnings shared in this paper are designed to inform and assist other subsurface data custodians whether they are energy companies, national data repositories, service companies or academia. Cloud technology enables us to reduce seismic data duplication to a single data version which can be accessed securely by all appropriate stakeholders. We discuss a highly automated, scalable data migration process which included seismic ingestion, machine learning and serverless architecture. This automated the data management process, progressing data from loading to global analysis in minutes. The project has provided access to subsurface data anytime, anywhere and on any device, delivering a more accessible data environment at lower costs and connecting via an API to traditional workflows. By approaching the subsurface data challenge in an innovative way, this project has provided multiple learnings to share and built a greater understanding of the value case for faster adoption of public cloud infrastructure. © 81st EAGE Conference and Exhibition 2019. All rights reserved." "57215345521;24766116400;10044906200;57215343685;57219734114;","EarthNET a native cloud web based solution for next generation subsurface workflows",2019,"10.3997/2214-4609.201901974","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084021587&doi=10.3997%2f2214-4609.201901974&partnerID=40&md5=80dcde21cc5cd666339b46ec447515fd","Faster and better data driven decision-making and shorter times to first oil and gas top the list of expected benefits that digital technologies can drive for upstream oil and gas companies. In the oil industry, Artificial Intelligence (AI) and Machine Learning (ML) tools have already moved from R&D projects into G&G tool boxes, slowly transforming the subsurface workflow. We will discuss about cloud platforms and demonstrate how such an integrated platforms provides both the data access and applications required to apply ML at scale with examples that include integration of multi regional datasets. We will show that such platforms are not only enhancing further creativity and enabling data driven decisions but in addition will shorten time to oil which seems to be the next challenge of the industry. © 81st EAGE Conference and Exhibition 2019 Workshop Programme. All rights reserved." "56610006900;57202441572;57210218564;14052416300;","RoofN3D: A database for 3d building reconstruction with deep learning",2019,"10.14358/PERS.85.6.435","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069962419&doi=10.14358%2fPERS.85.6.435&partnerID=40&md5=1232a616325793faf535573bd423c446","Machine learning methods, in particular those based on deep learning, have gained in importance through the latest development of artificial intelligence and computer hardware. However, the direct application of deep learning methods to improve the results of 3D building reconstruction is often not possible due, for example, to the lack of suitable training data. To address this issue, we present RoofN3D which provides a three-dimensional (3D) point cloud training dataset that can be used to train machine learning models for different tasks in the context of 3D building reconstruction. The details about RoofN3D and the developed framework to automatically derive such training data are described in this paper. Furthermore, we provide an overview of other available 3D point cloud training data and approaches from current literature in which solutions for the application of deep learning to 3D point cloud data are presented. Finally, we exemplarily demonstrate how the provided data can be used to classify building roofs with the PointNet framework. © 2019 American Society for Photogrammetry and Remote Sensing." "57210202528;57210205399;","Continuous authentication using smart health monitoring system",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85069929615&partnerID=40&md5=a6bb79c399bf3e9dd808737bab68d1d3","The main security issue in most of the computer is user identification and authentication. In traditional authentication schemes, the user is only validated during initial login which provides low security for the system. Hence continuous authentication has to be done to resolve the security issue. With increase in the number of smartphone users, continuous validation of the authenticated user is important. Continuous authentication mechanism can be made using two behavioral traits: app usage and touch based. There is a worldwide increase in the usage of apps that works on touchscreen, hence both can be used for authentication. Hence, the continuous authentication will be based on app usage and touch screen based which provide high security. In this paper, smart health monitoring system is used for continuous authentication. The data which is collected from wearable biomedical sensors for continuous health monitoring can also be used for continuous authentication. Although the biomedical signals are not highly discriminative a robust machine learning to obtain high accuracy levels is used. An android app is developed to gather data and send to cloud for data storage. The user is validated based on the decisions from the classifiers. The proposed work does not need any extra model for data collection as it uses the data gathered for health monitoring purpose, it can be used for low cost applications. © BEIESP." [No author id available],"2018 IEEE Region 10 Symposium, Tensymp 2018",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065105334&partnerID=40&md5=00c5515dd02d1e2a858fd90f4374dfe8","The proceedings contain 58 papers. The topics discussed include: optimal overcurrent relay coordination of a multi-loop distribution network with distributed generation using dual simplex algorithm; a customized system to assess foot plantar pressure: a case study on calloused and normal feet; distinguishing between cyber injection and faults using machine learning algorithms; computing the relations among three views based on artificial neural network; on-chip transient detection circuit for microelectronic systems against electrical transient disturbances due to ESD events; sparsity representation of beat signal in weather radar for compressive sampling; development of a compact-sized biodigester for pig manure and organic wastes with raspberry pi-based temperature, pressure, and pH level monitoring; an effective hexapod robot control design based on a fuzzy neural network and a kalman filter; potential of IoT system and cloud services for predicting agricultural pests and diseases; and performance evaluation of ideal nearest replica routing (NRR) against several forwarding strategies on named data networking(NDN)." "57204728675;57206315647;","Overcoming challenges in connected autonomous vehicles development: Open source vehicular data analytics platform",2019,"10.4271/2019-01-1083","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064682024&doi=10.4271%2f2019-01-1083&partnerID=40&md5=32d2d20489e2058beb09f81fe00dcf5f","Data Science and Machine Learning are buzzwords in our everyday lives, as is evident from its applications, such as voice recognition features in vehicles and on cell phones, automatic facial and traffic sign recognition, etc. Analyzing big data on the basis of internet searches, pattern recognition, and learning algorithms, provides deep understanding of the behaviour of processes, systems, nature, and ultimately the people. The already implementable idea of autonomous driving is nearly a reality for many drivers today with the aid of ""lane keeping assistance"" and ""adaptive cruise control systems"" in the vehicle. The drift towards connected, autonomous, and artificially intelligent systems that constantly learns from big data and is able to make best-suited decisions, is progressing in ways that is fundamental to the growth of automotive industry. The paper envisages the future of connected-and-autonomous-vehicles (CAVs) as computers-on-wheels. These are pictured as sophisticated systems with sensors on board as data sources and a lot of other functions and running services to support autonomous-driving. These services are considered to be computationally expensive. The unit performing on-board has limited computing resources while on the other hand, the cloud-based architecture has unconstrained resources, but it suffers from extended unexpected latency that requires large-scale internet data transfer. To deal with this conflicting scenario, Open-Vehicular-Data-Analytics-Platform (Open VDAP) for CAVs may be used. This allows CAVs to detect dynamically status of each service, computation-overhead and the optimal-offloading-destination such that each service could be finished within an acceptable-latency. Open VDAP is an open-source platform that offers free APIs and real-field vehicle data to researchers and developers in the community, allowing them to deploy and evaluate applications in real environment. © 2019 SAE International. All Rights Reserved." "56461352400;57210411252;57210414069;57210418335;","Prediction and detection of heart attack using machine learning and internet of things",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070610136&partnerID=40&md5=c328e679e2016e0c987aa1dc9e36fe39","In today’s modern world Cardiovascular infections are the basic reason for death worldwide in the course of the most recent couple of years in the developed as well as developing countries. Early recognition of heart diseases and persistent supervision of clinicians can lessen the death rate. Be that as it may, the exact discovery of heart diseases and meeting of a patient for 24 hours by a specialist isn’t accessible since it requires more insightfulness, time and aptitude. In this examination, a speculative plan of a cloud-based coronary illness identification and forecast the framework had been proposed to identify coronary illness utilizing Machine learning strategies. For the exact recognition of the coronary illness, a productive AI procedure ought to be utilized and we utilize numerous relapses for the expectation of heart attack illness. We use this algorithm to predict heart attack by taking different independent variables and we take pulse beat time to time as it varies from time to time. We use multiple regression to predict heart attack and we use IOT to communicate to the person and we use IOT devices and cloud platform in order to remind the person about his health condition of a heart attack. Besides, to screen the coronary illness tolerant a constant patient checking framework was created and displayed utilizing Arduino, able to do taking a parameter like a heartbeat rate utilizing beat sensor IOT device. The created framework can transmit the recorded information to a server which is updated at regular intervals. In this paper, I clarified the engineering for pulse or heartbeat rate and other information observing system and I likewise disclosed how to utilize an AI calculation like MULTIPLE REGRESSION calculations to foresee the heart attack by utilizing the gathered pulse information and another wellbeing related edge and how we use IOT devices for the location of a heart attack. © BEIESP." "57209471021;57209463826;57209472902;57209474569;","A machine learning algorithm for jitter reduction and video quality enhancement in IoT environment",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85067872532&partnerID=40&md5=fbf10404bb62e7f9b12ebc5387ea4c1a","Multimedia traffic has been abnormally increasing nowadays due to its greater usage and necessity. CCTV cameras (closed circuit television camera) are widely used in these days as it a matter of security concern. From shopping malls to home the usage CCTV camera plays a vital role. But the challenging part arises when the media data captured by the camera is to be transmitted to the display monitors of the owner. There are scenarios where more than one CCTV camera covers a particular region. The 70% of network traffic is caused due to CCTV surveillance. It is important to reduce the traffic and delay packets to deliver the data to the user on time. We have used wireless SDN network (software defined network) to transfer the multimedia data to the display monitor. The SDN control switch is integrated with AI module where machine learning algorithm is implemented (BAT algorithm) to prioritise the data packets. For more efficient storage data is uploaded to the IOT cloud. © BEIESP." "55665406600;","Advanced IT-based Future Sustainable Computing (2017-2018)",2019,"10.3390/su11082264","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85066877682&doi=10.3390%2fsu11082264&partnerID=40&md5=df013124a3524621672aaf1c1caa9799","Future Sustainability Computing (FSC) is an emerging concept that holds various types of paradigms, rules, procedures, and policies to support breadth and length of the deployment of Information Technology (IT) for abundant life. However, advanced IT-based FCS is facing several sustainability problems in different information processing and computing environments. Solutions to these problems can call upon various computational and algorithmic frameworks that employ optimization, integration, generation, and utilization technique within cloud, mobile, and cluster computing, such as meta-heuristics, decision support systems, prediction and control, dynamical systems, machine learning, and so on. Therefore, this special issue deals with various software and hardware design, novel architectures and frameworks, specific mathematical models, and efficient modeling-simulation for advance IT-based FCS. We accepted eighteen articles in the six different IT dimensions: machine learning, blockchain, optimized resource provision, communication network, IT governance, and information security. All accepted articles contribute to the applications and research in the FCS, such as software and information processing, cloud storage organization, smart devices, efficient algorithmic information processing and distribution. © 2019 by the authors." "57207992885;57207981392;57207982328;57193866840;57193770709;","Leading Edge Assembly Real Time Process Monitoring Using Industrial Internet of Things (IIoT)",2019,"10.4271/2019-01-1367","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063433788&doi=10.4271%2f2019-01-1367&partnerID=40&md5=00e610f6f0409f45874cd5092949940c","The increasing global demand for commercial aircraft creates many new challenges in manufacturing including an increased need to maximize the automation of manufacturing processes. The purpose of this research is to develop the understanding of leading edge assembly processes using robot mounted tooling and an automated fixture with advanced process monitoring. Within this research real-time process monitoring data is acquired from an assembly operation and processed into an open cloud environment enabling advanced data analytics. Implementation of advanced analytics utilising process data could be developed for the use of machine learning algorithms which can lead to superior fault finding. The aim of this research is to improve product quality, reduce cost and increase process knowledge, enabling the potential for maximized online and offline process feedback. This paper details how an open Industrial Internet of Things (IIoT) environment can be used to interface with a number of proprietary devices to enable real-time process monitoring in aerospace manufacturing. The implementation of the software and hardware is detailed and followed by an initial evaluation of the system architecture by performing leading edge assembly operations. In addition, a baseline of current IIoT systems is discussed, with the comparison to the specific architecture used here. The system is based on a wireless integration of proprietary devices and sensors feeding real-time data to an open cloud environment. Data is analysed and visualised in real time with online access and report generation. Data is supplied from the automated fixture and Restricted Access Drilling Unit (RADU) end-effector prototype developed by the Manufacturing Technology Centre. Supplied process data includes hardware and system health, and environmental measurements during the assembly operation. Access to analysed data grants the ability to identify occurring abnormalities during the assembly process which, ultimately, will allow for advancements in increased component quality and reduce manufacturing costs. © 2019 SAE International. All Rights Reserved." "57204728675;","Autonomous Connected Aerial Vehicles: Overview of Application of Open Source Vehicular Data Analytics Platform to Urban Air Taxi",2019,"10.4271/2019-01-1351","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063426508&doi=10.4271%2f2019-01-1351&partnerID=40&md5=9130b4c2d13e91c9cda41df96eb3a12d","Data Science and Machine Learning are playing a key role in our everyday lives, as is evident from its applications, such as voice recognition features in vehicles and on cell phones, automatic facial and traffic sign recognition etc. Analysing big data on the basis of searches, pattern recognition and learning algorithms provides deep understanding of the behaviour of processes, systems, nature, and ultimately the people. The already implementable idea of autonomous driving is nearly a reality for many drivers today with the help of lane keeping assistance and adaptive cruise control systems in the vehicle. The drift towards connected, autonomous, and artificially intelligent systems that constantly learns from the big data and is able to make best suited decisions is advancing in ways that are fundamentally important to many automotive industries. The paper envisages the future of connected-autonomous-urban-airtaxi (CUTs) as computers on wheels. These are pictured as sophisticated systems with on-board sensors as data sources and a lot of other functions and services running to support autonomous aerial driving/flying. These services are considered to be computationally expensive. The on-board computation unit has limited computing resources. On the other hand, the cloud-based architecture has unconstrained resources but it suffers from unexpected extended latency that leads to the large-scale Internet data transmission. To deal with this dilemma, Open Vehicular Data Analytics Platform (OpenVDAP) for CUTs may be used. This allows CUTs to dynamically detect status of each service, computation overhead and the optimal offloading destination so that each service could be finished within an acceptable latency. OpenVDAP is an open-source platform that offers free APIs and real-field aerial vehicle data to the researchers and developers in the community, allowing them to deploy and evaluate applications on the real environment. © 2019 SAE International. All Rights Reserved." "57200918295;24072139200;57207450851;56545781200;8632957700;","The utilisation of cloud computing and remote sensing approach to assess environmental sustainability in Malaysia",2019,"10.1088/1755-1315/230/1/012109","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85062512275&doi=10.1088%2f1755-1315%2f230%2f1%2f012109&partnerID=40&md5=d41eaa12c6a1016de50e502fce2b0f30","Monitoring of the environment over a large area will require huge amount of data. The implementation of conventional methods will be time consuming and very costly. Furthermore, one way to assess the environmental sustainability is by analysing the changes of land cover over two different periods. Therefore, this study implemented a cloud computing approach utilising an open source Remote Ecosystem Monitoring Assessment Pipeline (REMAP) to map the changes of Land use land cover (LULC) over Peninsular Malaysia utilising Landsat data obtained from two different periods (2003 and 2017). This approach has utilised a powerful inbuilt machine learning algorithm, Random Forest (RF) to test the performance of cloud computing using REMAP to produce LULC maps over Peninsular Malaysia. The results showed an acceptable LULC maps and the changes between two periods were analysed. Therefore, the utilisation of machine learning algorithm with the integration of cloud computing using REMAP can reduce the cost, lessen the processing time, produce LULC maps, and perform change analysis over large area. © 2019 Institute of Physics Publishing. All rights reserved." "57209835052;57073683700;56586072300;","An analysis of user oriented behaviour based malware variants detection system",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85068856096&partnerID=40&md5=79edbcdf8bfbca7093791a8bb0d64339","A virtualized infrastructure (VI) is implemented through one or more virtual machines that depend on built-in software defined multiple instances of hosting hardware. This infrastructure model added an advantage of gathering different computing resources and on-demand resource scaling that facilitated extensive deployment of VI to cloud computing services. BDSA finds the potential attack and protect in VI opposing vulnerabilities. HDFS is used to store the backend information. Security analytical algorithm applied on logs captured at various points within network to identify the attach existence. Ref[1]Graph based event correlation and Map Reduce Parser methodologies are used to identify the attack paths through the network logs obtained.Ref[1] Two step machine learning is used to determine the attack presence, attack’s conditional probability based on attributes is calculated through logistic regression and existence of attack on network is calculated through belief propagation. This has steered way for cyber attackers to launch attacks for illegal access on virtualized infrastructures. © BEIESP." "57204927880;57205443317;57205445803;","Methodology to Recognize Vehicle Loading Condition - An Indirect Method Using Telematics and Machine Learning",2019,"10.4271/2019-26-0019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85060051881&doi=10.4271%2f2019-26-0019&partnerID=40&md5=f2b0edb50cf4b2952b03c9420d71b8a2","Connected vehicles technology is experiencing a boom across the globe. Vehicle manufacturers have started using telematics devices which leverage mobile connectivity to pool the data. Though the primary purpose of the telematics devices is location tracking, the additional vehicle information gathered through the devices can bring in much more insights about the vehicles and its working condition. Cloud computing is one of the major enabled for connected vehicles and its data-driven solutions. On the other hand, machine learning and data analytics enable a rich customer experience understanding different inferences from the available data. From a fleet owner perspective, the revenue and the maintenance costs are directly related to the usage conditions of the vehicle. Usage information like load condition could help in efficient vehicle planning, drive mode selection and proactive maintenance [1]. A common approach to vehicle load condition detection is by using exclusive load sensing sensors. This paper explores a possibility of detecting vehicle load conditions without making use of any sensors. Instead, a supervised machine learning model is developed to recognize real-time loading condition, by analyzing vehicle driving behavior. This paper covers a machine learning based approach for load detection of small commercial vehicles, which are less then 1Ton of loading capacity. In this study, the focus is given to differentiating the vehicle behavior at different loading conditions and to select the accurate parameters for machine learning model development. These selected features are based on the domain expertise in vehicle dynamics and statistics of the data. The output of this novel method can be used for optimizing different ADAS functionalities [2] at very low-cost, leveraging telematics units. © 2019 SAE International. All Rights Reserved." "36158741000;57203041761;57215132765;","Drainage strategy optimization - Making better decisions under uncertainty",2019,"10.2118/196683-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088773612&doi=10.2118%2f196683-ms&partnerID=40&md5=4c332c53cbffda633bbd17287434e192","Improved reservoir knowledge is key to extracting additional value from existing oil and gas assets. However, given the uncertainty in the subsurface, it is always a question if our current development strategy is the most robust choice, or if there are alternatives that can further increase the value of our field. This paper presents a novel solution that enables the asset team to answer these questions in a new way. Furthermore, the solution helps teams quickly identify and screen new opportunities that ultimately increase both subsurface understanding and the value of the field. The solution combines a quasi-Newton gradient based numerical optimization scheme with a stochastic simplex approximate gradient (StoSAG) algorithm. Because the algorithm is non-intrusive with respect to the fluid flow simulator, we can directly apply the solution on any flow optimization problem without the need to access the simulator source code. The solution is implemented using a microservice architecture that allows for efficient scaling and deployment either on cloud-based or internal systems. We demonstrate the proposed solution on a field containing 11 oil producers and 7 water injectors by optimizing the water injection and oil production rates. The machine learning algorithm allows us to quickly explore different drainage strategies, given the current understanding and associated uncertainties of the reservoir. Specifically, the software solution suggests that 6 of the 18 pre-defined well targets are high risk and/or of little value. Running a second development scenario where we do not drill these six wells reduces the investment cost of this field by 163 MUSD and increases the expected net present value per well of the field by 48 percent. Compared with the reactive control drainage strategy approach, we increase the expected net present value of the field by 9.0 %, while simultaneously lowering the associated risk. Copyright 2019, Society of Petroleum Engineers." "57215883934;55356184400;","Pipeline corrosion failure probability assessment from qualitative inspection leveraging artificial intelligence",2019,"10.2118/197484-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088771496&doi=10.2118%2f197484-ms&partnerID=40&md5=0e35767a5ab9ec0da127560e483cc55d","Assessing asset corrosion related failure risks without involving expensive testing, field trials, and intensive data mining followed by complex computational analyses is a significant challenge in Oil & Gas industry. More reliably, such assessment is still governed by practical experience and fundamental conception of the main corrosion mechanisms, though often with the significant advancement of corrosion failure risk assessment and computer technology. To that end the practicality of artificial intelligence and machine learning (AIML) techniques leveraging the characteristics inspection data to predict corrosion failure mimicking human practical experience were investigated. This paper attempts to merge the two complex areas of corrosion engineering and computer aided analyses for the direct beneficiary of the fast-growing discipline of corrosion and integrity management (C&IM). Over the past few years, the authors from their respective areas of expertise have combined forces to learn about each other's world and help make tangible progress and resolve common problems in pipeline engineering. In this paper, qualitative risks due to external and internal corrosion of pipeline were studied and assessed using six different artificially intelligent enabled machine learning algorithms leveraging the existing PHMSA Hazardous Liquid Accident's report. Relative importance insights or governing mechanisms that results in corrosion related failure were identified utilizing several AIMC algorithms. Where relevant and possible semi quantitative risk assessments have been appraised to perform a 'calibration check' on the findings determined. This is best done by both verification and validation (V&V). The former is usually a desktop or laboratory exercise, and the latter often a field trial. Based on a comparative study using various learning algorithms like Gradient boosting, Deep Learning, Support Vectors, Random Forest, Decision Trees and Linear regression, it is observed that a specific corrosion mechanistic pipe failure can be predicted within 97% accuracy. The process is automated in a cloud-based web application form for prediction from raw characteristics data of accident report, along with a Corrosion SME review of logic and reasonability. High accuracy and reliability demonstrate that use of artificial intelligent would enable decision making engineering and project managers to give a faster and more confident assessment of corrosion risks. © 2019, Society of Petroleum Engineers" "57200247152;57203115730;57203121765;56593035700;7004709202;","Automated pressure transient analysis: A cloud-based approach",2019,"10.2523/iptc-19060-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85088069110&doi=10.2523%2fiptc-19060-ms&partnerID=40&md5=0d29cee62fab00b56bd60b06e9f02037","Pressure transient analysis provides useful information to evaluate injection induced fracture geometry, permeability damage near wellbore and pressure elevation in injection zone. Manual analysis of pressure data after each injection cycle could be subjective and time-consuming. In this study a cloud-based approach to automatically analyze pressure data will be presented, which is aimed to improve the reliability and efficiency of pressure transient analysis. There are two fundamental requirements for the automated pressure transient analysis: 1) Pressure data needs to be automatically retrieved from field sites and fed to analyzer; 2) Analyzer can automatically select instantaneous shut-in pressure (ISIP), identify flow regimes, and determine fracture closure point. To meet these requirements and also take the advantages of cloud storage and computing technologies, a web based application has been developed to pull real time injection data from any field sites and push it to a cloud database. Besides analyzing any existing pressure data in the cloud database, a built-in pressure transient analyzer can also detect any real-time pressure data and perform pressure analysis automatically when required data is available. The automated, cloud-based pressure transient analysis has been applied to multiple injection projects. In general, the analysis results including permeability, fracture half length, skin factor, and fracture closure pressure are comparable to these yielded from manual analysis. The discrepancy is mainly caused by poor data quality. The inconsistent selections of ISIP and different slopes defined for G-function and flow regime analyses also contribute to the divergence. Overall, the automated pressure transient analysis provides consistent results as the exact same criteria are applied to the pressure data, and analysis results are independent on analyzer's experience and knowledge. In addition, machine learning algorithms are applied to continuously refine the criteria and improve the quality of analysis results. As data from oil/gas industry increases exponentially over time, automated data transmission, storage, analysis and access are essential to maximize the value of the data and reduce operation cost. The automated pressure transient analysis presented here demonstrates that cloud storage and computing combined with automated analysis tools is an optimal way to overcome big data challenges facing by oil/gas industry. © 2019, International Petroleum Technology Conference" "57214910519;57201316372;57214915718;","Seismic data management for big data era",2019,"10.2118/197369-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85086058520&doi=10.2118%2f197369-ms&partnerID=40&md5=2ac570dfb21123763485da58b17ca338","Seismic data is one of earliest data acquired in a prospect evaluation and the data are utilized throughout the exploration and production stages of a prospect. With recent advances in the handling of big data, it is essential to re-evaluate the best practices in the seismic data ecosystem. This paper presents the idea to leveragingthe technology advancement in big data and cloud computing for Seismic data ecosystem with the aim to providing an improve user experience. This new seismic platform would be capable of handling, managing and delivering the full spectrum of seismic data varieties starting from acquired field data to interpretation ready processed data. The system to have the following capabilities: • Capability to entitle the right portion of data to every user as per interest • Organization of seismic data as per the business units • Data security by sharing data only with legitimate users/groups. • Direct or indirect integration with all the data sources and applications who are consuming and/ or generating data • Sharing of and collaboration on data within company and/or across organization for shareholding partner, perspective seismic buyer for trading and relinquishment, regulatory agency resource certifying agencies and service providers etc. over limited network connectivity. • Provide intergration/data deliverivey to End Users applications where this seismic data will be utilizaed Implementation of Seismic ecosystem will enable: • Sharing of seismic data by the acquisition, quality control, data processing and interpretation with user communities from one centralized storage • Collaboration of stake holders in real time over an encrypted network • Leveraging cloud and mobility technology advancement for agility and interaction. The system will be connected and interactive yet has the power of complex high-performance computing infrastructure on the background. • Data delivery and auditing to wider and more diverse user community that consumes data from different platforms. • Secure data access based on organizational business units to make sure data does not fall into unauthorized hand. • Reduction in seismic data turnaround time by reading and ingesting large volume of data through parallel input/output operation. • Improved data delivery and map interface with contextual information out of the centralized data store. • Augment traditional workflows with machine learning and artificial intelligence for example automated fault detection, etc., The proposed best practice aims to bring all of the different disciplines working with seismic data to one centralized seismic data repository and enable them to consume and share seismic data from big data lake. This is live and interactive when compared to traditional technologies of using the archive and restore system in standalone application. © 2019, Society of Petroleum Engineers" "57215884104;57215883959;57205425645;57205427894;36696718600;26026419000;57215884723;6602822699;36053538900;","Embracing the digital and artificial intelligence revolution for reservoir management - Intelligent integrated subsurface modelling IISM",2019,"10.2118/197388-ms","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084780776&doi=10.2118%2f197388-ms&partnerID=40&md5=ff7e24e91e17f48a296b22c93f594d6e","Challenges associated with volatile oil and gas prices and an enhanced emphasis on a cleaner energy world are pushing the oil and gas industry to re-consider its fundamental existing business-models and establish a long-term, more sustainable vision for the future. That vision needs to be more competitive, innovative, sustainable and profitable. To move along that path the oil and gas industry must proactively embrace the 4th Industrial Revolution (oil and gas 4.0) across every part of its business. This will help to overcome time constraints in the understanding and utilization of the terabytes of data that have been and are continuously being produced. There is a clear need to streamline and enhance the critical decision-making processes to deliver on key value drivers, reducing the cost per barrel, enabling greater efficiencies, enhanced sustainability and more predictable production. Latest advances in software and hardware technologies enabled by virtually unlimited cloud compute and artificial intelligence (AI) capabilities are used to integrate the different petro-technical disciplines that feed into massive reservoir management programs. The presented work in this paper is the foundation of a future ADNOC digital reservoir management system that can power the business for the next several decades. In order to achieve that goal, we are integrating next generation data management systems, reservoir modeling workflows and AI assisted interpretation systems across all domains through the Intelligent Integrated Subsurface Modelling (IISM) program. The IISM is a multi-stage program, aimed at establishing a synergy between all domains including drilling, petrophysics, geology, geophysics, fluid modeling and reservoir engineering. A continuous feedback loop helps identify and deliver optimum solutions across the entire reservoir characterization and management workflow. The intent is to dramatically reduce the turnaround time, improve accuracy and understanding of the reservoir for better and more timely reservoir management decisions. This would ultimately make the management of the resources more efficient, agile and sustainable. Data-driven machine learning (ML) workflows are currently being built across numerous petro-technical domains to enable quicker data processing, interpretation and insights from both structured and unstructured data. Automated quality controls and cross domain integration are integral to the system. This would ensure a better performance and deliver improvements in safety, efficiency and economics. This paper highlights how applying artificial intelligence, automation and cloud computing to complex reservoir management processes can transform a traditionally slow and disconnected set of processes into a near real time, fully integrated, workflow that can optimize efficiency, safety, performance and drive long term sustainability of the resource. © 2019, Society of Petroleum Engineers" "57205425143;57214913847;57214915820;","Using machine learning-based predictive models to enable preventative maintenance and prevent ESP downtime",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084166530&partnerID=40&md5=4ef62b9d3311b8f58ed52aeabb45c0de","This paper focuses on the use of artificial intelligence (AI) and machine learning (ML) algorithms to implement anomaly detection and shows how this concept can be extended to implement autonomous well surveillance. Today, critical equipment is monitored by implementing automation and control systems with built-in protection logic for safe operation of equipment and for shutdown of the system in the event that operation deviates outside of valid process conditions. These automation and control systems require constant surveillance by a human operator to verify that all processes are running normally. It is the human operator's responsibility to react to any alarm conditions that occur during operation. Often, the alarm trigger event occurs without early notification and the operator has a short period of time to react. This way of controlling operations requires skilled operators with a great deal of experience to monitor and control the system in an effective manner. It also limits the amount of time the operator can allocate toward optimization. Autonomous surveillance is the concept of training an AI system to provide early detection of abnormal behavior. In this way, the system can take over the task of constant surveillance of process operation, leaving the experienced human operator with additional capacity to focus his time on more productive actions. For example, the human operator can use a combination of process data and decision support information from the AI system to consider current operating conditions and implement a more optimized setpoint, which could increase production or extend equipment life expectancy. In this paper, an example is presented, which is based on monitoring electric submersible pumps (ESPs) using a deep learning neural network that was trained on historical data from the process control system historian. In addition to outlining the benefits that AI-assisted surveillance provides when compared to conventional methods, the paper provides a system architecture blueprint for implementing an autonomous monitoring application across different types of ESP fleets by connecting sensor data directly to a cloud-based monitoring system. The paper builds on the work of previous SPE papers by providing up-to-date results of a pilot project, where a predictive maintenance model has been running for 10 months. On the project, 30 ESPs ranging in power from as low as 200-kW to as high as 500-kW were deployed and monitored using an AI-supported predictive maintenance model. To date, the results have been extremely positive. In one case, an ESP interruption was predicted by the application 12 days before the actual failure occurred. Following several months of testing/use of the ML-based predictive maintenance solution during the pilot deployment, the ESP fleet operator concluded that the system can detect multiple kinds of anomalies in advance, even previously unknown ones. This capability is a significant distinction between the AI-based model and conventional models employed by other ESP diagnostic tools. Although these new, unknown types of complex ESP operational anomalies were difficult to interpret as to their root causes, they could still have led to ESP performance degradation and possible failure nonetheless, if not mitigated or remediated. © 2019, Society of Petroleum Engineers" "56135789500;57194177429;10340225800;56136748300;","Traffic-related air pollution monitoring laboratory with computer vision based traffic count and cloud based data visualization techniques",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084019676&partnerID=40&md5=f3a4a474335ff15727f4ccee64edd0bf","Motor vehicle exhaust is a significant source of urban air pollution. The most widely reported pollutants in vehicular exhaust include carbon monoxide, nitrogen oxides, unburned hydrocarbons, particulate matter, polycyclic aromatic hydrocarbons, and other organic compounds. People living or otherwise spending substantial time near busy highways are exposed to these pollutants and it is becoming increasingly important to understand the emission and health effects of road traffic. Many roadside areas have no or limited sustained monitoring capabilities, making it difficult to quantify the major pollution sources. Real-time mobile air monitoring has been demonstrated to have an advantage of identifying spatial and temporal differences of on-road traffic pollutants from different road types, traffic intensities, and road features. The Houston Advanced Research Center (HARC) has developed the Mobile Acquisition of Real-time Concentrations (MARC) laboratory, which consists of a Ford F-350 passenger van outfitted with a Proton Transfer Reaction Mass Spectrometer (PTR-MS), NOx analyzer, CO analyzer, GPS unit, camera, and metrological equipment. Front view camera mounted on the front of the van is used to get the traffic intensities using a machine learning based computer vision algorithm by tracking and counting the number of the vehicle in the vicinity. This allows for the fast measurement of geotagged gaseous pollutant concentrations correlated with traffic intensities. The data recorded from the van is uploaded to an off-site database via an adaptive interface that supports 3G/4G infrastructure and the information is broadcasted to a website in real-time. MARC is used during 2018 the Health Impact Assessment (HIA) of the North Houston Highway Improvement Project (NHHIP) in Houston, TX. The main goal of the study was to quantify and explain the potential co-benefits and co-harms to health associated with the proposed changes to the I-45 corridor from Beltway 8 North to the I-69 spur south of downtown Houston. © 2019 Air and Waste Management Association. All rights reserved." "57211205900;","Remote sensing for leak detection and quantification",2019,"10.2118/195794-MS","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084017163&doi=10.2118%2f195794-MS&partnerID=40&md5=a0c858aa0e2d60ae27a8516f547250e1","Offshore oil and gas installations are (by their nature) located in remote locations that are both difficult and costly to access. While such challenges exist, the operate & maintain requirements associated with such assets are consistent and must be addressed, requiring operators to identify the most efficient form of service to reduce staffing levels, risk and cost. Offshore hydrocarbon production assets commonly incorporate equipment and processes that can lead to significant (fugitive) gas emissions. The consequences are both economic and social (environmental) in nature, requiring operators to perform emissions surveys with the objective of leak identification and remediation within the shortest possible timeframe. The frequency of this activity is naturally limited and must be balanced with the staffing and operating needs of the broader facility, which in-turn can lead to sub-optimal leak detection to fix timing and reliability. Addressing the three key challenges of access productivity, detection reliability and results quantification, Worley has developed a remote sensing platform that incorporates the use of productive remote access equipment such as unmanned aerial vehicles (UAV) and in-situ monitoring, with machine based emissions detection and algorithmic quantification to provide a solution that allows the operator to increase survey frequency, obtain more reliable results at lower cost, and perform the work in a manner consistent with safe and low-risk operations. In both testing and field deployments, the results have provided for significant reductions in both false positive and negatives and have produced datasets that allow for accurate indications of greenhouse gas reduction via comparison of volumetric emissions before and after leak repair activity has taken place. The technology is largely mathematical, utilizing coded routines for machine learning to perform gas detection under (initially) supervised modeling conditions, and algorithmic gas dispersion models for further emission quantification. The performance of the survey is typically carried out through the integration of existing, proven manufactured sensing equipment across several types of UAV or in-situ monitors which collect field data for transmission to a cloud-based portal which further processes the results. The approach has been shown effective in accessing hard or costly to reach areas, improving survey productivities, while the data processing and quantification allows the operator to benefit from improved measurability and prioritize leak repair accordingly. Copyright 2019, Society of Petroleum Engineers" "57211206961;57211206666;","Machine learning image analysis for asset inspection",2019,"10.2118/195773-MS","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85084013834&doi=10.2118%2f195773-MS&partnerID=40&md5=ce3a7d7b71241f6488d0bf3b9067903e","Total E&P UK Ltd and Merkle have undertaken a proof of concept project to investigate machine learning image analysis applied to image and video data from onshore and offshore sites, in preparation for the deployment of an autonomous asset inspection ground robot in 2019. The aim is to better understand the feasibility of these methods, and to demonstrate the benefits of robotic inspection with regard to improving safety and efficiency, enhancing data capture and reducing operational costs. An object detection model was developed based on high-performance open source algorithms. Transfer learning was applied using a custom-built image library and the result is a model able to detect a range of different types of items in the industrial environment including mobile equipment, process equipment, infrastructure and personnel (with and without PPE). The object detection model is used to feed into object classification anomaly detection models to look at the state of selected pieces of identified equipment, such as whether a valve is open or closed, which can be placed in the context of the expected state of the process equipment by relating it to the digital twin for the asset. An additional object detection model was developed to operate as a gas leak detection system for infrared cameras. The object detection model achieved good results and model performance was driven by the number and quality of images used for the training. An anomaly detection model designed to detect whether ball valves were open or closed delivered good results, with high accuracy and balanced false positive and false negative detection rates. The overall performance of the infrared gas detection model was restricted by the limited volume and variability of the training data, although the false positive detection rate was very low. A significant part of the machine learning was devoted to the development of a consistent labelled image library for oil and gas equipment, infrastructure and gas leaks. Image transformations were tested but boosting the number of images using transforms gave variable results. Additional training and testing data is needed to ensure that the models are as robust as possible, especially for the gas leak detection model. Once the models are productionised and in use, additional data can be used to periodically retrain the models for improved performance. In addition to the machine learning algorithms, a fundamental aspect of the project is the development of the overall technical architecture, supporting the data science. This includes enabling the transfer of data from inspection robots or other connected camera devices into a data store in a cloud computing environment and returning the results to dashboard systems with different levels of detail depending on user requirements. Copyright 2019, Society of Petroleum Engineers" "57191583173;57214793809;57214793588;57214800500;57214797515;","Blocksat: On-demand access to shared-use satellite constellations",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079138023&partnerID=40&md5=94823c1be896751d15b26f031db23e92","As space launch costs drop and the feasibility of “small-satellite” distributed sensing and imaging improves, we note a growing interest in satellite swarms across academic, government, and industry labs. Many corporate proposals already explore satellite constellations as the backbone of a global, space-based Internet service [1], [2]. However, these constellation proposals are usually owned by a single entity or conglomerate, and operated to serve a particular business model. Building on the theme of crowd-sharing models and innovative research in distributed system-control algorithms, we propose a constellation-management system and multi-functional CubeSat hardware platform that would allow communal use of satellite functionality, including opportunities for non-conflicting concurrent processes and re-purposing of satellite hardware. By coupling communication and data ledger protocols (e.g. modified blockchains) with machine learning algorithms for smart task distribution and execution management, this research describes a way to orchestrate peer-to-peer network collaborations where satellite constellation functionality is dynamically rented, shared or reused between many applications. We envision multi-purpose, shared- use satellite constellations, bringing this category of space hardware into the “on-demand” services market (along the lines of Amazon Web Services, but for broad space applications) and moving beyond more narrow applications of Distributed Ledger Technologies (DLTs) currently deployed in orbit [3]. This model for allocation of a shared research resource, where a ledger is kept to distribute access fairly, builds on years of scientific collaboration processes [4] created around expensive, limited-use hardware resources (e.g., multinational collaborations that split time on large telescopes or the CERN particle accelerator [5]). This research proposes to automate and enhance this process, exploring remote-activation and operations of space-based resources. This paper will discuss the BlockSat mission architecture and ConOps (Concept of Operations), showing how the integration of a multi-functional hardware platform and cloud-computing participation model can create a new class of small satellite operations. We aim to free satellite deployments from the limitations of single-use applications and propose a new “open space” market that can democratize access to LEO space technology. Copyright © 2019 by the International Astronautical Federation (IAF). All rights reserved." "57208753024;56448070900;","Implementation of machine learning methods on FPGA for On-board spacecraft operation",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85079124867&partnerID=40&md5=1fec763b3e230fa456042fe41e114983","Machine Learning (ML) techniques are increasingly being used in terrestrial applications. They have also been proposed for various space applications such as on-board data processing, planetary explorations, autonomous operations and various mission-specific applications. Terrestrial application of ML is facilitated by shared resources such as cloud services, powerful desktop computers and framework APIs. However, these enablers are limited in space environment. To take advantage of the progress in ML, space applications need to incorporate on-board inference. The first part of this paper presents a review on some space applications in which such algorithms are suitable or have been used. The review scope is limited to on-board and in-flight applications that have already been realized and doesn't factor in majority of those that are at a proposal stage. ML techniques at ground stations for analysing and processing data is not covered. The review also presents some hardware that has been or can be utilized in these implementations. Though there has been a plethora of hardware platforms geared towards ML inference on the edge, this review notes that Field Programmable Gate Arrays (FPGAs) and associated Systems-on-Chip (SoCs) are more suitable for ML inference in space applications. They also have flight-heritage, having been applied in mission-specific and satellite subsystem operations. They are also reconfigurable and can therefore be adopted for different tasks on the fly. Available models and networks optimized for edge ML inference are presented in this study as well. Such models are more appropriate for space applications since a custom implementation is time-consuming and prone to failure due to tight FPGA fabric timing constraints. Nevertheless, the workflow for a custom Artificial Neural Network (ANN) implementation on an FPGA has been presented. Xilinx Kintex-7 FPGA has been used as the target FPGA in the implementation and evaluation of the network. Vivado IDE and Verilog Hardware Description Language have been used in the demonstration. Copyright © 2019 by the International Astronautical Federation (IAF). All rights reserved." [No author id available],"14th 3D GeoInfo Conference 2019",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078019390&partnerID=40&md5=23b92da0f5e61589ef44364fe466e3ce","The proceedings contain 19 papers. The topics discussed include: generating 3D city models based on the semantic segmentation of lidar data using convolutional neural networks; voxel-based visibility analysis for safety assessment of urban environments; cityreconstruction from airborne lidar: a computational geometry approach; raise the roof: towards generating LOD2 models without aerial surveys using machine learning; a big data approach for comprehensive urban shadow analysis from airborne laser scanning point clouds; and an improved automatic pointwise semantic segmentation of a 3D urban scene from mobile terrestrial and airborne lidar point clouds: a machine learning approach." "36614966600;54930662100;56055664200;55542665400;55893847400;","Active sampling and model based prediction for fast and robust detection and reconstruction of complex roofs in 3D point clouds",2019,"10.5194/isprs-annals-IV-4-W8-43-2019","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077959601&doi=10.5194%2fisprs-annals-IV-4-W8-43-2019&partnerID=40&md5=293352e189996bf8d36129daedcc3405","3D city models in Level-of-Detail 2 (LoD2) are nowadays inevitable for many applications such as solar radiation calculation and energy demand estimation. City-wide models are required which can solely be acquired by fully automatic approaches. In this paper we propose a novel method for the 3D-reconstruction of LoD2 buildings with structured roofs and dormers from LIDAR data. We apply a hybrid strategy which combines the strengths of top-down and bottom-up methods. The main contribution is the introduction of an active sampling strategy which applies a cascade of filters focusing on promising samples in an early stage and avoiding the pitfalls of RANSAC based approaches. Such filters are based on prior knowledge represented by (non-parametric) density distributions. Samples are pairs of surflets, i.e. 3D points together with normal vectors derived from a plane approximation of their neighborhood. Surflet pairs imply immediately important roof parameters such as azimuth, inclination and ridge height, as well as parameters for internal precision and consistency, giving a good base for assessment and ranking. Ranking of samples leads to a small number of promising hypotheses. Model selection is based on predictions for example of ridge positions which can easily be falsified based on the given observations. Our approach does not require building footprints as prerequisite. They are derived in a preprocessing step using machine learning methods, in particular Support Vector Machines (SVM). © Authors 2019. CC BY 4.0 License." "57212882647;54893965400;57212866934;57212877790;","Potential distribution model of leontochir ovallei using remote sensing data [Modelo de distribución potencial de leontochir ovallei con datos de sensores remotos]",2019,"10.4995/raet.2019.12792","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85077460375&doi=10.4995%2fraet.2019.12792&partnerID=40&md5=2457565e13138944a884c1da5e9a4187","Predicting the potential distribution of short-lived species with a narrow natural distribution range is a difficult task, especially when there is limited field data. The possible distribution of L. ovallei was modeled using the maximum entropy approach. This species has a very restricted distribution along the hyperarid coastal desert in northern Chile. Our results showed that local and regional environmental factors define its distribution. Changes in altitude and microhabitat related to the landforms are of critical importance at the local scale, whereas cloud cover variations associated with coastal fog was the principal factor determining the presence of L. ovallei at the regional level. This study verified the value of the maximum entropy in understanding the factors that influence the distribution of plant species with restricted distribution ranges. © 2019, Universitat Politecnica de Valencia. All rights reserved." "57211500142;57201118368;","IoT platforms for the Mining Industry: An overview",2019,"10.29227/IM-2019-01-47","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85074188305&doi=10.29227%2fIM-2019-01-47&partnerID=40&md5=9ff1d7c6c5fa4f5731543fbd9816b59e","Industry 4.0 and the Internet of Things are now very common concepts as solutions that can revolutionize the industry. Constant technological progress increases the possibilities of using computer tools and solutions to support processes in industry and production optimization. The use of the Internet of Things is particularly important in complex processes in mining, enabling the extraction of valuable information from data. The integration of physical facilities in the enterprise enables the digitization of production processes and the increase of efficiency and security. This article presents an overview of the selected internet of things platforms and analytical tools that can be used in industry, with particular emphasis on the mining sector. It is pointed out, that the number of suppliers of IoT technologies and analytical tools offering advanced data analytics services for industry is significant and constantly evolving. The aim of the article is to evaluate selected IoT solutions based on the following criteria: offering predictive analytics, implemented artificial intelligence (AI) or machine learning (ML) algorithms, a mining-oriented process approach, advanced data visualization, interoperability, real-time data capture, remote device management and cloud-based technology. The review was prepared to provide knowledge about IoT vendors operating on the market, as well as to indicate the functionalities that are the most popular among solutions. © 2019 Polish Mineral Engineering Society. All rights reserved." "55927897800;6507214354;35488207000;44061832600;55257235200;6506805816;","Editorial for the Special Issue ""Frontiers in spectral imaging and 3D technologies for geospatial solutions""",2019,"10.3390/rs11141714","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071529952&doi=10.3390%2frs11141714&partnerID=40&md5=eefaea7ac07c55cb355ae2539b1d3e06","This Special Issue hosts papers on the integrated use of spectral imaging and 3D technologies in remote sensing, including novel sensors, evolving machine learning technologies for data analysis, and the utilization of these technologies in a variety of geospatial applications. The presented results showed improved results when multimodal data was used in object analysis. © 2019 by the authors." "36082483100;56132443500;57210283137;","How Cost Effective Is Machine Learning/AI Applied to Leak Detection and Pipe Replacement Prioritization?",2019,"10.1061/9780784482506.029","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070217696&doi=10.1061%2f9780784482506.029&partnerID=40&md5=41a73bd907f914899b006a883d66c0aa","Artificial intelligence (AI) and machine learning (ML) offer a cost-effective way of evaluating the condition of buried water mains taking in hundreds of variables which is more than normal management consulting practices. Machine learning as a cloud-based solution is able to leverage existing water main pipe data and produce a more accurate model than typical age-based models. Detecting more high-risk water main pipes using ML likelihood of failure (LoF) probability analysis provides improved decision making for directing leak detection, direct pipe inspection, and renewal and replacement activities. This paper explores the applications of AI/ML in underground water pipe distribution systems using example data sets from for a large water utility and a medium sized water utility analyzing the accuracy of an aged-based methodology versus applied ML. © 2019 American Society of Civil Engineers." "44061451800;57206996179;57205673606;57197936167;","Advancing In-line Inspection technology and pipeline risk management through advanced analytics of big data",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064745247&partnerID=40&md5=15d27070dc69bc1db6fd04d7d5ee75ed","Every year we hear more and more about Big Data, how it's going to be the future, how machine learning is going to change what we are capable of and what we do. However, after all this time, the same questions remain as it applies to our industry: How do I get Big Data? How will it make a difference to integrity management program efficiency and ultimately pipeline safety? And, when will it be here? This paper focuses on answering these simple questions by providing a context for existing, active systems that are successfully employed today in the in-line inspection world and, particularly, their subsequent role in pipeline integrity decision making. Additionally, we'll examine how such computing technologies as Big Data, ""machine learning"", AI, and cloud computing are evolving with the increasing expectation of getting more from the significant volumes of information and performance data available to the pipeline operator and Inline Inspection (ILI) vendors. We'll discuss key considerations necessary for establishing a robust and viable framework needed to manage and analyse the vast quantities of pipeline inspection data being collected every day as a means for performance and reliability improvement, including current and historical digital technologies of both inline and NDE dig excavation systems. Furthermore, how, with an ability to analyse pipeline defect characteristics over a wide spectrum of pipelines and conditions, development processes for advanced characterization and accuracy improvements can be dramatically accelerated relative to historic methods. Insight will be provided as to how advanced analytics are now being further enabled with the massive processing power readily available through these cloud-based systems. Lastly, regarding higher accuracy and more consistent field calibration data, when we combined these Big Data environments with the ongoing improvements in ILI, we will outline how that can open up the possibilities for operators to better understand their pipeline risks and how to prioritize most efficiently when it comes to managing repair and maintenance. © 2019 Clarion Technical Conferences." "57205610424;55538601800;","Mobile observatory uses sight and sound to quanify drainage conditions for O&M and real time control",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85060818314&partnerID=40&md5=30611c27d9063ec55d8edd8cb295753c","Sight&Sound patterns produced by sewer flows can be modeled with Machine Learning to affordable quantify flow parameters at a vast number of drainage sites. Such quantifications produce low cost data to fill in gaps of understanding about very expensive invisible infrastructure lying beneath cities. As a more complete operational understanding of infrastructure becomes affordable, a more cognitive ML system can use that performance data to creation and fine tune a predictive mathematical model for system wide drainage fine tuned by on going flow parameters supplied from affordable flow monitors. With such modeling real time or near real time flow control holds much promise to mitigate overflows into our basements, streets, streams, lakes, rivers and oceans. A final note-the altaStation design also comprehends the need for actuators in sewers. An over-the-air software update available in 2019 will provide closed loop control functions that facilitate cloud based, supervisory control system. Copyright © 2018 Water Environment Federation" "57205427916;57205427465;57205428926;","From insight to foresight: Knowing how to apply artificial intelligence in the oil & gas industry",2019,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059978832&partnerID=40&md5=d3bc1967b1cbd4b60dc12d1eb62ca26a","We are in an era where digital technologies are developing at exponential rates and transforming industries wholesale. The confluence of machine learning advances, accelerated growth in acquired data, on-demand CPU and GPU driven computing such as cloud infrastructure, and other advances in automation and robotics are causing an industrial revolution that some term as the ""Fourth Industrial Revolution"". Given that all these transformative technologies are now available and rapidly reinventing other industries, why is the rate of adoption in the oil and gas industry so slow? How can we best utilize these advances to stop drowning in data and instead transform this data into information and knowledge in order to enable secure and intelligent automation in oilfield operations? The oil and gas industry has attempted, at times successfully, a multitude of big data and analytical techniques to further describe and analyze the systems' or system of systems' subsurface interactions. While the proofs of concepts have shown promise, structural difficulties embedded in the design of 20th century systems hinder the implementation of the methods and procedures now part and parcel of the 21st century, driven forth because of the Fourth Industrial Revolution. Unfortunately, 20th century procedures are not able to incorporate 21st century driven processes and methods of conducting business. We outline some of the structural challenges facing the oil and gas industry and describe a few of the solutions that have been developed to help companies in the industry. These include applications from the subsurface in geophysics, completions design, and production. Overcoming data silos in traditional data infrastructure requires a novel approach to cloud infrastructure that respects user access, data privacy, and data residency requirements of companies. Assessing data for quality and for reasonable diversity and variation in order to answer questions posed by oil & gas companies can be quite profound. This critical step prevents companies from spending lots of non-productive time and money trying to develop and tune machine learning algorithms to produce answers that are simply not available in the data. Further, getting data to be in a form suitable to apply artificial intelligence can be quite involved. We illustrate the above challenges by several subsurface examples and then describe the implementation of novel solutions. What we will show is that the oil and gas digital highway presently has data traffic jams preventing it from moving at the speed of light. Removing these traffic jams offers decision-makers the Revolution on the previous one while showcasing the concept of [3] Creative Destruction that supplants one technology with another. According to Joseph Schumpeter, the ""gale of creative destruction"" describes the ""process of industrial mutation that incessantly revolutionizes the economic structure from within, incessantly destroying the old one, incessantly creating a new one"". The various Industrial Revolutions are described in Figure 1 as: 1. IR 1.0 was steam power and the advent of industrialization 2. IR 2.0 was the introduction of factories, mass manufacturing and other techniques of mass production 3. IR 3.0 was the semiconductor revolution with advances in computers, electronics and robotics/ automation 4. IR4.0 is about cyberphysical systems and the marriage of hardware and software to produce autonomous systems [4] The Cyber-Physical Systems (CPS) in Figure 2 are integrations of computation, networking, and physical processes. Embedded computers and networks monitor and control the physical processes, with feedback loops where physical processes affect computations and vice versa. Independent systems have existed and are continuously moving towards becoming autonomous in nature, for example, the above surface rotating equipment in the oilfield whose health is monitored in real-time and event triggers manage and alleviate catastrophic failures. [5] Artificial Intelligence is realizing the intelligent interoperations of system of systems defined as a collection of task-oriented or dedicated systems that pool their resources and capabilities together to create a new, more complex system which offers more functionality and performance than simply the sum of the constituent systems. Upstream oil and gas has specialized over the last 100 years into highly developed scientific and commercial niches. While the outcome has been engineering and academic disciplines driving forth accuracy in decisions within the disciplines, the integration of the various niches, which can be stated as the intelligent communication of information between and among systems, has remained a struggle. We discuss this further under the section entitled ""Data Management"". It is easy to recognize that where there is high quality communication of information among systems, artificial intelligence has made significant strides. For example, facial recognition is now used routinely for expediting in airports for expediting immigration lines, online retail and companies like Netflix use artificial intelligence to understand customer preferences and refine and target their offerings, personal assistants like Microsoft Cortana, Amazon Alexa, Google Now, and Apple Siri use voice recognition for many applications such as home automation. opportunity to move from insight to foresight - looking out in front instead of the rearview mirror to drive change. © Copyright 2018, Society of Petroleum Engineers." "57192386104;57205735747;57217749096;57205747497;57205742396;7006093015;","Remote cloud-based automated stroke rehabilitation assessment using wearables",2018,"10.1109/eScience.2018.00063","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85061384336&doi=10.1109%2feScience.2018.00063&partnerID=40&md5=321330a162eaa20badba3d9ace5b91a7","We outline a system enabling accurate remote assessment of stroke rehabilitation levels using wrist worn accelerometer time series data. The system is built based on features generated from clustering models across sliding windows in the data and makes use of computation in the cloud. Predictive models are built using advanced machine learning techniques. © 2018 IEEE." "57205500439;56366080900;22634069200;57037171600;55967354600;","Study of enteromorpha idendification based on machine learning technology",2018,"10.1109/OCEANSKOBE.2018.8559374","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85060296421&doi=10.1109%2fOCEANSKOBE.2018.8559374&partnerID=40&md5=89dddcb3a49d6a55c05ab796718845e0","In this paper, we construct an enteromorpha remote sensing dataset. We take the eight channels data, from Band 1 to Band 8, in Landset 8 as spectral features of the dataset and consider the features extracted by Local Binary Patterns algorithm(LBP) and Gray-level Co-occurrence Matrix algorithm (GLCM) as texture features. Combining these two types of features, we create a novel machine learning dataset including enteromorpha, land, seawater, and cloud samples. Then we use the single-channel threshold method and the multi-channel ratio method to label the entire data set pixel-by-pixel based on the feature of spectral characteristics and ground reflection spectra. Four classifiers are trained via Support Vector Machine (SVM), K Nearest Neighbor (KNN), Expectation Maximization (EM) and Stack Autoencoders (SAE) on our dataset respectively. A new classifier with a set of weight coefficients, which is called weighted classifier, can be obtained by combining above classifiers according to the accuracy of the four classifiers on all kinds of samples. Experiments show that the classification accuracy of these classifiers in our dataset are higher than 90% and weighted classifier reaches 93.59%. If we only focus on enteromorpha data, the best classifier is KNN classifier which has the correct rate of 95.8%. © 2018 IEEE." "57189042712;57203813157;57203820620;57203820084;","Mobile Mechanic - An innovative step towards Digital Automobile Service",2018,"10.1109/ICSCET.2018.8537243","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059410451&doi=10.1109%2fICSCET.2018.8537243&partnerID=40&md5=bbf6dacc30e06a18b74ad80ea40654a3","In the 21 st century's work engrossed environment, everyone is in a rush to reach their destination on time, nothing should be a hindrance to their way of achieving their goal. Considering a real-life problem where a person faces troubles when their car breaks down in the middle of the road, can really become a major problem especially when this occurs on highways or distant places. This can be solved by using our application which can help the users to find nearby location of service or a repairing center. The application would be subdivided into two main sections namely the Mechanic's portal and the User's portal. Both these portals would require the respective user to either login or register themselves up before using the application. The location can be provided or obtained using the GPS available in mobile phones. The User's portal would display a map with the current location of themselves along with all the available repair shops around them, which shall be consistently updated using the plethora of location services available on Microsoft Azure Cloud. User can get detailed information about these shops along with a route to reach the desired mechanic shop, using Google Map API. Further, the same data shall be utilized for providing interest/ choice based mechanic shops to future users, by implementing Data Analytics and Machine learning available as API's on Azure Cloud Services. This same data can be thereafter be cleansed and provided as employment opportunities for locations which have lesser mechanic reach and for various other assistance, by using Azure Big Data. The application will also be providing a SOS feature for women's safety and a DIY section consisting of videos and documents within the application would let the users try their hands at fixing their own vehicles on occasion of no network coverage or poor internet connectivity. © 2018 IEEE." "56019512200;7102383497;","Data mining for classification of high volume dense lidar data in an Urban Area",2018,"10.5194/isprs-annals-IV-5-391-2018","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85057600992&doi=10.5194%2fisprs-annals-IV-5-391-2018&partnerID=40&md5=2e79e30f940d35407ec0f51fa8852d02","3D LiDAR point cloud obtained from the laser scanner is too dense and contains millions of points with information. For such huge volume of data to be sorted, identified, validated and be used for prediction, data mining provides immense scope and has been used to achieve the same. Certain unique attributes were selected as an input for creating models through machine learning. Supervised models were thus built for prediction of classes through the available LiDAR data using random forest algorithm. The algorithm was chosen owing to its efficiency and accuracy over other data mining algorithms. The models created using random forest were then tested on an unclassified point cloud data of an urban area. The method shows promising results in terms of classification accuracy as overall accuracy of 91.71% was achieved for pixel-based classification. The method also displays enhanced efficiency over common classification algorithms as the time taken to make predictions about the data is reduced considerably for a set of dense LiDAR data. This shows positive foresight of making use of data mining and machine learning to handle large volume of LiDAR data and can go a long way in augmenting efficient processing of LiDAR data. © 2018 Authors." [No author id available],"1st International Conference on Data Science and Analytics, PuneCon 2018 - Proceedings",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85070317866&partnerID=40&md5=327aa6d9ece80221278f9401385fc965","The proceedings contain 71 papers. The topics discussed include: a working paper on use of social media by selected Indian public sector banks; a comprehensive analysis between popular symmetric encryption algorithms; CReP: a packet forwarding protocol in mobile social networks; real-time disease forecasting using climatic factors: supervised analytical methodology; standards and techniques to remove data remanence in cloud storage; using machine learning to find out when to use box-cox transformation on time series data; embedded control system for flash lamp pumped solid-state Nd:glass laser power supply; diabetes readmission prediction using distributed and collaborative paradigms; and text detection from scene videos having blurriness and text of different sizes." "55574821400;6701728686;21741875100;55346329500;15046810600;55587759100;22943323900;57196442844;55407788800;57208247654;","Bridging climate and earth observation data analytics in a federated cloud infrastructure using interoperable multidisciplinary workflows",2018,"10.1109/IGARSS.2018.8519076","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85064210594&doi=10.1109%2fIGARSS.2018.8519076&partnerID=40&md5=0cde5233d7b9caff4a848db4739c1afc","Managing large data sets is a challenge that is compounded when data is distributed across multiple sites. In the case of the combined use of Earth Observation (EO) and climate data, traditional big data approaches are not adequate. This has lead to the development of a new generation of federated cyberinfrastructures, such as the Platform for the Analysis and Visualization of Climate Science (PAVICS), developed by the Computer Research Institute of Montreal (CRIM) and Ouranos, a Montreal-based consortium on regional climatology and adaptation to climate change. PAVICS also reuses key components of Birdhouse, a collaborative project led by the German Climate Computing Center (DKRZ). PAVICS uses standards, such as the Web Processing Service (WPS) developed by the Open Geospatial Consortium (OGC) as a standard interface for providing access to pre-defined geospatial processing services to facilitate integration and foster reuse. This paper presents some of the significant advancements that have been made in recent projects conducted in Canada, in the United States, and in Europe. © 2018 IEEE." "57189391340;6602942477;57200597357;6701615368;","Temporal difference and density-based learning method applied for deforestation detection using ALOS-2/PALSAR-2",2018,"10.1109/IGARSS.2018.8518412","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85063129681&doi=10.1109%2fIGARSS.2018.8518412&partnerID=40&md5=4e9e8a462135c6cd3891427d8cd22096","Remote sensing has established as key technology for monitoring of environmental degradation such as forest clearing. One of the state-of-the-art microwave EO systems for forest monitoring is Japan's L-band ALOS-2/PALSAR-2 which provides outstanding means for observing tropical forests due its cloud and canopy penetration capability. However, the complexity of the physical backscattering properties of forests and the associated spatial and temporal variabilities, render straightforward change detection methods based on simple thresholding rather inaccurate with high false alarm rates. In this paper, we develop a framework to alleviate problems caused by forest backscatter variability. We define three essential elements, namely ""structures of density"", ""speed of change"", and ""expansion patterns"" which are obtained by differential computing between two repeat-pass PALSAR-2 images. To improve both the detection and assessing of deforestation, a ""deforestation behavior pattern"" is sought through temporal machine learning mechanism of the three proposed elements. Our results indicate that the use of ""structure of density"" can introduce a more robust performance for detecting deforestation. Meanwhile, ""speed of change"" and ""expansion pattern"" are capable to provide additional information with respect to the drivers of deforestation and the land-use change. © 2018 IEEE." "57209097561;37015389000;","Augmented Map Based Traffic Density Estimation for Robot Navigation",2018,"10.1109/TENCONSpring.2018.8692049","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065071563&doi=10.1109%2fTENCONSpring.2018.8692049&partnerID=40&md5=14254f2497037f853ce236f0fc3ec826","Most of the work done on robot navigation is focused on map building, localization and obstacle avoidance. In this paper, we particularly focus on intelligent robot navigation based on scene traffic. In this course we initially generate augmented maps using scene traffic flow captured at discrete times of same days, different days and various weather conditions. For facilitating the same, we manually navigate a mobile robot(X80SV) for capturing data(image and sonar map data) across a campus; segmented into zones. Zone specific point cloud maps are then created by merging scene object detection results with sonar map. A Gaussian model is further used for estimating trend for traffic density from the point clouds. This information is later utilized for navigational purposes, to detect the minimum crowded path from a source to destination. Object detection and classification is done using a fine tuned AlexNet. For labelling the training set used for fine tuning, we perform foreground extraction, transfer learning and SVM classification of the scenes. 16 classes of objects including background is used. During realtime navigation, we also perform object detection and classification so that the current traffic trend can be matched with the augmented maps. © 2018 IEEE." "57208553138;55556221300;","Potential of IoT System and Cloud Services for Predicting Agricultural Pests and Diseases",2018,"10.1109/TENCONSpring.2018.8691951","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065055826&doi=10.1109%2fTENCONSpring.2018.8691951&partnerID=40&md5=8e78e66899bbd0081f47cec8372d4f83","Controlling the outbreaks of pests and diseases in agricultural environment, it is still a big challenge to the farmers due to the changing climatic conditions. In this paper we are proposing the alternative method of predicting occurrences of pest and diseases in the plantation, by combining the advantage of IoT farmland monitoring system and Amazon Machine Learning cloud-based services to find hidden patterns into data. Logistic regression algorithm used to train our IoT collected dataset and classify the data with acceptable model quality score, to estimate the diseases forecasting based on sensing technology. © 2018 IEEE." "57217699055;57205580374;57202620886;57188878642;57205578338;","Elasticdocs as an automated information retrieval platform for unstructured reservoir data utilizing a sequence of smart machine learning methods within a hybrid cloud container",2018,"10.3997/2214-4609.201803242","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85093689436&doi=10.3997%2f2214-4609.201803242&partnerID=40&md5=119c8cebc117deec49ef0cfb8202b351","There is a tremendous amount of information available and stored in digital geoscientific documents and published reports in the energy industry. These documents contain a distillation of reservoir information from diverse discipline of geologists, geophysicists, petrophysicists and drillers, that are stored in unstructured format, which find further use in succeeding reservoir modeling stages. In particular, national data management repositories and oil companies hosts these huge amounts of historical well reports containing information such as lithology, hydrocarbon shows, and other reservoir data. Due to the large volume, vintage variety, and non-standardized formats, extraction of valuable information that are used as inputs for interpretation, is an arduous, very time-consuming task. Our solution is to develop ElasticDocs a machine learning-enabled platform in a hybrid cloud container that automatically reads and understand hundreds or thousand of technical documents with little human supervision through a smart combination of machine learning algorithms including optical character recognition (OCR), elatic search, natural language processing (NLP), clustering and deep convolutional neural network. The platform uses a hybrid, 2-tier data service architecture leveraging on the strength of both the strength of local servers and cloud to enhance data security, integrity, and accessibility. © EAGE Reservoir Geoscience Conference, ResGeo 2018. All rights reserved." "57209498999;57194870778;26424302500;56428234100;","Cloud-based borehole image interpretation workflow using machine learning",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083938338&partnerID=40&md5=b9113ad487a010bfb70ef0fffce3c447","We propose a new workflow for borehole image interpretation incorporating machine learning methods and recent cloud technologies. In order to predict facies from borehole image, clean labels, i.e. the borehole image with annotated facies intervals, must be collected. We propose a solution to accelerate the labelling process by introducing an automatic segmentation and unsupervised learning algorithm based on image texture features. We then present a web based application and pipeline from the user interface to the cloud database that allows the user to easily construct and update the digital facies library and we show first results of our automated image recognition. © 2018 Society of Petroleum Engineers. All rights reserved." "57215911633;57215913812;57217951270;36781870300;","Image processing and restriction of video downloads using cloud",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082353982&partnerID=40&md5=cf8cd87311f13e313084c5958f2ba4fc","Flower image classification using deep learning and convolutional neural network (CNN) based on machine learning in Tensor flow. Tensor flow IDE is used to implement machine learning algorithms. Flower image processing is based on supervised learning which detects the parameters of image. Parameters of the image were compared by decision algorithms. These images are classified by neurons in convolutional neural network. Video processing based on machine learning is used in restriction of downloading the videos by preventing the second response from the server and enabling the debugging of the video by removing the request from the user. © 2018 Authors." "57203576845;57208855690;57201914273;55445979600;56401125500;55446100900;","A Novel approach for prediction of heart disease: Machine learning techniques",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082344458&partnerID=40&md5=ebd40c790915a462e6489d18ba36b2b8","Heart disease and machine learning are the two different words where one is related to medical field and another one to artificial intelligence. In medical filed most of them are facing the problems with the heart disease and machine learning is developing area in computer science. Heart disease is general called cardiac disease where it gives the more data or information, it is to be collected to give the reports for the patients and the machine learning also requires the data for predicting and to solve the problems. Machine learning techniques are used in prediction of heart diseases where it gives the faster prediction with less computation time and better accuracy to progress their health. Heart disease prediction requires lot of data for predicting and in cloud computing also we have more data and the data available in cloud it is difficult to analyze. So we use machine learning algorithms or techniques to predict the heart disease and the in the similar way we can apply these algorithms or techniques to predict or analyze the data that is available in cloud. In this paper we are going to use machine learning algorithms called Backpropagation Algorithm and later we use optimization algorithm later. Backpropagation algorithm deals with the artificial neural networks. Backpropagation is a method used to calculate the error contribution of each neuron after a batch of data (in image recognition, multiple images) is processed. This is used by an enveloping optimization algorithm to adjust the weight of each neuron, completing the learning process for that case. Machine learning algorithms and techniques are used for recognize the intensity of risk issues in humans and it helps the patients to take safety measures in well advances to save the patient's life. © 2018 Authors." "57215914751;55150308500;","An understanding of machine learning techniques in big data analytics: A survey",2018,"10.14419/ijet.v7i3.12.16450","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082343877&doi=10.14419%2fijet.v7i3.12.16450&partnerID=40&md5=25abd611f3e4e2b900eea96b03be76d1","Big data is a Firing Term in the recent era of the modern world, due to the information exploita-tion; there is an enormous amount of data produced. Big data is a powerful momentum of infor-mation and communication technology field due to the effect of growing data in healthcare, IOT, cloud computing, online education, online businesses, and public management. The produced data is not only large but also complex. Big data has a large amount of unstructured data so that there is a need to develop advanced tools and techniques for han-dling big data. Machine Learning is a prominent area of Artificial Intelligence. It makes the system to make intelligent resolutions by giving the knowledge to achieve the goals. This study reviews the various challenges and innovative ideas for big data analytics with machine learning in different fields over the past ten years. This paper mainly organized to identify the research projects based on the discussions over machine learning techniques for big data analytics and provide suggestions to develop the new projects. © 2018 Authors." "57210929194;24402703700;57204647920;","The guidance seeker onboard romote sensing satellite to enhance mission effectiveness",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85071949027&partnerID=40&md5=2b8e463546fb362d128812403f9957dd","Nowadays, the swath has trend to get smaller with the development of high-resolution satellite for fixed number of pixels. So it is getting more important to capture clear target image efficiently. Otherwise, it might take several times to have a clear image if there is cloud between the remote sensing instrument (RSI) and target image. The objective of this paper is to develop an on-board guidance seeker on RSI. The on-board guidance seeker is used to get wideangle images and detect target before the task target is captured. Based on wide-Angle images and detected-Target, the path of satellite can be fine turned to direct ground target without cloud blocking ground target to get clear image or to find target ships and so on. The on-board guidance seeker includes a wide-Angle camera, an on-board machine learning unit and an on-board real time path fine-Tuning unit. This kind of on-board AI-driven design can enhance the mission effectiveness. It also can be applied to rescue in the open sea. © 2018 Asian Association on Remote Sensing. All Rights Reserved." "55357841300;57208632282;47161085800;57208628181;57205390249;57207195307;","Research and application of machine-learning-oriented spacecraft health management platform",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065341564&partnerID=40&md5=d234c8a94632d492afa4177a9f536cd9","The complex environment, conditions, aging failure and other comprehensive factors, making the spacecraft fault detection, diagnosis, prediction exceptionally difficult. The capability of traditional expert knowledge systems in spacecraft's system-level fault handling is limited, and still require designers and domain experts to spend a lot of time on mechanism analysis, formula derivation, and experimental verification. The traditional manual-analysis-based work model obviously cannot meet the development requirements of the spacecraft's high-reliability and quantity growth. In recent years, the new machine learning platforms (e.g., Microsoft's Azure, Google's Cloud Machine Learning, Alibaba's PAI), which have friendly process analysis framework, rich plugs and play machine learning tools and distributed services, can provide new ideas of complex problem handling in the fields of spacecraft. It proposed a machine-learning-oriented spacecraft health management platform design based on the analysis of the difficulties in spacecraft fault diagnosis and fault prediction, including modeling, health management platform architecture, massive data preprocessing methods, TensorFlow and other typical machine learning tools integration method, diagnosis and prediction of distributed service design and the results display and evaluation design, etc. Finally, the actual application effect is verified with the solar array power forecasting and other cases. The experiment results show that the research can provide technical reference for the research and application of spacecraft health management technology based on machine learning, and ultimately improve the safety of the spacecraft. © 2018 by the International Astronautical Federation (IAF). All rights reserved." [No author id available],"Delivering solutions at the intersection satellite big data, cloud computing, machine learning and IoT technology - The case of SatSure",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85065339959&partnerID=40&md5=96a9e219274f676fcd9618ce4a60bcae","Doubling farm income by 2022-23 translates into enabling farmers to grow at more than 1.5 times the rate of Gross Domestic Product (GDP) growth in India over the next 5 years. This is no mean feat in leapfrogging to be achieved in a community where 67% of the 193.76 million hectares cropped area in the country is held by marginal farmers with holdings below one hectare. The three pillars that support farmers in the entire lifecycle of agriculture from a monetary perspective are credit, insurance and trade institutions. While much of the discussion and debate today centres around institutions that are involved in producing better seeds, fertilizer and equipment for better farm outcomes, it is important to acknowledge that discussing the future of monetary institutions and trade aspects in both technology and policy is as necessary as every other aspect of the farm and farmers. Utilizing satellite Big Data for driving efficiencies in financing agriculture, as global digitization drive is set to help both sides (farmers and monetary institutions). Gaining access to credit and insurance based on farm/farmer level analytics will also mean fixing problems in fraudulent utilization of loans as well as claims. For example, a bank providing loans for a particular crop can also have an automated flagging system for the entire lifecycle of the farm starting from sowing to harvest. This shall make sure that the credit is utilized effectively and will also give a perspective to the banker on the estimations of yield based, the health of the crop, etc. Similarly, an underwriter can get a perspective on the risks of monsoon, availability of irrigation, land fertility over time, weather uncertainties over cropping period, etc., providing complete transparency in sizing of the claim in the underwriting process. Copyright © 2018 by the International Astronautical Federation." "57202998689;36730895700;57205433409;57191746914;55002247800;57204106697;57205433321;57203025148;","It's raining barrels: Cloud computing in the O&G industry",2018,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85060023883&partnerID=40&md5=bd92f4a7803a068792b57f7655b82fc6","Cloud computing has become the buzz word in the last few years. All the service industries from different fields are using cloud computing based data analytics to optimize their operations in order to improve the customer user experiences and the overall efficiency. It is mainly due to the use of high volume computing backed by significant development in advanced hardware capabilities. This paper describes the 'what', 'why' and 'how' based questions on cloud computing. To start with, a brief introduction of cloud computing has been discussed along with the history of computation usage in oil industry. Followed by that a brief introduction highlighting the significance of Artificial intelligence and machine learning in the current computing environment has been explained. As the industry moves towards more and more usage of digital oilfield techniques, the dominance of high end computing and data analytics in the oil and gas industry is also showcased. Once cloud computing is established as a standard, the paper further discusses different modes of cloud delivery service models- Infrastructure as a service (IaaS), Platform as a service (PaaS) and Software as a Service (SaaS). The paper also identifies the challenges and concerns/issues that comes as a part of cloud computing methodology and then describes how these all challenges are being addresses. This paper elucidates the comprehensive view of the cloud computing landscape in oil and gas arena through a review of available noteworthy open source literature. © 2018 Society of Petroleum Engineers. All rights reserved." "57205244267;55758943200;","Machine learning in cloud: Sentiment analyzing system",2018,"10.14419/ijet.v7i4.40.24387","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85059261090&doi=10.14419%2fijet.v7i4.40.24387&partnerID=40&md5=1b2a83bde69e4f5193b10f0b2ed7b935","As the number of computer users increases, numerous content has been generated by them. Machine learning as one of the main direction of natural language processing, allows computer systems to extract various information from the generated content. Processing results determine the sentiments of the text to extract the author's emotional evaluation that is expressed in the text. The aim of the project was to develop the Sentiment Analyzing system by using Machine Learning algorithms on cloud-based system. The paper describes the development process of Sentiment Analyzing System in English language. Two Machine Learning algorithms, SVM and Naïve Bayes classifier, have been inspected and Cloud computing used to develop and publish web application. The testing results demonstrate the accuracy of the work in proposed method. © 2018 Authors." "57195102154;57204491819;56217320100;57202647606;","Automatic LOD0 classification of airborne LiDAR data in urban and non-urban areas",2018,"10.1080/22797254.2018.1522934","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85055260551&doi=10.1080%2f22797254.2018.1522934&partnerID=40&md5=8177a266b6cce93946555996c72c9c27","Point clouds are a very detailed and accurate vector data model of 3D geographic information. In contrast to other data models, no standard has been defined for visualization and management of point clouds based on levels of detail (LOD). This paper proposes the application of the concept of LODs to point clouds and defines the LOD0 for point cloud classification (the lowest possible level of detail) as urban and non-urban. A methodology based on the use of machine learning techniques is developed to perform LOD0 classification to airborne LiDAR data. Point clouds acquired with airborne laser scanner (ALS) are structured in grid maps and geometric features related with Z distribution and roughness are extracted from each cell. Six machine learning classifiers have been trained with datasets including urban (cities) and non-urban samples (farmlands and forests). The influence of grid size, point density, number of features and classifier type are analysed in detail. The classifiers have been tested in three case studies. The best results correspond to a grid size of 100 m and the use of 12 geometric features. The accuracy is around 90% in all tests and Cohen’s Kappa index reaches 81% in the best of cases. © 2018, © 2018 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group." "57218593297;57190957993;","Securing cloud by mitigating insider data theft attacks with decoy technology using Hadoop",2018,"10.14419/ijet.v7i2.31.13407","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85047856614&doi=10.14419%2fijet.v7i2.31.13407&partnerID=40&md5=1882bcc94adc67c572c7440c40d06ce9","Cloud Computing has been intrinsically changing the way we utilize computers to keep and retrieve our personal & business data. With the advent of this emerging paradigm of computing, it arises the new security challenges. Existent cryptographic data security techniques i.e., encryption deteriorated in preventing data theft attacks once the key is compromised, especially those perpetrated by insiders. Cloud Security Alliance reckoned this threat as a significant danger of Cloud Computing. Although the majority of Cloud users are very much known of this risk, they are leftover with the only choice of trusting the cloud service provider, regards to their data protection. In this paper, we propose an alternate way to secure data on the cloud which is more efficient and secure by the concoction of user profile mapping using Hadoop framework and offensive decoy technology. © 2018 Authors." "57194041416;57201646954;55312192000;57194253460;","LiDAR Based Sensor Verification",2018,"10.4271/2018-01-0043","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045507351&doi=10.4271%2f2018-01-0043&partnerID=40&md5=43971302da8f9b3dfd98732008e0ab2a","In the world of automated driving, sensing accuracy is of the utmost importance, and proving that your sensors can do the job is serious business. This is where ground-truth labeling has an important role in Autoliv's validation process. Currently, annotating ground-truth data is a tedious and manual effort, involving finding the important events of interest and using the human eye to determine objects from LiDAR point cloud images. We present a workflow we developed in MATLAB to alleviate some of the pains associated with labeling point cloud data from a LiDAR sensor and the advantages that the workflow provides to the labeler. We discuss the capabilities of a tool we developed to assist users in visualizing, navigating, and annotating objects in point cloud data, tracking these objects through time over multiple frames, and then using the labeled data for developing machine learning based classifiers. We describe how the output of the labeling process is used to train deep neural nets to provide a fully automated way to produce vehicle objects of interest which can be used to find false-negative events. To do this with a human analyst takes as much time as to play back the entire data set. However, with a fully automated approach it can be run on many computers to reduce the analysis time. We present this time savings as well as the accuracy of the labels achieved and show how this approach provides substantial benefit to Autoliv's validation process. © 2018 SAE International. All Rights Reserved." "57201469360;57201362771;57201368459;57201475091;57201475351;","A proposals of convolution neural network system for malicious code analysis based on cloud systems",2018,"10.14419/ijet.v7i2.12.11040","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85045007763&doi=10.14419%2fijet.v7i2.12.11040&partnerID=40&md5=d0fcebe8fcba09ea4c5db00b90d76cf0","Background/Objectives: In the information security field, artificial intelligence must be applied first. This is because the frequency of malicious code is too high and the processing method is too difficult, which is very difficult for human to handle. Methods/Statistical analysis: In this paper, we developed a program to classify malicious codes into images and a Tensorflow system to classify malicious codes. The malware used as input was the computer virus code used in the BIG 2015 Challenge. This dataset, called a Kaggle dataset, consists of 10,868 bytes of train set. Findings: We used the Tensorflow SLIM library to develop this machine learning malware learning machine. This resulted in more than 80% accuracy. Especially, when the CRIS-Ensemble algorithm was added, the accuracy was 97%. The study of malicious code analysis using machine learning consists of two major parts. First, the process of making the virus into images is important. To classify 10,868 Kaggle malware datasets that the BIG 2015 winner showed 99.6% accuracy, Tensorflow's accuracy and parameter tuning are important, but finding the way to make good images is the most important technique Improvements/Applications: The results show that the malicious code classification system using machine learning can be an effective method to classify malicious code of malicious code by the accuracy of the result and ease of use. © 2018 Yong-kyu Park et. al." [No author id available],"Proceedings - 2017 International Conference on Green Informatics, ICGI 2017",2017,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85041684643&partnerID=40&md5=3d2ecb82b6f6b0568554d30e975c2014","The proceedings contain 47 papers. The topics discussed include: efficient GPU-based parallel Kriging algorithm for predicting the air quality index; hybridization of PMF and LSTM for recommendation of intelligent resource; research and application of clustering algorithm based on shared nearest neighbor; a low complexity extended Kalman filter algorithm for neural network digital predistortion of power amplifier; a monocular ranging algorithm for detecting illegal vehicle jumping; abstract super points from core network by unique candidate list; character-based convolutional grid neural network for breast cancer classification; an independent forwarding algorithm based on multidimensional spatial superposition model in SDN; machine learning based LncRNA function prediction; resource reconfiguration module for cross-cloud system; the implementation of a GPU-accelerated virtual desktop infrastructure platform; implement a virtual development platform based on QEMU; improving tasks scheduling performance in cloud computing environment by using analytic hierarchy process model; cloud bank liquidity risk prediction and identification, liquidity creation, and resource fragility; a task scheduling scheme for preventing temperature hotspot on GPU heterogeneous cluster; exploring critical success factors of building green logistics business in Fuzhou; cooperative self-organized energy-saving mechanism of cellular network based on hybrid energy supplies; encouraging knowledge sharing among green fashion communities; the key success factors of developing intelligent logistics within pharmaceutical industry in Fujian free trade area; and to assess the core competitiveness of Taiwanese design departments' students." "6507534695;","Machine-learning point cloud classification",2017,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85035035463&partnerID=40&md5=5ea0eb317d6bcf3d03056dd22d6feb8e",[No abstract available] "48161127100;7006517690;","Geocomputation: Data, Methods, and Applications in a New Era",2017,"10.1016/B978-0-12-409548-9.09600-7","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85082357776&doi=10.1016%2fB978-0-12-409548-9.09600-7&partnerID=40&md5=a409bf27164db92e9e4bf2fd18a3487e","Geocomputation is a research field where computational technology and methods are applied to geographic data. We are in the midst of a fundamental change that affects how computers are used in handling geographic data. The Web has provided access to much new data as well as hardware agnostic software platforms for mapping and analysis. These new data and capabilities offer both opportunities and challenges to researchers. Traditional methods of geocomputation face new challenges to cope with and take advantage of these new opportunities. In this article we survey these changes and discuss how they are influencing new directions of spatial data applications. © 2018 Elsevier Inc. All rights reserved." "56783909900;55918988300;23013616900;6701381246;","Erratum: A machine learning method for co-registration and individual tree matching of forest inventory and airborne laser scanning data [Remote Sens., 9, 505 (2017)] DOI: 10.3390/rs9050505",2017,"10.3390/rs9070692","https://www.scopus.com/inward/record.uri?eid=2-s2.0-85022337303&doi=10.3390%2frs9070692&partnerID=40&md5=04f9ba7ee1450f546c2cbafdb8d008ab","Since Equation (2) has been rearranged incorrectly during preparation for this article [1], the authors would like to correct the relevant text of Section 3.4.3 as follows: 3.4.3. Synthetic Tree Detections A tree crown is defined by the tree species-specific parameters b and c (see Table 2), the tree height h, crown base height cbh, and its crown radius r (all provided by the Waldplaner 2.0 software, [2]). The height above ground ei of a point with distance di to the tree location less than or equal to r is calculated by Equation (2). Please note that since the parameters of the Douglas fir were not given by Pretzsch [3], these were assumed to be similar to the parameters of the Norway spruce. ei = h - c · (h - cbh) · (di/r)1/b (2) For each plot, we generated uniquely-distributed xy-coordinates with the same pulse density as the original point cloud. For each point, the z-coordinate was calculated by applying Equation (2) for all trees and adding the value of the DTM (of the original ALS point cloud) at the tree location. Finally, the point is linked to the tree which resulted in the maximum z-coordinate and-for a subsequent evaluation-it is labeled with the ID of that tree. We apologize for any inconvenience caused to the readers by these changes. The changes do not affect the scientific results. The manuscript will be updated and the original will remain online on the article webpage. © 2017 by the authors." "6507194541;57219176330;7005583247;57219173568;57219173295;57219176366;57219174081;7401595700;7004902724;","Online advanced uncertain reasoning architecture with binomial event discriminator system for novelty detection in smart water networks",2017,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85091596172&partnerID=40&md5=322e650678bf3047dbb0b68a96fc1079","Minimising the loss of treated water from water supply systems due to bursts is an ongoing issue for water service providers around the world. Sensor technology and the 'big data' they generate combined with machine learning based analytics are providing an opportunity for automated event detection. AURA-Alert has been developed as an online (Software as a Service) system which automates the training data selection (by selecting data with acceptable Match Strength and with regular retraining). The addition of a Binomial Event Discriminator service can produce alerts based on windows of thresholded match distances. A pilot deployment on over 200 live data streams in the cloud has been deployed as part of the SmartWater4Europe project. Examples of analysis for real events are presented. For a historic subset of eight data streams over a three month period up to 58% of bursts were detected (depending on window used for evaluation). It is concluded that the system is an effective and viable tool for novelty detection for water network time series data with potential for wider applicability. Key strengths include lack of per site configuration, data-driven self-learning (from periods of normality), real-time, high scalability and full automation of model retraining. © 2020 CCWI 2017 - 15th International Conference on Computing and Control for the Water Industry. All rights reserved." "57196831038;57203371490;","Cloud based processing of free and commercial earth observation data with PCI GXL, populating and analysing data with the Australian Geoscience Data Cube software",2017,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85051431120&partnerID=40&md5=49b90a935d7c2f77de4a29ca2fa75293","Over recent years, corporate and public satellite operators have provided API’s to their image archives and distributed image processing has moved into the cloud. With this in place, the development of multi-sensor cloud enabled automated image pre-processing and analysis tools, feeding data cubes such as released by Geoscience Australia, have become necessary. Such a system is PCI GXL. The system can run on an Amazon cloud and extract time series of data from the archive of Planet Labs or ESA through their respective API’s. A range of image preprocessing techniques can be undertaken with a set of software libraries and executables. GXL powered automated workflows would typically include image calibration, atmospheric corrections, image to image registration to 1/10th of the image pixel, image compositing and image fusion. Pre-processing is optimized in a distributed way to feed into a data cube released as open source by Geoscience Australia. Once imagery has been deposited in the data cube, quantitative time series analysis can be performed by GXL to enable rapid identification of changes. A new approach to the analysis of multi resolution data supports the combined use of high spatial resolution commercial data with lower spatial resolution imagery collected at a high temporal frequency. Image segmentation and feature calculation based on pure pixels only allow the abstraction of data to units of homogeneous surfaces reducing computational efforts and complexity for a multi resolution spatial and temporal image analysis. A set of software libraries dedicated to SAR image processing for polarimetry and interferometry enable the use of SAR data with all its attributes in the data stack. Near real time change detection and mapping applications for disaster management can be addressed in this way. Machine learning classifiers are able to convert data in the data cube to thematic geo-spatial information efficiently and accurately. The system can be integrated with cloud enabled multi-mission satellite ground segments such as FarEarth to complete the loop from satellite tasking to geo-spatial information delivery to end users. © 2018 International Astronautical Federation IAF. All rights reserved." "57189347919;6603684955;57218448132;","Using Cloud-Based Analytics to Save Lives",2016,"10.1016/B978-0-12-803192-6.00012-8","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84969730169&doi=10.1016%2fB978-0-12-803192-6.00012-8&partnerID=40&md5=9d7a1f5095a607d9f2dc84a6856caa2d","The National Flood Interoperability Experiment (NFIE) is research initiative among government, academia, and industry to help demonstrate the next generation of national flood hydrology modeling to enable early warning systems and emergency response. The goal of NFIE is to answer the questions-What if it were easier to predict more accurately where floods will occur? What if more flood information could be shared in real time to aid in response that is more effective for planning and prevent deaths and property damage? This paper presents an approach to these problems based on cloud computing and machine learning. The paper also addresses the characteristics of cloud computing that can make it an attractive alternative to on-premises infrastructure for much scientific collaboration. © 2016 Elsevier Inc. All rights reserved." "55331191600;26655335900;6603227150;24765165900;22735922800;36944449700;","Large scale thematic mapping by supervised machine learning on 'big data' distributed cluster computing frameworks",2015,"10.1109/IGARSS.2015.7326065","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84962510511&doi=10.1109%2fIGARSS.2015.7326065&partnerID=40&md5=635b34400d2f45a070d8f3337791942c","The Petabyte-scale data volumes in Earth Observation (EO) archives are not efficiently manageable with serial processes running on large isolated servers. Distributed storage and processing based on 'big data' cloud computing frameworks needs to be considered as a part of the solution. © 2015 IEEE." "55308134300;6701525647;57195650018;","Do open clusters have distinguishable chemical signatures?",2014,"10.1051/eas/1567023","https://www.scopus.com/inward/record.uri?eid=2-s2.0-84937806199&doi=10.1051%2feas%2f1567023&partnerID=40&md5=638109ea41b3cfb63dba9b17ebfc40ed","Past studies have already shown that stars in open clusters are chemically homogeneous (e.g. De Silva et al. 2006, 2007 and 2009). These results support the idea that stars born from the same giant molecular cloud should have the same chemical composition. In this context, the chemical tagging technique was proposed by Freeman et al. (2002). The principle is to recover disrupted stellar clusters by looking only to the stellar chemical composition. In order to evaluate the feasibility of this approach, it is necessary to test if we can distinguish between stars born from different molecular clouds. For this purpose, we studied the chemical composition of stars in 32 old and intermediate-age open clusters, and we applied machine learning algorithms to recover the original cluster by only considering the chemical signatures. © EAS, EDP Sciences, 2015." [No author id available],"Proceedings - 7th International Congress on Environmental Modelling and Software: Bold Visions for Environmental Modeling, iEMSs 2014",2014,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84912141184&partnerID=40&md5=2d9f8f03df6a5bb62de6895101d8825d","The proceedings contain 310 papers. The topics discussed include: use of near real time earth observation data infrastructures and open source tools for water resources monitoring and assessment; the virtual machine (VM) scaler: an infrastructure manager supporting environmental modeling on infrastructure-as-a-service clouds; physically based landslide susceptibility models with different degree of complexity: integration in OMS, calibration and verification; a spatial planning tool for the evaluation of the effect of hydrological and land-use changes on ecosystem services; modelling land-use changes in Godavari river basin : a comparison of two districts in Andhra Pradesh; abstractions from sensor data with complex event processing and machine learning; graph clustering based on social network community detection algorithms; linking governance storylines with the D-EXPANSE model to explore the power system transition pathways; and object-based analysis of multispectral RS data and GIS for detection of climate change impact on the Karakoram range northern Pakistan." [No author id available],"Proceedings - 7th International Congress on Environmental Modelling and Software: Bold Visions for Environmental Modeling, iEMSs 2014",2014,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84912126716&partnerID=40&md5=80a36610703d119c81553cdf6b93ae43","The proceedings contain 310 papers. The topics discussed include: use of near real time earth observation data infrastructures and open source tools for water resources monitoring and assessment; the virtual machine (VM) scaler: an infrastructure manager supporting environmental modeling on infrastructure-as-a-service clouds; physically based landslide susceptibility models with different degree of complexity: integration in OMS, calibration and verification; a spatial planning tool for the evaluation of the effect of hydrological and land-use changes on ecosystem services; modelling land-use changes in Godavari river basin : a comparison of two districts in Andhra Pradesh; abstractions from sensor data with complex event processing and machine learning; graph clustering based on social network community detection algorithms; linking governance storylines with the D-EXPANSE model to explore the power system transition pathways; and object-based analysis of multispectral RS data and GIS for detection of climate change impact on the Karakoram range northern Pakistan." [No author id available],"Proceedings - 7th International Congress on Environmental Modelling and Software: Bold Visions for Environmental Modeling, iEMSs 2014",2014,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84912102891&partnerID=40&md5=47e05dae7b53af68795265e83ba42fcd","The proceedings contain 310 papers. The topics discussed include: use of near real time earth observation data infrastructures and open source tools for water resources monitoring and assessment; the virtual machine (VM) scaler: an infrastructure manager supporting environmental modeling on infrastructure-as-a-service clouds; physically based landslide susceptibility models with different degree of complexity: integration in OMS, calibration and verification; a spatial planning tool for the evaluation of the effect of hydrological and land-use changes on ecosystem services; modelling land-use changes in Godavari river basin : a comparison of two districts in Andhra Pradesh; abstractions from sensor data with complex event processing and machine learning; graph clustering based on social network community detection algorithms; linking governance storylines with the D-EXPANSE model to explore the power system transition pathways; and object-based analysis of multispectral RS data and GIS for detection of climate change impact on the Karakoram range northern Pakistan." [No author id available],"Proceedings - 7th International Congress on Environmental Modelling and Software: Bold Visions for Environmental Modeling, iEMSs 2014",2014,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-84912073084&partnerID=40&md5=7621fca50733c52856391bc79137f46c","The proceedings contain 310 papers. The topics discussed include: use of near real time earth observation data infrastructures and open source tools for water resources monitoring and assessment; the virtual machine (VM) scaler: an infrastructure manager supporting environmental modeling on infrastructure-as-a-service clouds; physically based landslide susceptibility models with different degree of complexity: integration in OMS, calibration and verification; a spatial planning tool for the evaluation of the effect of hydrological and land-use changes on ecosystem services; modelling land-use changes in Godavari river basin : a comparison of two districts in Andhra Pradesh; abstractions from sensor data with complex event processing and machine learning; graph clustering based on social network community detection algorithms; linking governance storylines with the D-EXPANSE model to explore the power system transition pathways; and object-based analysis of multispectral RS data and GIS for detection of climate change impact on the Karakoram range northern Pakistan." [No author id available],"ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences",2013,,"https://www.scopus.com/inward/record.uri?eid=2-s2.0-85060427922&partnerID=40&md5=b36c702b10aaa44ace64caa1e6400cc8","The proceedings contain 31 papers. The topics discussed include: canopy surface reconstruction and tropical forest parameters prediction from airborne laser scanner for large forest area; comparison of spatiotemporal mapping techniques for enormous ETL and exploitation patterns; a spatiotemporal prediction framework for air pollution based on deep RNN; applying Thiessen polygon catchment areas and gridded population weights to estimate conflict-driven population changes in south Sudan; detection of behavior patterns of interest using big data which have spatial and temporal attributes; a simple spatially weighted measure of temporal stability for data with limited temporal observations; detecting vessels carrying migrants using machine learning; spatio-temporal data model for integrating evolving nation-level datasets; spatio-temporal visualization of time-series satellite-derived CO2 flux data using volume rendering and gpu-based interpolation on a cloud-driven digital earth; extracting spatiotemporal objects from raster data to represent physical features and analyze related processes; spatiotemporal patterns and socioeconomic dimensions of shared accommodations: the case of AirBNB in Los Angeles, California; elastic cloud computing architecture and system for heterogeneous spatiotemporal computing; efficient lidar point cloud data managing and processing in a hadoop-based distributed framework; a novel approach of indexing and retrieving spatial polygons for efficient spatial region queries; and normalization strategies for enhancing spatio-temporal analysis of social media responses during extreme events: a case study based on analysis of four extreme events using socioenvironmental data explorer (SEDE)."