Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
57
result(s) for
"baseline learning algorithms"
Sort by:
Integrating Machine Learning Ensembles for Landslide Susceptibility Mapping in Northern Pakistan
by
Altalbe, Ali
,
Ali, Rashid
,
Ali, Nafees
in
Algorithms
,
baseline learning algorithms
,
data collection
2024
Natural disasters, notably landslides, pose significant threats to communities and infrastructure. Landslide susceptibility mapping (LSM) has been globally deemed as an effective tool to mitigate such threats. In this regard, this study considers the northern region of Pakistan, which is primarily susceptible to landslides amid rugged topography, frequent seismic events, and seasonal rainfall, to carry out LSM. To achieve this goal, this study pioneered the fusion of baseline models (logistic regression (LR), K-nearest neighbors (KNN), and support vector machine (SVM)) with ensembled algorithms (Cascade Generalization (CG), random forest (RF), Light Gradient-Boosting Machine (LightGBM), AdaBoost, Dagging, and XGBoost). With a dataset comprising 228 landslide inventory maps, this study employed a random forest classifier and a correlation-based feature selection (CFS) approach to identify the twelve most significant parameters instigating landslides. The evaluated parameters included slope angle, elevation, aspect, geological features, and proximity to faults, roads, and streams, and slope was revealed as the primary factor influencing landslide distribution, followed by aspect and rainfall with a minute margin. The models, validated with an AUC of 0.784, ACC of 0.912, and K of 0.394 for logistic regression (LR), as well as an AUC of 0.907, ACC of 0.927, and K of 0.620 for XGBoost, highlight the practical effectiveness and potency of LSM. The results revealed the superior performance of LR among the baseline models and XGBoost among the ensembles, which contributed to the development of precise LSM for the study area. LSM may serve as a valuable tool for guiding precise risk-mitigation strategies and policies in geohazard-prone regions at national and global scales.
Journal Article
A hybrid approach using support vector machine rule-based system: detecting cyber threats in internet of things
2024
While the proliferation of the Internet of Things (IoT) has revolutionized several industries, it has also created severe data security concerns. The security of these network devices and the dependability of IoT networks depend on efficient threat detection. Device heterogeneity, computing resource constraints, and the ever-changing nature of cyber threats are a few of the obstacles that make detecting cyber threats in IoT systems difficult. Complex threats often go undetected by conventional security measures, requiring more sophisticated, adaptive detection methods. Therefore, this study presents the Hybrid approach based on the Support Vector Machines Rule-Based Detection (HSVMR-D) method for an all-encompassing approach to identifying cyber threats to the IoT. The HSVMR-D employs SVM to categorize known and unknown threats using attributes acquired from IoT data. Identifying known attack signatures and patterns using rule-based approaches improves detection efficiency without retraining by adapting pre-trained models to new IoT contexts. Moreover, protecting vital infrastructure and sensitive data, HSVMR-D provides a thorough and adaptable solution to improve the security posture of IoT deployments. Comprehensive experiment analysis and simulation results compared to the baseline study have confirmed the efficiency of the proposed HSVMR-D. Furthermore, increased resilience to completely novel changing threats, fewer false positives, and improved accuracy in threat detection are all outcomes that show the proposed work outperforms others. The HSVMR-D approach is helpful where the primary objective is a secure environment in the Internet of Things (IoT) when resources are limited.
Journal Article
WeatherBench 2: A Benchmark for the Next Generation of Data‐Driven Global Weather Models
by
Rasp, Stephan
,
Sha, Fei
,
Bromberg, Carla
in
Algorithms
,
Artificial intelligence
,
Baseline studies
2024
WeatherBench 2 is an update to the global, medium‐range (1–14 days) weather forecasting benchmark proposed by (Rasp et al., 2020, https://doi.org/10.1029/2020ms002203), designed with the aim to accelerate progress in data‐driven weather modeling. WeatherBench 2 consists of an open‐source evaluation framework, publicly available training, ground truth and baseline data as well as a continuously updated website with the latest metrics and state‐of‐the‐art models: https://sites.research.google/weatherbench. This paper describes the design principles of the evaluation framework and presents results for current state‐of‐the‐art physical and data‐driven weather models. The metrics are based on established practices for evaluating weather forecasts at leading operational weather centers. We define a set of headline scores to provide an overview of model performance. In addition, we also discuss caveats in the current evaluation setup and challenges for the future of data‐driven weather forecasting. Plain Language Summary Traditionally, weather forecasts are made by models that attempt to replicate the physical processes of the atmosphere. This has been very successful over the last few decades as better computers, better observations and model upgrades have lead to steadily improving weather forecasts. However, with rapid advances in artificial intelligence (AI), the question can be asked whether one can simply learn a weather model from past observations or reanalyzes. In the last couple of years, we have seen tremendous progress with state‐of‐the‐art AI models rivaling the best “traditional” weather models in skill. WeatherBench 2 is a benchmark data set designed to evaluate and compare the quality of AI and traditional models. By setting a standard for evaluation, alongside providing open‐source data and code, this project aims to accelerate this research direction and lead to better weather prediction. Key Points WeatherBench 2 is a framework for evaluating and comparing data‐driven and traditional numerical weather forecasting models It provides an evaluation framework, publicly available data sets and a website to assess the state‐of‐the‐art weather models The evaluation protocol has been designed following best practices established in the operational weather forecasting community
Journal Article
Weed Identification by Single-Stage and Two-Stage Neural Networks: A Study on the Impact of Image Resizers and Weights Optimization Algorithms
by
Velayudhan, Kesini Krishnan
,
Arif, Khalid Mahmood
,
Potgieter, Johan
in
Accuracy
,
Algorithms
,
Artificial neural networks
2022
The accurate identification of weeds is an essential step for a site-specific weed management system. In recent years, deep learning (DL) has got rapid advancements to perform complex agricultural tasks. The previous studies emphasized the evaluation of advanced training techniques or modifying the well-known DL models to improve the overall accuracy. In contrast, this research attempted to improve the mean average precision (mAP) for the detection and classification of eight classes of weeds by proposing a novel DL-based methodology. First, a comprehensive analysis of single-stage and two-stage neural networks including Single-shot MultiBox Detector (SSD), You look only Once (YOLO-v4), EfficientDet, CenterNet, RetinaNet, Faster Region-based Convolutional Neural Network (RCNN), and Region-based Fully Convolutional Network (RFCN), has been performed. Next, the effects of image resizing techniques along with four image interpolation methods have been studied. It led to the final stage of the research through optimization of the weights of the best-acquired model by initialization techniques, batch normalization, and DL optimization algorithms. The effectiveness of the proposed work is proven due to a high mAP of 93.44% and validated by the stratified k-fold cross-validation technique. It was 5.8% improved as compared to the results obtained by the default settings of the best-suited DL architecture (Faster RCNN ResNet-101). The presented pipeline would be a baseline study for the research community to explore several tasks such as real-time detection and reducing the computation/training time. All the relevant data including the annotated dataset, configuration files, and inference graph of the final model are provided with this article. Furthermore, the selection of the DeepWeeds dataset shows the robustness/practicality of the study because it contains images collected in a real/complex agricultural environment. Therefore, this research would be a considerable step toward an efficient and automatic weed control system.
Journal Article
Exploring the Steps of Infrared (IR) Spectral Analysis: Pre-Processing, (Classical) Data Modelling, and Deep Learning
by
Mokari, Azadeh
,
Guo, Shuxia
,
Bocklitz, Thomas
in
Algorithms
,
analytical technologies
,
Cosmic rays
2023
Infrared (IR) spectroscopy has greatly improved the ability to study biomedical samples because IR spectroscopy measures how molecules interact with infrared light, providing a measurement of the vibrational states of the molecules. Therefore, the resulting IR spectrum provides a unique vibrational fingerprint of the sample. This characteristic makes IR spectroscopy an invaluable and versatile technology for detecting a wide variety of chemicals and is widely used in biological, chemical, and medical scenarios. These include, but are not limited to, micro-organism identification, clinical diagnosis, and explosive detection. However, IR spectroscopy is susceptible to various interfering factors such as scattering, reflection, and interference, which manifest themselves as baseline, band distortion, and intensity changes in the measured IR spectra. Combined with the absorption information of the molecules of interest, these interferences prevent direct data interpretation based on the Beer–Lambert law. Instead, more advanced data analysis approaches, particularly artificial intelligence (AI)-based algorithms, are required to remove the interfering contributions and, more importantly, to translate the spectral signals into high-level biological/chemical information. This leads to the tasks of spectral pre-processing and data modeling, the main topics of this review. In particular, we will discuss recent developments in both tasks from the perspectives of classical machine learning and deep learning.
Journal Article
Review of Wide-Baseline Stereo Image Matching Based on Deep Learning
by
Yilmaz, Alper
,
Yao, Guobiao
,
Meng, Fei
in
affine invariant feature
,
Algorithms
,
Artificial neural networks
2021
Strong geometric and radiometric distortions often exist in optical wide-baseline stereo images, and some local regions can include surface discontinuities and occlusions. Digital photogrammetry and computer vision researchers have focused on automatic matching for such images. Deep convolutional neural networks, which can express high-level features and their correlation, have received increasing attention for the task of wide-baseline image matching, and learning-based methods have the potential to surpass methods based on handcrafted features. Therefore, we focus on the dynamic study of wide-baseline image matching and review the main approaches of learning-based feature detection, description, and end-to-end image matching. Moreover, we summarize the current representative research using stepwise inspection and dissection. We present the results of comprehensive experiments on actual wide-baseline stereo images, which we use to contrast and discuss the advantages and disadvantages of several state-of-the-art deep-learning algorithms. Finally, we conclude with a description of the state-of-the-art methods and forecast developing trends with unresolved challenges, providing a guide for future work.
Journal Article
A Comparative Performance Assessment of Ensemble Learning for Credit Scoring
2020
Extensive research has been performed by organizations and academics on models for credit scoring, an important financial management activity. With novel machine learning models continue to be proposed, ensemble learning has been introduced into the application of credit scoring, several researches have addressed the supremacy of ensemble learning. In this research, we provide a comparative performance evaluation of ensemble algorithms, i.e., random forest, AdaBoost, XGBoost, LightGBM and Stacking, in terms of accuracy (ACC), area under the curve (AUC), Kolmogorov–Smirnov statistic (KS), Brier score (BS), and model operating time in terms of credit scoring. Moreover, five popular baseline classifiers, i.e., neural network (NN), decision tree (DT), logistic regression (LR), Naïve Bayes (NB), and support vector machine (SVM) are considered to be benchmarks. Experimental findings reveal that the performance of ensemble learning is better than individual learners, except for AdaBoost. In addition, random forest has the best performance in terms of five metrics, XGBoost and LightGBM are close challengers. Among five baseline classifiers, logistic regression outperforms the other classifiers over the most of evaluation metrics. Finally, this study also analyzes reasons for the poor performance of some algorithms and give some suggestions on the choice of credit scoring models for financial institutions.
Journal Article
An efficient approach for textual data classification using deep learning
by
Ullah Khan, Habib
,
Iqbal, Tayyab
,
Alsubai, Shtwai
in
Accuracy
,
Algorithms
,
Artificial intelligence
2022
Text categorization is an effective activity that can be accomplished using a variety of classification algorithms. In machine learning, the classifier is built by learning the features of categories from a set of preset training data. Similarly, deep learning offers enormous benefits for text classification since they execute highly accurately with lower-level engineering and processing. This paper employs machine and deep learning techniques to classify textual data. Textual data contains much useless information that must be pre-processed. We clean the data, impute missing values, and eliminate the repeated columns. Next, we employ machine learning algorithms: logistic regression, random forest, K-nearest neighbors (KNN), and deep learning algorithms: long short-term memory (LSTM), artificial neural network (ANN), and gated recurrent unit (GRU) for classification. Results reveal that LSTM achieves 92% accuracy outperforming all other model and baseline studies.
Journal Article
The sensitivity of pCO2 reconstructions to sampling scales across a Southern Ocean sub-domain: a semi-idealized ocean sampling simulation approach
by
Djeutchouang, Laique M
,
Luke, Gregor
,
Chang, Nicolette
in
Antarctic front
,
Antarctic zone
,
Baseline studies
2022
The Southern Ocean is a complex system yet is sparsely sampled in both space and time. These factors raise questions about the confidence in present sampling strategies and associated machine learning (ML) reconstructions. Previous studies have not yielded a clear understanding of the origin of uncertainties and biases for the reconstructions of the partial pressure of carbon dioxide (pCO2) at the surface ocean (pCO2ocean). We examine these questions through a series of semi-idealized observing system simulation experiments (OSSEs) using a high-resolution (± 10 km) coupled physical and biogeochemical model (NEMO-PISCES, Nucleus for European Modelling of the Ocean, Pelagic Interactions Scheme for Carbon and Ecosystem Studies). Here we choose 1 year of the model sub-domain of 10∘ of latitude (40–50∘ S) by 20∘ of longitude (10∘ W–10∘ E). This domain is crossed by the sub-Antarctic front and thus includes both the sub-Antarctic zone and the polar frontal zone in the south-east Atlantic Ocean, which are the two most sampled sub-regions of the Southern Ocean. We show that while this sub-domain is small relative to the Southern Ocean scales, it is representative of the scales of variability we aim to examine. The OSSEs simulated the observational scales ofpCO2ocean in ways that are comparable to existing ocean CO2 observing platforms (ships, Wave Gliders, carbon floats, Saildrones) in terms of their temporal sampling scales and not necessarily their spatial ones. The pCO2 reconstructions were carried out using a two-member ensemble approach that consisted of two machine learning (ML) methods, (1) the feed-forward neural network and (2) the gradient boosting machines. The baseline data were from the ship-based simulations mimicking ship-based observations from the Surface Ocean CO2 Atlas (SOCAT). For each of the sampling-scale scenarios, we applied the two-member ensemble method to reconstruct the full sub-domainpCO2ocean. The reconstruction skill was then assessed through a statistical comparison of reconstructedpCO2ocean and the model domain mean. The analysis shows that uncertainties and biases forpCO2ocean reconstructions are very sensitive to both the spatial and the temporal scales of pCO2 sampling in the model domain. The four key findings from our investigation are as follows: (1) improving ML-based pCO2 reconstructions in the Southern Ocean requires simultaneous high-resolution observations (<3 d) of the seasonal cycle of the meridional gradients ofpCO2ocean; (2) Saildrones stand out as the optimal platforms to simultaneously address these requirements; (3) Wave Gliders with hourly/daily resolution in pseudo-mooring mode improve on carbon floats (10 d period), which suggests that sampling aliases from the 10 d sampling period might have a greater negative impact on their uncertainties, biases, and reconstruction means; and (4) the present seasonal sampling biases (towards summer) in SOCAT data in the Southern Ocean may be behind a significant winter bias in the reconstructed seasonal cycle of pCO2ocean.
Journal Article
Microbial indicators of environmental perturbations in coral reef ecosystems
2019
We would like to acknowledge the contribution of the Marine Microbes(MM) and Biomes of Australian Soil Environments (BASE) projects, throughthe Australian Microbiome Initiative in the generation of data used in thispublication. The Australian Microbiome Initiative is supported by fundingfrom Bioplatforms Australia through the Australian Government NationalCollaborative Research Infrastructure Strategy (NCRIS). The study was furtherfunded by the Advance Queensland PhD Scholarship, the Great Barrier ReefMarine Park Authority Management Award and a National EnvironmentalScience Program (NESP) grant awarded to BG.The funders had no role in the study design, data collection and analysis,decision to publish, or preparation of the manuscript.
Journal Article