Journal articles on the topic 'Machine Learning Model Robustness'

To see the other types of publications on this topic, follow the link: Machine Learning Model Robustness.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Machine Learning Model Robustness.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Arslan, Ayse. "Rethinking Robustness in Machine Learning: Use of Generative Adversarial Networks for Enhanced Robustness." Scholars Journal of Engineering and Technology 10, no. 3 (March 28, 2022): 9–15. http://dx.doi.org/10.36347/sjet.2022.v10i03.001.

Full text
Abstract:
Machine learning (ML) is increasingly being used in real-world applications, so understanding the uncertainty and robustness of a model is necessary to ensure performance in practice. This paper explores approximations for robustness which can meaningfully explain the behavior of any black box model. Starting with a discussion on components of a robust model this paper offers some techniques based on the Generative Adversarial Network (GAN) approach to improve the robustness of a model. The study concludes that a clear understanding of robust models for ML allows improving information for practitioners, and helps to develop tools that assess the robustness of ML. Also, ML tools and libraries could benefit from a clear understanding on how information should be presented and how these tools are used.
APA, Harvard, Vancouver, ISO, and other styles
2

Einziger, Gil, Maayan Goldstein, Yaniv Sa’ar, and Itai Segall. "Verifying Robustness of Gradient Boosted Models." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 2446–53. http://dx.doi.org/10.1609/aaai.v33i01.33012446.

Full text
Abstract:
Gradient boosted models are a fundamental machine learning technique. Robustness to small perturbations of the input is an important quality measure for machine learning models, but the literature lacks a method to prove the robustness of gradient boosted models.This work introduces VERIGB, a tool for quantifying the robustness of gradient boosted models. VERIGB encodes the model and the robustness property as an SMT formula, which enables state of the art verification tools to prove the model’s robustness. We extensively evaluate VERIGB on publicly available datasets and demonstrate a capability for verifying large models. Finally, we show that some model configurations tend to be inherently more robust than others.
APA, Harvard, Vancouver, ISO, and other styles
3

Thapa, Chandra, Pathum Chamikara Mahawaga Arachchige, Seyit Camtepe, and Lichao Sun. "SplitFed: When Federated Learning Meets Split Learning." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 8 (June 28, 2022): 8485–93. http://dx.doi.org/10.1609/aaai.v36i8.20825.

Full text
Abstract:
Federated learning (FL) and split learning (SL) are two popular distributed machine learning approaches. Both follow a model-to-data scenario; clients train and test machine learning models without sharing raw data. SL provides better model privacy than FL due to the machine learning model architecture split between clients and the server. Moreover, the split model makes SL a better option for resource-constrained environments. However, SL performs slower than FL due to the relay-based training across multiple clients. In this regard, this paper presents a novel approach, named splitfed learning (SFL), that amalgamates the two approaches eliminating their inherent drawbacks, along with a refined architectural configuration incorporating differential privacy and PixelDP to enhance data privacy and model robustness. Our analysis and empirical results demonstrate that (pure) SFL provides similar test accuracy and communication efficiency as SL while significantly decreasing its computation time per global epoch than in SL for multiple clients. Furthermore, as in SL, its communication efficiency over FL improves with the number of clients. Besides, the performance of SFL with privacy and robustness measures is further evaluated under extended experimental settings.
APA, Harvard, Vancouver, ISO, and other styles
4

Balakrishnan, Charumathi, and Mangaiyarkarasi Thiagarajan. "CREDIT RISK MODELLING FOR INDIAN DEBT SECURITIES USING MACHINE LEARNING." Buletin Ekonomi Moneter dan Perbankan 24 (March 8, 2021): 107–28. http://dx.doi.org/10.21098/bemp.v24i0.1401.

Full text
Abstract:
We develop a new credit risk model for Indian debt securities rated by major credit rating agencies in India using the ordinal logistic regression (OLR). The robustness of the model is tested by comparing it with classical models available for ratings prediction. We improved the model’s accuracy by using machine learning techniques, such as the artificial neural networks (ANN), support vector machines (SVM) and random forest (RF). We found that the accuracy of our model has improved from 68% using OLR to 82% when using ANN and above 90% when using SVM and RF.
APA, Harvard, Vancouver, ISO, and other styles
5

Nguyen, Ngoc-Kim-Khanh, Quang Nguyen, Hai-Ha Pham, Thi-Trang Le, Tuan-Minh Nguyen, Davide Cassi, Francesco Scotognella, Roberto Alfierif, and Michele Bellingeri. "Predicting the Robustness of Large Real-World Social Networks Using a Machine Learning Model." Complexity 2022 (November 9, 2022): 1–16. http://dx.doi.org/10.1155/2022/3616163.

Full text
Abstract:
Computing the robustness of a network, i.e., the capacity of a network holding its main functionality when a proportion of its nodes/edges are damaged, is useful in many real applications. The Monte Carlo numerical simulation is the commonly used method to compute network robustness. However, it has a very high computational cost, especially for large networks. Here, we propose a methodology such that the robustness of large real-world social networks can be predicted using machine learning models, which are pretrained using existing datasets. We demonstrate this approach by simulating two effective node attack strategies, i.e., the recalculated degree (RD) and initial betweenness (IB) node attack strategies, and predicting network robustness by using two machine learning models, multiple linear regression (MLR) and the random forest (RF) algorithm. We use the classic network robustness metric R as a model response and 8 network structural indicators (NSI) as predictor variables and trained over a large dataset of 48 real-world social networks, whose maximum number of nodes is 265,000. We found that the RF model can predict network robustness with a mean squared error (RMSE) of 0.03 and is 30% better than the MLR model. Among the results, we found that the RD strategy has more efficacy than IB for attacking real-world social networks. Furthermore, MLR indicates that the most important factors to predict network robustness are the scale-free exponent α and the average node degree <k>. On the contrary, the RF indicates that degree assortativity a, the global closeness, and the average node degree <k> are the most important factors. This study shows that machine learning models can be a promising way to infer social network robustness.
APA, Harvard, Vancouver, ISO, and other styles
6

Wu, Zhijing, and Hua Xu. "A Multi-Task Learning Machine Reading Comprehension Model for Noisy Document (Student Abstract)." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 10 (April 3, 2020): 13963–64. http://dx.doi.org/10.1609/aaai.v34i10.7254.

Full text
Abstract:
Current neural models for Machine Reading Comprehension (MRC) have achieved successful performance in recent years. However, the model is too fragile and lack robustness to tackle the imperceptible adversarial perturbations to the input. In this work, we propose a multi-task learning MRC model with a hierarchical knowledge enrichment to further improve the robustness for noisy document. Our model follows a typical encode-align-decode framework. Additionally, we apply a hierarchical method of adding background knowledge into the model from coarse-to-fine to enhance the language representations. Besides, we optimize our model by jointly training the answer span and unanswerability prediction, aiming to improve the robustness to noise. Experiment results on benchmark datasets confirm the superiority of our method, and our method can achieve competitive performance compared with other strong baselines.
APA, Harvard, Vancouver, ISO, and other styles
7

Chuah, Joshua, Uwe Kruger, Ge Wang, Pingkun Yan, and Juergen Hahn. "Framework for Testing Robustness of Machine Learning-Based Classifiers." Journal of Personalized Medicine 12, no. 8 (August 14, 2022): 1314. http://dx.doi.org/10.3390/jpm12081314.

Full text
Abstract:
There has been a rapid increase in the number of artificial intelligence (AI)/machine learning (ML)-based biomarker diagnostic classifiers in recent years. However, relatively little work has focused on assessing the robustness of these biomarkers, i.e., investigating the uncertainty of the AI/ML models that these biomarkers are based upon. This paper addresses this issue by proposing a framework to evaluate the already-developed classifiers with regard to their robustness by focusing on the variability of the classifiers’ performance and changes in the classifiers’ parameter values using factor analysis and Monte Carlo simulations. Specifically, this work evaluates (1) the importance of a classifier’s input features and (2) the variability of a classifier’s output and model parameter values in response to data perturbations. Additionally, it was found that one can estimate a priori how much replacement noise a classifier can tolerate while still meeting accuracy goals. To illustrate the evaluation framework, six different AI/ML-based biomarkers are developed using commonly used techniques (linear discriminant analysis, support vector machines, random forest, partial-least squares discriminant analysis, logistic regression, and multilayer perceptron) for a metabolomics dataset involving 24 measured metabolites taken from 159 study participants. The framework was able to correctly predict which of the classifiers should be less robust than others without recomputing the classifiers itself, and this prediction was then validated in a detailed analysis.
APA, Harvard, Vancouver, ISO, and other styles
8

Sepulveda, Natalia Espinoza, and Jyoti Sinha. "Parameter Optimisation in the Vibration-Based Machine Learning Model for Accurate and Reliable Faults Diagnosis in Rotating Machines." Machines 8, no. 4 (October 23, 2020): 66. http://dx.doi.org/10.3390/machines8040066.

Full text
Abstract:
Artificial intelligence (AI)-based machine learning (ML) models seem to be the future for most of the applications. Recent research effort has also been made on the application of these AI and ML methods in the vibration-based faults diagnosis (VFD) in rotating machines. Several research studies have been published over the last decade on this topic. However, most of the studies are data driven, and the vibration-based ML (VML) model is generally developed on a typical machine. The developed VML model may not predict faults accurately if applied on other identical machines or a machine with different operation conditions or both. Therefore, the current research is on the development of a VML model by optimising the vibration parameters based on the dynamics of the machine. The developed model is then blindly tested at different machine operation conditions to show the robustness and reliability of the proposed VML model.
APA, Harvard, Vancouver, ISO, and other styles
9

Zhang, Lingwen, Ning Xiao, Wenkao Yang, and Jun Li. "Advanced Heterogeneous Feature Fusion Machine Learning Models and Algorithms for Improving Indoor Localization." Sensors 19, no. 1 (January 2, 2019): 125. http://dx.doi.org/10.3390/s19010125.

Full text
Abstract:
In the era of the Internet of Things and Artificial Intelligence, the Wi-Fi fingerprinting-based indoor positioning system (IPS) has been recognized as the most promising IPS for various applications. Fingerprinting-based algorithms critically rely on a fingerprint database built from machine learning methods. However, currently methods are based on single-feature Received Signal Strength (RSS), which is extremely unstable in performance in terms of precision and robustness. The reason for this is that single feature machines cannot capture the complete channel characteristics and are susceptible to interference. The objective of this paper is to exploit the Time of Arrival (TOA) feature and propose a heterogeneous features fusion model to enhance the precision and robustness of indoor positioning. Several challenges are addressed: (1) machine learning models based on heterogeneous features, (2) the optimization of algorithms for high precision and robustness, and (3) computational complexity. This paper provides several heterogeneous features fusion-based localization models. Their effectiveness and efficiency are thoroughly compared with state-of-the-art methods.
APA, Harvard, Vancouver, ISO, and other styles
10

Drews, Samuel, Aws Albarghouthi, and Loris D'Antoni. "Proving Data-Poisoning Robustness in Decision Trees." Communications of the ACM 66, no. 2 (January 20, 2023): 105–13. http://dx.doi.org/10.1145/3576894.

Full text
Abstract:
Machine learning models are brittle, and small changes in the training data can result in different predictions. We study the problem of proving that a prediction is robust to data poisoning , where an attacker can inject a number of malicious elements into the training set to influence the learned model. We target decision tree models, a popular and simple class of machine learning models that underlies many complex learning techniques. We present a sound verification technique based on abstract interpretation and implement it in a tool called Antidote. Antidote abstractly trains decision trees for an intractably large space of possible poisoned datasets. Due to the soundness of our abstraction, Antidote can produce proofs that, for a given input, the corresponding prediction would not have changed had the training set been tampered with or not. We demonstrate the effectiveness of Antidote on a number of popular datasets.
APA, Harvard, Vancouver, ISO, and other styles
11

Niroumand-Jadidi, M., and F. Bovolo. "TEMPORALLY TRANSFERABLE MACHINE LEARNING MODEL FOR TOTAL SUSPENDED MATTER RETRIEVAL FROM SENTINEL-2." ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences V-3-2022 (May 17, 2022): 339–45. http://dx.doi.org/10.5194/isprs-annals-v-3-2022-339-2022.

Full text
Abstract:
Abstract. The empirical (regression-based) models have long been used for retrieving water quality parameters from optical imagery by training a model between image spectra and collocated in-situ data. However, a need clearly exists to examine and enhance the temporal transferability of models. The performance of a model trained in a specific period can deteriorate when applied at another time due to variations in the composition of constituents, atmospheric conditions, and sun glint. In this study, we propose a machine learning approach that trains a neural network using samples distributed in space and time, enabling the temporal robustness of the model. We explore the temporal transferability of the proposed neural network and standard band ratio models in retrieving total suspended matter (TSM) from Sentinel-2 imagery in San Francisco Bay. Multitemporal Sentinel-2 imagery and in-situ data are used to train the models. The transferability of models is then examined by estimating the TSM for imagery acquired after the training period. In addition, we assess the robustness of the models concerning the sun glint correction. The results imply that the neural network-based model is temporally transferable (R2 &amp;approx; 0.75; RMSE &amp;approx; 7 g/m3 for retrievals up to 70 g/m3) and is minimally impacted by the sun glint correction. Conversely, the ratio model showed relatively poor temporal robustness with high sensitivity to the glint correction.
APA, Harvard, Vancouver, ISO, and other styles
12

Liu, Molei, Yi Zhang, and Doudou Zhou. "Double/debiased machine learning for logistic partially linear model." Econometrics Journal 24, no. 3 (June 11, 2021): 559–88. http://dx.doi.org/10.1093/ectj/utab019.

Full text
Abstract:
Summary We propose double/debiased machine learning approaches to infer a parametric component of a logistic partially linear model. Our framework is based on a Neyman orthogonal score equation consisting of two nuisance models for the nonparametric component of the logistic model and conditional mean of the exposure with the control group. To estimate the nuisance models, we separately consider the use of high dimensional (HD) sparse regression and (nonparametric) machine learning (ML) methods. In the HD case, we derive certain moment equations to calibrate the first order bias of the nuisance models, which preserves the model double robustness property. In the ML case, we handle the nonlinearity of the logit link through a novel and easy-to-implement ‘full model refitting’ procedure. We evaluate our methods through simulation and apply them in assessing the effect of the emergency contraceptive pill on early gestation and new births based on a 2008 policy reform in Chile.
APA, Harvard, Vancouver, ISO, and other styles
13

Schröder, Laura, Nikolay Krasimirov Dimitrov, David Robert Verelst, and John Aasted Sørensen. "Using Transfer Learning to Build Physics-Informed Machine Learning Models for Improved Wind Farm Monitoring." Energies 15, no. 2 (January 13, 2022): 558. http://dx.doi.org/10.3390/en15020558.

Full text
Abstract:
This paper introduces a novel, transfer-learning-based approach to include physics into data-driven normal behavior monitoring models which are used for detecting turbine anomalies. For this purpose, a normal behavior model is pretrained on a large simulation database and is recalibrated on the available SCADA data via transfer learning. For two methods, a feed-forward artificial neural network (ANN) and an autoencoder, it is investigated under which conditions it can be helpful to include simulations into SCADA-based monitoring systems. The results show that when only one month of SCADA data is available, both the prediction accuracy as well as the prediction robustness of an ANN are significantly improved by adding physics constraints from a pretrained model. As the autoencoder reconstructs the power from itself, it is already able to accurately model the normal behavior power. Therefore, including simulations into the model does not improve its prediction performance and robustness significantly. The validation of the physics-informed ANN on one month of raw SCADA data shows that it is able to successfully detect a recorded blade angle anomaly with an improved precision due to fewer false positives compared to its purely SCADA data-based counterpart.
APA, Harvard, Vancouver, ISO, and other styles
14

Zhao, Yue, Xuejian Wang, Cheng Cheng, and Xueying Ding. "Combining Machine Learning Models Using combo Library." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 09 (April 3, 2020): 13648–49. http://dx.doi.org/10.1609/aaai.v34i09.7111.

Full text
Abstract:
Model combination, often regarded as a key sub-field of ensemble learning, has been widely used in both academic research and industry applications. To facilitate this process, we propose and implement an easy-to-use Python toolkit, combo, to aggregate models and scores under various scenarios, including classification, clustering, and anomaly detection. In a nutshell, combo provides a unified and consistent way to combine both raw and pretrained models from popular machine learning libraries, e.g., scikit-learn, XGBoost, and LightGBM. With accessibility and robustness in mind, combo is designed with detailed documentation, interactive examples, continuous integration, code coverage, and maintainability check; it can be installed easily through Python Package Index (PyPI) or {https://github.com/yzhao062/combo}.
APA, Harvard, Vancouver, ISO, and other styles
15

Spelda, Petr, and Vit Stritecky. "Human Induction in Machine Learning." ACM Computing Surveys 54, no. 3 (June 2021): 1–18. http://dx.doi.org/10.1145/3444691.

Full text
Abstract:
As our epistemic ambitions grow, the common and scientific endeavours are becoming increasingly dependent on Machine Learning (ML). The field rests on a single experimental paradigm, which consists of splitting the available data into a training and testing set and using the latter to measure how well the trained ML model generalises to unseen samples. If the model reaches acceptable accuracy, then an a posteriori contract comes into effect between humans and the model, supposedly allowing its deployment to target environments. Yet the latter part of the contract depends on human inductive predictions or generalisations, which infer a uniformity between the trained ML model and the targets. The article asks how we justify the contract between human and machine learning. It is argued that the justification becomes a pressing issue when we use ML to reach “elsewhere” in space and time or deploy ML models in non-benign environments. The article argues that the only viable version of the contract can be based on optimality (instead of on reliability, which cannot be justified without circularity) and aligns this position with Schurz's optimality justification. It is shown that when dealing with inaccessible/unstable ground-truths (“elsewhere” and non-benign targets), the optimality justification undergoes a slight change, which should reflect critically on our epistemic ambitions. Therefore, the study of ML robustness should involve not only heuristics that lead to acceptable accuracies on testing sets. The justification of human inductive predictions or generalisations about the uniformity between ML models and targets should be included as well. Without it, the assumptions about inductive risk minimisation in ML are not addressed in full.
APA, Harvard, Vancouver, ISO, and other styles
16

Li, Yibo, Chao Liu, Senyue Zhang, Wenan Tan, and Yanyan Ding. "Reproducing Polynomial Kernel Extreme Learning Machine." Journal of Advanced Computational Intelligence and Intelligent Informatics 21, no. 5 (September 20, 2017): 795–802. http://dx.doi.org/10.20965/jaciii.2017.p0795.

Full text
Abstract:
Conventional kernel support vector machine (KSVM) has the problem of slow training speed, and single kernel extreme learning machine (KELM) also has some performance limitations, for which this paper proposes a new combined KELM model that build by the polynomial kernel and reproducing kernel on Sobolev Hilbert space. This model combines the advantages of global and local kernel function and has fast training speed. At the same time, an efficient optimization algorithm called cuckoo search algorithm is adopted to avoid blindness and inaccuracy in parameter selection. Experiments were performed on bi-spiral benchmark dataset, Banana dataset, as well as a number of classification and regression datasets from the UCI benchmark repository illustrate the feasibility of the proposed model. It achieves the better robustness and generalization performance when compared to other conventional KELM and KSVM, which demonstrates its effectiveness and usefulness.
APA, Harvard, Vancouver, ISO, and other styles
17

Espinheira, Patrícia, Luana da Silva, Alisson Silva, and Raydonal Ospina. "Model Selection Criteria on Beta Regression for Machine Learning." Machine Learning and Knowledge Extraction 1, no. 1 (February 8, 2019): 427–49. http://dx.doi.org/10.3390/make1010026.

Full text
Abstract:
Beta regression models are a class of supervised learning tools for regression problems with univariate and limited response. Current fitting procedures for beta regression require variable selection based on (potentially problematic) information criteria. We propose model selection criteria that take into account the leverage, residuals, and influence of the observations, both to systematic linear and nonlinear components. To that end, we propose a Predictive Residual Sum of Squares (PRESS)-like machine learning tool and a prediction coefficient, namely P 2 statistic, as a computational procedure. Monte Carlo simulation results on the finite sample behavior of prediction-based model selection criteria P 2 are provided. We also evaluated two versions of the R 2 criterion. Finally, applications to real data are presented. The new criterion proved to be crucial to choose models taking into account the robustness of the maximum likelihood estimation procedure in the presence of influential cases.
APA, Harvard, Vancouver, ISO, and other styles
18

Papadopoulos, Pavlos, Oliver Thornewill von Essen, Nikolaos Pitropakis, Christos Chrysoulas, Alexios Mylonas, and William J. Buchanan. "Launching Adversarial Attacks against Network Intrusion Detection Systems for IoT." Journal of Cybersecurity and Privacy 1, no. 2 (April 23, 2021): 252–73. http://dx.doi.org/10.3390/jcp1020014.

Full text
Abstract:
As the internet continues to be populated with new devices and emerging technologies, the attack surface grows exponentially. Technology is shifting towards a profit-driven Internet of Things market where security is an afterthought. Traditional defending approaches are no longer sufficient to detect both known and unknown attacks to high accuracy. Machine learning intrusion detection systems have proven their success in identifying unknown attacks with high precision. Nevertheless, machine learning models are also vulnerable to attacks. Adversarial examples can be used to evaluate the robustness of a designed model before it is deployed. Further, using adversarial examples is critical to creating a robust model designed for an adversarial environment. Our work evaluates both traditional machine learning and deep learning models’ robustness using the Bot-IoT dataset. Our methodology included two main approaches. First, label poisoning, used to cause incorrect classification by the model. Second, the fast gradient sign method, used to evade detection measures. The experiments demonstrated that an attacker could manipulate or circumvent detection with significant probability.
APA, Harvard, Vancouver, ISO, and other styles
19

Yanjun Li, Yanjun Li, Huan Huang Yanjun Li, Qiang Geng Huan Huang, Xinwei Guo Qiang Geng, and Yuyu Yuan Xinwei Guo. "Fairness Measures of Machine Learning Models in Judicial Penalty Prediction." 網際網路技術學刊 23, no. 5 (September 2022): 1109–16. http://dx.doi.org/10.53106/160792642022092305019.

Full text
Abstract:
<p>Machine learning (ML) has been widely adopted in many software applications across domains. However, accompanying the outstanding performance, the behaviors of the ML models, which are essentially a kind of black-box software, could be unfair and hard to understand in many cases. In our human-centered society, an unfair decision could potentially damage human value, even causing severe social consequences, especially in decision-critical scenarios such as legal judgment. Although some existing works investigated the ML models in terms of robustness, accuracy, security, privacy, quality, etc., the study on the fairness of ML is still in the early stage. In this paper, we first proposed a set of fairness metrics for ML models from different perspectives. Based on this, we performed a comparative study on the fairness of existing widely used classic ML and deep learning models in the domain of real-world judicial judgments. The experiment results reveal that the current state-of-the-art ML models could still raise concerns for unfair decision-making. The ML models with high accuracy and fairness are urgently demanding.</p> <p>&nbsp;</p>
APA, Harvard, Vancouver, ISO, and other styles
20

Saarela, Mirka, and Lilia Geogieva. "Robustness, Stability, and Fidelity of Explanations for a Deep Skin Cancer Classification Model." Applied Sciences 12, no. 19 (September 23, 2022): 9545. http://dx.doi.org/10.3390/app12199545.

Full text
Abstract:
Skin cancer is one of the most prevalent of all cancers. Because of its being widespread and externally observable, there is a potential that machine learning models integrated into artificial intelligence systems will allow self-screening and automatic analysis in the future. Especially, the recent success of various deep machine learning models shows promise that, in the future, patients could self-analyse their external signs of skin cancer by uploading pictures of these signs to an artificial intelligence system, which runs such a deep learning model and returns the classification results. However, both patients and dermatologists, who might use such a system to aid their work, need to know why the system has made a particular decision. Recently, several explanation techniques for the deep learning algorithm’s decision-making process have been introduced. This study compares two popular local explanation techniques (integrated gradients and local model-agnostic explanations) for image data on top of a well-performing (80% accuracy) deep learning algorithm trained on the HAM10000 dataset, a large public collection of dermatoscopic images. Our results show that both methods have full local fidelity. However, the integrated gradients explanations perform better with regard to quantitative evaluation metrics (stability and robustness), while the model-agnostic method seem to provide more intuitive explanations. We conclude that there is still a long way before such automatic systems can be used reliably in practice.
APA, Harvard, Vancouver, ISO, and other styles
21

Mello, Flávio Luis de. "A Survey on Machine Learning Adversarial Attacks." Journal of Information Security and Cryptography (Enigma) 7, no. 1 (January 20, 2020): 1–7. http://dx.doi.org/10.17648/jisc.v7i1.76.

Full text
Abstract:
It is becoming notorious several types of adversaries based on their threat model leverage vulnerabilities to compromise a machine learning system. Therefore, it is important to provide robustness to machine learning algorithms and systems against these adversaries. However, there are only a few strong countermeasures, which can be used in all types of attack scenarios to design a robust artificial intelligence system. This paper is structured and comprehensive overview of the research on attacks to machine learning systems and it tries to call the attention from developers and software houses to the security issues concerning machine learning.
APA, Harvard, Vancouver, ISO, and other styles
22

Lai and Tsai. "Improving GIS-based Landslide Susceptibility Assessments with Multi-temporal Remote Sensing and Machine Learning." Sensors 19, no. 17 (August 27, 2019): 3717. http://dx.doi.org/10.3390/s19173717.

Full text
Abstract:
This study developed a systematic approach with machine learning (ML) to apply the satellite remote sensing images, geographic information system (GIS) datasets, and spatial analysis for multi-temporal and event-based landslide susceptibility assessments at a regional scale. Random forests (RF) algorithm, one of the ML-based methods, was selected to construct the landslide susceptibility models. Different ratios of landslide and non-landslide samples were considered in the experiments. This study also employed a cost-sensitive analysis to adjust the decision boundary of the developed RF models with unbalanced sample ratios to improve the prediction results. Two strategies were investigated for model verification, namely space- and time-robustness. The space-robustness verification was designed for separating samples into training and examining data based on a single event or the same dataset. The time-robustness verification was designed for predicting subsequent landslide events by constructing a landslide susceptibility model based on a specific event or period. A total of 14 GIS-based landslide-related factors were used and derived from the spatial analyses. The developed landslide susceptibility models were tested in a watershed region in northern Taiwan with a landslide inventory of changes detected through multi-temporal satellite images and verified through field investigation. To further examine the developed models, the landslide susceptibility distributions of true occurrence samples and the generated landslide susceptibility maps were compared. The experiments demonstrated that the proposed method can provide more reasonable results, and the accuracies were found to be higher than 93% and 75% in most cases for space- and time-robustness verifications, respectively. In addition, the mapping results revealed that the multi-temporal models did not seem to be affected by the sample ratios included in the analyses.
APA, Harvard, Vancouver, ISO, and other styles
23

Mei, Wenjuan, Zhen Liu, Yuanzhang Su, Li Du, and Jianguo Huang. "Evolved-Cooperative Correntropy-Based Extreme Learning Machine for Robust Prediction." Entropy 21, no. 9 (September 19, 2019): 912. http://dx.doi.org/10.3390/e21090912.

Full text
Abstract:
In recent years, the correntropy instead of the mean squared error has been widely taken as a powerful tool for enhancing the robustness against noise and outliers by forming the local similarity measurements. However, most correntropy-based models either have too simple descriptions of the correntropy or require too many parameters to adjust in advance, which is likely to cause poor performance since the correntropy fails to reflect the probability distributions of the signals. Therefore, in this paper, a novel correntropy-based extreme learning machine (ELM) called ECC-ELM has been proposed to provide a more robust training strategy based on the newly developed multi-kernel correntropy with the parameters that are generated using cooperative evolution. To achieve an accurate description of the correntropy, the method adopts a cooperative evolution which optimizes the bandwidths by switching delayed particle swarm optimization (SDPSO) and generates the corresponding influence coefficients that minimizes the minimum integrated error (MIE) to adaptively provide the best solution. The simulated experiments and real-world applications show that cooperative evolution can achieve the optimal solution which provides an accurate description on the probability distribution of the current error in the model. Therefore, the multi-kernel correntropy that is built with the optimal solution results in more robustness against the noise and outliers when training the model, which increases the accuracy of the predictions compared with other methods.
APA, Harvard, Vancouver, ISO, and other styles
24

Yang, Rui, Yongbao Liu, Xing He, and Zhimeng Liu. "Gas Turbine Model Identification Based on Online Sequential Regularization Extreme Learning Machine with a Forgetting Factor." Energies 16, no. 1 (December 27, 2022): 304. http://dx.doi.org/10.3390/en16010304.

Full text
Abstract:
Due to the advantages of high convergence accuracy, fast training speed, and good generalization performance, the extreme learning machine is widely used in model identification. However, a gas turbine is a complex nonlinear system, and its sampling data are often time-sensitive and have measurement noise. This article proposes an online sequential regularization extreme learning machine algorithm based on the forgetting factor (FOS_RELM) to improve gas turbine identification performance. The proposed FOS_RELM not only retains the advantages of the extreme learning machine algorithm but also enhances the learning effect by rapidly discarding obsolete data during the learning process and improves the anti-interference performance by using the regularization principle. A detailed performance comparison of the FOS_RELM with the extreme learning machine algorithm and regularized extreme learning machine algorithm is carried out in the model identification of a gas turbine. The results show that the FOS_RELM has higher accuracy and better robustness than the extreme learning machine algorithm and regularized extreme learning machine algorithm. All in all, the proposed algorithm provides a candidate technique for modeling actual gas turbine units.
APA, Harvard, Vancouver, ISO, and other styles
25

Zhou, Zhiyu, Xu Gao, Jianxin Zhang, Zefei Zhu, and Xudong Hu. "A novel hybrid model using the rotation forest-based differential evolution online sequential extreme learning machine for illumination correction of dyed fabrics." Textile Research Journal 89, no. 7 (March 20, 2018): 1180–97. http://dx.doi.org/10.1177/0040517518764020.

Full text
Abstract:
This study proposes an ensemble differential evolution online sequential extreme learning machine (DE-OSELM) for textile image illumination correction based on the rotation forest framework. The DE-OSELM solves the inaccuracy and long training time problems associated with traditional illumination correction algorithms. First, the Grey–Edge framework is used to extract the low-dimensional and efficient image features as online sequential extreme learning machine (OSELM) input vectors to improve the training and learning speed of the OSELM. Since the input weight and hidden-layer bias of OSELMs are randomly obtained, the OSELM algorithm has poor prediction accuracy and low robustness. To overcome this shortcoming, a differential evolution algorithm that has the advantages of good global search ability and robustness is used to optimize the input weight and hidden-layer bias of the DE-OSELM. To further improve the generalization ability and robustness of the illumination correction model, the rotation forest algorithm is used as the ensemble framework, and the DE-OSELM is used as the base learner to replace the regression tree algorithm in the original rotation forest algorithm. Then, the obtained multiple different DE-OSELM learners are aggregated to establish the prediction model. The experimental results show that compared with the textile color correction algorithm based on the support vector regression and extreme learning machine algorithms, the ensemble illumination correction method achieves high prediction accuracy, strong robustness, and good generalization ability.
APA, Harvard, Vancouver, ISO, and other styles
26

Lamilla, Erick, Christian Sacarelo, Manuel S. Alvarez-Alvarado, Arturo Pazmino, and Peter Iza. "Optical Encoding Model Based on Orbital Angular Momentum Powered by Machine Learning." Sensors 23, no. 5 (March 2, 2023): 2755. http://dx.doi.org/10.3390/s23052755.

Full text
Abstract:
Based on orbital angular momentum (OAM) properties of Laguerre–Gaussian beams LG(p,ℓ), a robust optical encoding model for efficient data transmission applications is designed. This paper presents an optical encoding model based on an intensity profile generated by a coherent superposition of two OAM-carrying Laguerre–Gaussian modes and a machine learning detection method. In the encoding process, the intensity profile for data encoding is generated based on the selection of p and ℓ indices, while the decoding process is performed using a support vector machine (SVM) algorithm. Two different decoding models based on an SVM algorithm are tested to verify the robustness of the optical encoding model, finding a BER =10−9 for 10.2 dB of signal-to-noise ratio in one of the SVM models.
APA, Harvard, Vancouver, ISO, and other styles
27

Zheng, Jiao, and Zhengyu Yu. "A Novel Machine Learning-Based Systolic Blood Pressure Predicting Model." Journal of Nanomaterials 2021 (June 7, 2021): 1–8. http://dx.doi.org/10.1155/2021/9934998.

Full text
Abstract:
Blood pressure (BP) is a vital biomedical feature for diagnosing hypertension and cardiovascular diseases. Traditionally, it is measured by cuff-based equipment, e.g., sphygmomanometer; the measurement is discontinued and uncomfortable. A cuff-less method based on different signals, electrocardiogram (ECG) and photoplethysmography (PPG), is proposed recently. However, this method is costly and inconvenient due to the collections of multisensors. In this paper, a novel machine learning-based systolic blood pressure (SBP) predicting model is proposed. The model was evaluated by clinical and lifestyle features (gender, marital status, smoking status, age, weight, etc.). Different machine learning algorithms and different percentage of training, validation, and testing were evaluated to optimize the model accuracy. Results were validated to increase the accuracy and robustness of the model. The performance of our model met both the level of grade A (British Hypertension Society (BHS) standard) and the American National Standard from the Association for the Advancement of Medical Instrumentation (AAMI) for SBP estimation.
APA, Harvard, Vancouver, ISO, and other styles
28

Ning, Kun-Peng, Lue Tao, Songcan Chen, and Sheng-Jun Huang. "Improving Model Robustness by Adaptively Correcting Perturbation Levels with Active Queries." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 10 (May 18, 2021): 9161–69. http://dx.doi.org/10.1609/aaai.v35i10.17106.

Full text
Abstract:
In addition to high accuracy, robustness is becoming increasingly important for machine learning models in various applications. Recently, much research has been devoted to improving the model robustness by training with noise perturbations. Most existing studies assume a fixed perturbation level for all training examples, which however hardly holds in real tasks. In fact, excessive perturbations may destroy the discriminative content of an example, while deficient perturbations may fail to provide helpful information for improving the robustness. Motivated by this observation, we propose to adaptively adjust the perturbation levels for each example in the training process. Specifically, a novel active learning framework is proposed to allow the model interactively querying the correct perturbation level from human experts. By designing a cost-effective sampling strategy along with a new query type, the robustness can be significantly improved with a few queries. Both theoretical analysis and experimental studies validate the effectiveness of the proposed approach.
APA, Harvard, Vancouver, ISO, and other styles
29

Diakonikolas, Ilias, Gautam Kamath, Daniel M. Kane, Jerry Li, Ankur Moitra, and Alistair Stewart. "Robustness meets algorithms." Communications of the ACM 64, no. 5 (May 2021): 107–15. http://dx.doi.org/10.1145/3453935.

Full text
Abstract:
In every corner of machine learning and statistics, there is a need for estimators that work not just in an idealized model, but even when their assumptions are violated. Unfortunately, in high dimensions, being provably robust and being efficiently computable are often at odds with each other. We give the first efficient algorithm for estimating the parameters of a high-dimensional Gaussian that is able to tolerate a constant fraction of corruptions that is independent of the dimension. Prior to our work, all known estimators either needed time exponential in the dimension to compute or could tolerate only an inverse-polynomial fraction of corruptions. Not only does our algorithm bridge the gap between robustness and algorithms, but also it turns out to be highly practical in a variety of settings.
APA, Harvard, Vancouver, ISO, and other styles
30

Simester, Duncan, Artem Timoshenko, and Spyros I. Zoumpoulis. "Targeting Prospective Customers: Robustness of Machine-Learning Methods to Typical Data Challenges." Management Science 66, no. 6 (June 2020): 2495–522. http://dx.doi.org/10.1287/mnsc.2019.3308.

Full text
Abstract:
We investigate how firms can use the results of field experiments to optimize the targeting of promotions when prospecting for new customers. We evaluate seven widely used machine-learning methods using a series of two large-scale field experiments. The first field experiment generates a common pool of training data for each of the seven methods. We then validate the seven optimized policies provided by each method together with uniform benchmark policies in a second field experiment. The findings not only compare the performance of the targeting methods, but also demonstrate how well the methods address common data challenges. Our results reveal that when the training data are ideal, model-driven methods perform better than distance-driven methods and classification methods. However, the performance advantage vanishes in the presence of challenges that affect the quality of the training data, including the extent to which the training data captures details of the implementation setting. The challenges we study are covariate shift, concept shift, information loss through aggregation, and imbalanced data. Intuitively, the model-driven methods make better use of the information available in the training data, but the performance of these methods is more sensitive to deterioration in the quality of this information. The classification methods we tested performed relatively poorly. We explain the poor performance of the classification methods in our setting and describe how the performance of these methods could be improved. This paper was accepted by Matthew Shum, marketing.
APA, Harvard, Vancouver, ISO, and other styles
31

Alharbi, Amal H., Aravinda C. V, Meng Lin, B. Ashwini, Mohamed Yaseen Jabarulla, and Mohd Asif Shah. "Detection of Peripheral Malarial Parasites in Blood Smears Using Deep Learning Models." Computational Intelligence and Neuroscience 2022 (May 24, 2022): 1–11. http://dx.doi.org/10.1155/2022/3922763.

Full text
Abstract:
Due to the plasmodium parasite, malaria is transmitted mostly through red blood cells. Manually counting blood cells is extremely time consuming and tedious. In a recommendation for the advanced technology stage and analysis of malarial disease, the performance of the XG-Boost, SVM, and neural networks is compared. In comparison to machine learning models, convolutional neural networks provide reliable results when analyzing and recognizing the same datasets. To reduce discrepancies and improve robustness and generalization, we developed a model that analyzes blood samples to determine whether the cells are parasitized or not. Experiments were conducted on 13,750 parasitized and 13,750 parasitic samples. Support vector machines achieved 94% accuracy, XG-Boost models achieved 90% accuracy, and neural networks achieved 80% accuracy. Among these three models, the support vector machine was the most accurate at distinguishing parasitized cells from uninfected ones. An accuracy rate of 97% was achieved by the convolution neural network in recognizing the samples. The deep learning model is useful for decision making because of its better accuracy.
APA, Harvard, Vancouver, ISO, and other styles
32

Gröhl, Janek, Thomas Kirchner, Tim Adler, and Lena Maier-Hein. "Confidence Estimation for Machine Learning-Based Quantitative Photoacoustics." Journal of Imaging 4, no. 12 (December 10, 2018): 147. http://dx.doi.org/10.3390/jimaging4120147.

Full text
Abstract:
In medical applications, the accuracy and robustness of imaging methods are of crucial importance to ensure optimal patient care. While photoacoustic imaging (PAI) is an emerging modality with promising clinical applicability, state-of-the-art approaches to quantitative photoacoustic imaging (qPAI), which aim to solve the ill-posed inverse problem of recovering optical absorption from the measurements obtained, currently cannot comply with these high standards. This can be attributed to the fact that existing methods often rely on several simplifying a priori assumptions of the underlying physical tissue properties or cannot deal with realistic noise levels. In this manuscript, we address this issue with a new method for estimating an indicator of the uncertainty of an estimated optical property. Specifically, our method uses a deep learning model to compute error estimates for optical parameter estimations of a qPAI algorithm. Functional tissue parameters, such as blood oxygen saturation, are usually derived by averaging over entire signal intensity-based regions of interest (ROIs). Therefore, we propose to reduce the systematic error of the ROI samples by additionally discarding those pixels for which our method estimates a high error and thus a low confidence. In silico experiments show an improvement in the accuracy of optical absorption quantification when applying our method to refine the ROI, and it might thus become a valuable tool for increasing the robustness of qPAI methods.
APA, Harvard, Vancouver, ISO, and other styles
33

Tedesco, Salvatore, Martina Andrulli, Markus Åkerlund Larsson, Daniel Kelly, Antti Alamäki, Suzanne Timmons, John Barton, Joan Condell, Brendan O’Flynn, and Anna Nordström. "Comparison of Machine Learning Techniques for Mortality Prediction in a Prospective Cohort of Older Adults." International Journal of Environmental Research and Public Health 18, no. 23 (December 4, 2021): 12806. http://dx.doi.org/10.3390/ijerph182312806.

Full text
Abstract:
As global demographics change, ageing is a global phenomenon which is increasingly of interest in our modern and rapidly changing society. Thus, the application of proper prognostic indices in clinical decisions regarding mortality prediction has assumed a significant importance for personalized risk management (i.e., identifying patients who are at high or low risk of death) and to help ensure effective healthcare services to patients. Consequently, prognostic modelling expressed as all-cause mortality prediction is an important step for effective patient management. Machine learning has the potential to transform prognostic modelling. In this paper, results on the development of machine learning models for all-cause mortality prediction in a cohort of healthy older adults are reported. The models are based on features covering anthropometric variables, physical and lab examinations, questionnaires, and lifestyles, as well as wearable data collected in free-living settings, obtained for the “Healthy Ageing Initiative” study conducted on 2291 recruited participants. Several machine learning techniques including feature engineering, feature selection, data augmentation and resampling were investigated for this purpose. A detailed empirical comparison of the impact of the different techniques is presented and discussed. The achieved performances were also compared with a standard epidemiological model. This investigation showed that, for the dataset under consideration, the best results were achieved with Random UnderSampling in conjunction with Random Forest (either with or without probability calibration). However, while including probability calibration slightly reduced the average performance, it increased the model robustness, as indicated by the lower 95% confidence intervals. The analysis showed that machine learning models could provide comparable results to standard epidemiological models while being completely data-driven and disease-agnostic, thus demonstrating the opportunity for building machine learning models on health records data for research and clinical practice. However, further testing is required to significantly improve the model performance and its robustness.
APA, Harvard, Vancouver, ISO, and other styles
34

Dong, M., H. Qiu, H. Wang, P. Zhi, and Z. Xu. "SONAR IMAGE RECOGNITION BASED ON MACHINE LEARNING FRAMEWORK." International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences XLVI-3/W1-2022 (April 22, 2022): 45–51. http://dx.doi.org/10.5194/isprs-archives-xlvi-3-w1-2022-45-2022.

Full text
Abstract:
Abstract. In order to improve the robustness and generalization ability of model recognition, sonar images are enhanced by preprocessing such as conversion coordinates, interpolation, denoising and enhancement, and the transfer learning method under the Caffe framework of MATLAB as an interface is used respectively (mainly composed of 8 layers of network structure, including 5 convolutional layers and 3 full chain layers) And the transfer learning method under the Python deep learning framework Inception-Resnet-v2 model for sonar image training and recognition. First of all, part of the sonar image dataset (derived from the 2021 National Robot Underwater Competition online competition data), using MATLAB as the interface Caffe framework, the sonar image is trained to obtain a training model, and then through parameter adjustment, the convolutional neural network model of sonar image automatic recognition is obtained, and the transfer learning method can use less sonar image data to solve the problem of insufficient sonar image data, and then make the training achieve a higher recognition rate in a shorter time. When the training data is randomly sampled for testing, the sonar data recognition model based on the Caffe framework is quickly and fully recognized, and the recognition rate can reach 92% when the test sample does not participate in the training of sonar image data; The transfer learning method under the Inception-Resnet-v2 model of python deep learning framework is used to train recognition on sonar images, and the recognition rate reaches about 97%. Using the two models in this paper, it is feasible to identify sonar images with high recognition rate, which is much higher than traditional recognition methods such as SVM classifiers, and the two sonar image data recognition models based on deep learning have better recognition ability and generalization ability.
APA, Harvard, Vancouver, ISO, and other styles
35

Jia, Jinyuan, Xiaoyu Cao, and Neil Zhenqiang Gong. "Intrinsic Certified Robustness of Bagging against Data Poisoning Attacks." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 9 (May 18, 2021): 7961–69. http://dx.doi.org/10.1609/aaai.v35i9.16971.

Full text
Abstract:
In a data poisoning attack, an attacker modifies, deletes, and/or inserts some training examples to corrupt the learnt machine learning model. Bootstrap Aggregating (bagging) is a well known ensemble learning method, which trains multiple base models on random subsamples of a training dataset using a base learning algorithm and uses majority vote to predict labels of testing examples. We prove the intrinsic certified robustness of bagging against data poisoning attacks. Specifically, we show that bagging with an arbitrary base learning algorithm provably predicts the same label for a testing example when the number of modified, deleted, and/or inserted training examples is bounded by a threshold. Moreover, we show that our derived threshold is tight if no assumptions on the base learning algorithm are made. We evaluate our method on MNIST and CIFAR10. For instance, our method achieves a certified accuracy of 91.1% on MNIST when arbitrarily modifying, deleting, and/or inserting 100 training examples. Code is available at: https://github.com/jjy1994/BaggingCertifyDataPoisoning.
APA, Harvard, Vancouver, ISO, and other styles
36

Ardabili, Sina F., Amir Mosavi, Pedram Ghamisi, Filip Ferdinand, Annamaria R. Varkonyi-Koczy, Uwe Reuter, Timon Rabczuk, and Peter M. Atkinson. "COVID-19 Outbreak Prediction with Machine Learning." Algorithms 13, no. 10 (October 1, 2020): 249. http://dx.doi.org/10.3390/a13100249.

Full text
Abstract:
Several outbreak prediction models for COVID-19 are being used by officials around the world to make informed decisions and enforce relevant control measures. Among the standard models for COVID-19 global pandemic prediction, simple epidemiological and statistical models have received more attention by authorities, and these models are popular in the media. Due to a high level of uncertainty and lack of essential data, standard models have shown low accuracy for long-term prediction. Although the literature includes several attempts to address this issue, the essential generalization and robustness abilities of existing models need to be improved. This paper presents a comparative analysis of machine learning and soft computing models to predict the COVID-19 outbreak as an alternative to susceptible–infected–recovered (SIR) and susceptible-exposed-infectious-removed (SEIR) models. Among a wide range of machine learning models investigated, two models showed promising results (i.e., multi-layered perceptron, MLP; and adaptive network-based fuzzy inference system, ANFIS). Based on the results reported here, and due to the highly complex nature of the COVID-19 outbreak and variation in its behavior across nations, this study suggests machine learning as an effective tool to model the outbreak. This paper provides an initial benchmarking to demonstrate the potential of machine learning for future research. This paper further suggests that a genuine novelty in outbreak prediction can be realized by integrating machine learning and SEIR models.
APA, Harvard, Vancouver, ISO, and other styles
37

Han, Bo, Bo He, Mengmeng Ma, Tingting Sun, Tianhong Yan, and Amaury Lendasse. "RMSE-ELM: Recursive Model Based Selective Ensemble of Extreme Learning Machines for Robustness Improvement." Mathematical Problems in Engineering 2014 (2014): 1–12. http://dx.doi.org/10.1155/2014/395686.

Full text
Abstract:
For blended data, the robustness of extreme learning machine (ELM) is so weak because the coefficients (weights and biases) of hidden nodes are set randomly and the noisy data exert a negative effect. To solve this problem, a new framework called “RMSE-ELM” is proposed in this paper. It is a two-layer recursive model. In the first layer, the framework trains lots of ELMs in different ensemble groups concurrently and then employs selective ensemble approach to pick out an optimal set of ELMs in each group, which can be merged into a large group of ELMs called candidate pool. In the second layer, selective ensemble approach is recursively used on candidate pool to acquire the final ensemble. In the experiments, we apply UCI blended datasets to confirm the robustness of our new approach in two key aspects (mean square error and standard deviation). The space complexity of our method is increased to some degree, but the result has shown that RMSE-ELM significantly improves robustness with a rapid learning speed compared to representative methods (ELM, OP-ELM, GASEN-ELM, GASEN-BP, and E-GASEN). It becomes a potential framework to solve robustness issue of ELM for high-dimensional blended data in the future.
APA, Harvard, Vancouver, ISO, and other styles
38

Gao, H., L. Jézéque, E. Cabrol, and B. Vitry. "Robust Design of Suspension System with Polynomial Chaos Expansion and Machine Learning." Science & Technique 19, no. 1 (February 5, 2020): 43–54. http://dx.doi.org/10.21122/2227-1031-2020-19-1-43-54.

Full text
Abstract:
During the early development of a new vehicle project, the uncertainty of parameters should be taken into consideration because the design may be perturbed due to real components’ complexity and manufacturing tolerances. Thus, the numerical validation of critical suspension specifications, such as durability and ride comfort should be carried out with random factors. In this article a multi-objective optimization methodology is proposed which involves the specification’s robustness as one of the optimization objectives. To predict the output variation from a given set of uncertain-but-bounded parameters proposed by optimization iterations, an adaptive chaos polynomial expansion (PCE) is applied to combine a local design of experiments with global response surfaces. Furthermore, in order to reduce the additional tests required for PCE construction, a machine learning algorithm based on inter-design correlation matrix firstly classifies the current design points through data mining and clustering. Then it learns how to predict the robustness of future optimized solutions with no extra simulations. At the end of the optimization, a Pareto front between specifications and their robustness can be obtained which represents the best compromises among objectives. The optimum set on the front is classified and can serve as a reference for future design. An example of a quarter car model has been tested for which the target is to optimize the global durability based on real road excitations. The statistical distribution of the parameters such as the trajectories and speeds is also taken into account. The result shows the natural incompatibility between the durability of the chassis and the robustness of this durability. Here the term robustness does not mean “strength”, but means that the performance is less sensitive to perturbations. In addition, a stochastic sampling verifies the good robustness prediction of PCE method and machine learning, based on a greatly reduced number of tests. This example demonstrates the effectiveness of the approach, in particular its ability to save computational costs for full vehicle simulation.
APA, Harvard, Vancouver, ISO, and other styles
39

Waldow, Fabian, Matthias Schnaubelt, Christopher Krauss, and Thomas Günter Fischer. "Machine Learning in Futures Markets." Journal of Risk and Financial Management 14, no. 3 (March 13, 2021): 119. http://dx.doi.org/10.3390/jrfm14030119.

Full text
Abstract:
In this paper, we demonstrate how a well-established machine learning-based statistical arbitrage strategy can be successfully transferred from equity to futures markets. First, we preprocess futures time series comprised of front months to render them suitable for our returns-based trading framework and compile a data set comprised of 60 futures covering nearly 10 trading years. Next, we train several machine learning models to predict whether the h-day-ahead return of each future out- or underperforms the corresponding cross-sectional median return. Finally, we enter long/short positions for the top/flop-k futures for a duration of h days and assess the financial performance of the resulting portfolio in an out-of-sample testing period. Thereby, we find the machine learning models to yield statistically significant out-of-sample break-even transaction costs of 6.3 bp—a clear challenge to the semi-strong form of market efficiency. Finally, we discuss sources of profitability and the robustness of our findings.
APA, Harvard, Vancouver, ISO, and other styles
40

Barandas, Marília, Duarte Folgado, Ricardo Santos, Raquel Simão, and Hugo Gamboa. "Uncertainty-Based Rejection in Machine Learning: Implications for Model Development and Interpretability." Electronics 11, no. 3 (January 28, 2022): 396. http://dx.doi.org/10.3390/electronics11030396.

Full text
Abstract:
Uncertainty is present in every single prediction of Machine Learning (ML) models. Uncertainty Quantification (UQ) is arguably relevant, in particular for safety-critical applications. Prior research focused on the development of methods to quantify uncertainty; however, less attention has been given to how to leverage the knowledge of uncertainty in the process of model development. This work focused on applying UQ into practice, closing the gap of its utility in the ML pipeline and giving insights into how UQ is used to improve model development and its interpretability. We identified three main research questions: (1) How can UQ contribute to choosing the most suitable model for a given classification task? (2) Can UQ be used to combine different models in a principled manner? (3) Can visualization techniques improve UQ’s interpretability? These questions are answered by applying several methods to quantify uncertainty in both a simulated dataset and a real-world dataset of Human Activity Recognition (HAR). Our results showed that uncertainty quantification can increase model robustness and interpretability.
APA, Harvard, Vancouver, ISO, and other styles
41

Chen, Zhanbo, and Qiufeng Wei. "Developing an Improved Survival Prediction Model for Disease Prognosis." Biomolecules 12, no. 12 (November 25, 2022): 1751. http://dx.doi.org/10.3390/biom12121751.

Full text
Abstract:
Machine learning has become an important research field in genetics and molecular biology. Survival analysis using machine learning can provide an important computed-aid clinical research scheme for evaluating tumor treatment options. However, the genomic features are high-dimensional, which limits the prediction performance of the survival learning model. Therefore, in this paper, we propose an improved survival prediction model using a deep forest and self-supervised learning. It uses a deep survival forest to perform adaptive learning of high-dimensional genomic data and ensure robustness. In addition, self-supervised learning, as a semi-supervised learning style, is designed to utilize unlabeled samples to improve model performance. Based on four cancer datasets from The Cancer Genome Atlas (TCGA), the experimental results show that our proposed method outperforms four advanced survival analysis methods in terms of the C-index and brier score. The developed prediction model will help doctors rethink patient characteristics’ relevance to survival time and personalize treatment decisions.
APA, Harvard, Vancouver, ISO, and other styles
42

Liu, Zhewei, Zijia Zhang, Yaoming Cai, Yilin Miao, and Zhikun Chen. "Semi-Supervised Classification via Hypergraph Convolutional Extreme Learning Machine." Applied Sciences 11, no. 9 (April 25, 2021): 3867. http://dx.doi.org/10.3390/app11093867.

Full text
Abstract:
Extreme Learning Machine (ELM) is characterized by simplicity, generalization ability, and computational efficiency. However, previous ELMs fail to consider the inherent high-order relationship among data points, resulting in being powerless on structured data and poor robustness on noise data. This paper presents a novel semi-supervised ELM, termed Hypergraph Convolutional ELM (HGCELM), based on using hypergraph convolution to extend ELM into the non-Euclidean domain. The method inherits all the advantages from ELM, and consists of a random hypergraph convolutional layer followed by a hypergraph convolutional regression layer, enabling it to model complex intraclass variations. We show that the traditional ELM is a special case of the HGCELM model in the regular Euclidean domain. Extensive experimental results show that HGCELM remarkably outperforms eight competitive methods on 26 classification benchmarks.
APA, Harvard, Vancouver, ISO, and other styles
43

Dunn, Corey, Nour Moustafa, and Benjamin Turnbull. "Robustness Evaluations of Sustainable Machine Learning Models against Data Poisoning Attacks in the Internet of Things." Sustainability 12, no. 16 (August 10, 2020): 6434. http://dx.doi.org/10.3390/su12166434.

Full text
Abstract:
With the increasing popularity of the Internet of Things (IoT) platforms, the cyber security of these platforms is a highly active area of research. One key technology underpinning smart IoT systems is machine learning, which classifies and predicts events from large-scale data in IoT networks. Machine learning is susceptible to cyber attacks, particularly data poisoning attacks that inject false data when training machine learning models. Data poisoning attacks degrade the performances of machine learning models. It is an ongoing research challenge to develop trustworthy machine learning models resilient and sustainable against data poisoning attacks in IoT networks. We studied the effects of data poisoning attacks on machine learning models, including the gradient boosting machine, random forest, naive Bayes, and feed-forward deep learning, to determine the levels to which the models should be trusted and said to be reliable in real-world IoT settings. In the training phase, a label modification function is developed to manipulate legitimate input classes. The function is employed at data poisoning rates of 5%, 10%, 20%, and 30% that allow the comparison of the poisoned models and display their performance degradations. The machine learning models have been evaluated using the ToN_IoT and UNSW NB-15 datasets, as they include a wide variety of recent legitimate and attack vectors. The experimental results revealed that the models’ performances will be degraded, in terms of accuracy and detection rates, if the number of the trained normal observations is not significantly larger than the poisoned data. At the rate of data poisoning of 30% or greater on input data, machine learning performances are significantly degraded.
APA, Harvard, Vancouver, ISO, and other styles
44

Tóth, Martos, and Nelson Sommerfeldt. "PV self-consumption prediction methods using supervised machine learning." E3S Web of Conferences 362 (2022): 02003. http://dx.doi.org/10.1051/e3sconf/202236202003.

Full text
Abstract:
The increased prevalence of photovoltaic (PV) self-consumption policies across Europe and the world place an increased importance on accurate predictions for life-cycle costing during the planning phase. This study presents several machine learning and regression models for predicting self-consumption, trained on a variety of datasets from Sweden. The results show that advanced ML models have an improved performance over simpler regressions, where the highest performing model, Random Forest, has a mean average error of 1.5 percentage points and an R2 of 0.977. Training models using widely available typical meteorological year (TMY) climate data is also shown to introduce small, acceptable errors when tested against spatially and temporally matched climate and load data. The ability to train the ML models with TMY climate data makes their adoption easier and builds on previous work by demonstrating the robustness of the methodology as a self-consumption prediction tool. The low error and high R2 are a notable improvement over previous estimation models and the minimal input data requirements make them easy to adopt and apply in a wide array of applications.
APA, Harvard, Vancouver, ISO, and other styles
45

Zhang, Fengyi, Xinyuan Cui, Renrong Gong, Chuan Zhang, and Zhigao Liao. "Key Experimental Factors of Machine Learning-Based Identification of Surgery Cancellations." Journal of Healthcare Engineering 2021 (February 20, 2021): 1–10. http://dx.doi.org/10.1155/2021/6247652.

Full text
Abstract:
This study aimed to provide effective methods for the identification of surgeries with high cancellation risk based on machine learning models and analyze the key factors that affect the identification performance. The data covered the period from January 1, 2013, to December 31, 2014, at West China Hospital in China, which focus on elective urologic surgeries. All surgeries were scheduled one day in advance, and all cancellations were of institutional resource- and capacity-related types. Feature selection strategies, machine learning models, and sampling methods are the most discussed topic in general machine learning researches and have a direct impact on the performance of machine learning models. Hence, they were considered to systematically generate complete schemes in machine learning-based identification of surgery cancellations. The results proved the feasibility and robustness of identifying surgeries with high cancellation risk, with the considerable maximum of area under the curve (AUC) (0.7199) for random forest model with original sampling using backward selection strategy. In addition, one-side Delong test and sum of square error analysis were conducted to measure the effects of feature selection strategy, machine learning model, and sampling method on the identification of surgeries with high cancellation risk, and the selection of machine learning model was identified as the key factors that affect the identification of surgeries with high cancellation risk. This study offers methodology and insights for identifying the key experimental factors for identifying surgery cancellations, and it is helpful to further research on machine learning-based identification of surgeries with high cancellation risk.
APA, Harvard, Vancouver, ISO, and other styles
46

Wan, Zhibin, Changqing Zhang, Pengfei Zhu, and Qinghua Hu. "Multi-View Information-Bottleneck Representation Learning." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 11 (May 18, 2021): 10085–92. http://dx.doi.org/10.1609/aaai.v35i11.17210.

Full text
Abstract:
In real-world applications, clustering or classification can usually be improved by fusing information from different views. Therefore, unsupervised representation learning on multi-view data becomes a compelling topic in machine learning. In this paper, we propose a novel and flexible unsupervised multi-view representation learning model termed Collaborative Multi-View Information Bottleneck Networks (CMIB-Nets), which comprehensively explores the common latent structure and the view-specific intrinsic information, and discards the superfluous information in the data significantly improving the generalization capability of the model. Specifically, our proposed model relies on the information bottleneck principle to integrate the shared representation among different views and the view-specific representation of each view, prompting the multi-view complete representation and flexibly balancing the complementarity and consistency among multiple views. We conduct extensive experiments (including clustering analysis, robustness experiment, and ablation study) on real-world datasets, which empirically show promising generalization ability and robustness compared to state-of-the-arts.
APA, Harvard, Vancouver, ISO, and other styles
47

Omri, Mohamed Nazih, and Wafa Mribah. "Towards an Intelligent Machine Learning-based Business Approach." International Journal of Intelligent Systems and Applications 14, no. 1 (February 8, 2022): 1–23. http://dx.doi.org/10.5815/ijisa.2022.01.01.

Full text
Abstract:
With the constant increase of data induced by stakeholders throughout a product life cycle, companies tend to rely on project management tools for guidance. Business intelligence approaches that are project-oriented will help the team communicate better, plan their next steps, have an overview of the current project state and take concrete actions prior to the provided forecasts. The spread of agile working mindsets are making these tools even more useful. It sets a basic understanding of how the project should be running so that the implementation is easy to follow on and easy to use. In this paper, we offer a model that makes project management accessible from different software development tools and different data sources. Our model provide project data analysis to improve aspects: (i) collaboration which includes team communication, team dashboard. It also optimizes document sharing, deadlines and status updates. (ii) planning: allows the tasks described by the software to be used and made visible. It will also involve tracking task time to display any barriers to work that some members might be facing without reporting them. (iii) forecasting to predict future results from behavioral data, which will allow concrete measures to be taken. And (iv) Documentation to involve reports that summarize all relevant project information, such as time spent on tasks and charts that study the status of the project. The experimental study carried out on the various data collections on our model and on the main models that we have studied in the literature, as well as the analysis of the results, which we obtained, clearly show the limits of these studied models and confirms the performance of our model as well as efficiency in terms of precision, recall and robustness.
APA, Harvard, Vancouver, ISO, and other styles
48

Stachl, Clemens, Florian Pargent, Sven Hilbert, Gabriella M. Harari, Ramona Schoedel, Sumer Vaid, Samuel D. Gosling, and Markus Bühner. "Personality Research and Assessment in the Era of Machine Learning." European Journal of Personality 34, no. 5 (September 2020): 613–31. http://dx.doi.org/10.1002/per.2257.

Full text
Abstract:
The increasing availability of high–dimensional, fine–grained data about human behaviour, gathered from mobile sensing studies and in the form of digital footprints, is poised to drastically alter the way personality psychologists perform research and undertake personality assessment. These new kinds and quantities of data raise important questions about how to analyse the data and interpret the results appropriately. Machine learning models are well suited to these kinds of data, allowing researchers to model highly complex relationships and to evaluate the generalizability and robustness of their results using resampling methods. The correct usage of machine learning models requires specialized methodological training that considers issues specific to this type of modelling. Here, we first provide a brief overview of past studies using machine learning in personality psychology. Second, we illustrate the main challenges that researchers face when building, interpreting, and validating machine learning models. Third, we discuss the evaluation of personality scales, derived using machine learning methods. Fourth, we highlight some key issues that arise from the use of latent variables in the modelling process. We conclude with an outlook on the future role of machine learning models in personality research and assessment.
APA, Harvard, Vancouver, ISO, and other styles
49

Kim, Jaehun. "Increasing trust in complex machine learning systems." ACM SIGIR Forum 55, no. 1 (June 2021): 1–3. http://dx.doi.org/10.1145/3476415.3476435.

Full text
Abstract:
Machine learning (ML) has become a core technology for many real-world applications. Modern ML models are applied to unprecedentedly complex and difficult challenges, including very large and subjective problems. For instance, applications towards multimedia understanding have been advanced substantially. Here, it is already prevalent that cultural/artistic objects such as music and videos are analyzed and served to users according to their preference, enabled through ML techniques. One of the most recent breakthroughs in ML is Deep Learning (DL), which has been immensely adopted to tackle such complex problems. DL allows for higher learning capacity, making end-to-end learning possible, which reduces the need for substantial engineering effort, while achieving high effectiveness. At the same time, this also makes DL models more complex than conventional ML models. Reports in several domains indicate that such more complex ML models may have potentially critical hidden problems: various biases embedded in the training data can emerge in the prediction, extremely sensitive models can make unaccountable mistakes. Furthermore, the black-box nature of the DL models hinders the interpretation of the mechanisms behind them. Such unexpected drawbacks result in a significant impact on the trustworthiness of the systems in which the ML models are equipped as the core apparatus. In this thesis, a series of studies investigates aspects of trustworthiness for complex ML applications, namely the reliability and explainability. Specifically, we focus on music as the primary domain of interest, considering its complexity and subjectivity. Due to this nature of music, ML models for music are necessarily complex for achieving meaningful effectiveness. As such, the reliability and explainability of music ML models are crucial in the field. The first main chapter of the thesis investigates the transferability of the neural network in the Music Information Retrieval (MIR) context. Transfer learning, where the pre-trained ML models are used as off-the-shelf modules for the task at hand, has become one of the major ML practices. It is helpful since a substantial amount of the information is already encoded in the pre-trained models, which allows the model to achieve high effectiveness even when the amount of the dataset for the current task is scarce. However, this may not always be true if the "source" task which pre-trained the model shares little commonality with the "target" task at hand. An experiment including multiple "source" tasks and "target" tasks was conducted to examine the conditions which have a positive effect on the transferability. The result of the experiment suggests that the number of source tasks is a major factor of transferability. Simultaneously, it is less evident that there is a single source task that is universally effective on multiple target tasks. Overall, we conclude that considering multiple pre-trained models or pre-training a model employing heterogeneous source tasks can increase the chance for successful transfer learning. The second major work investigates the robustness of the DL models in the transfer learning context. The hypothesis is that the DL models can be susceptible to imperceptible noise on the input. This may drastically shift the analysis of similarity among inputs, which is undesirable for tasks such as information retrieval. Several DL models pre-trained in MIR tasks are examined for a set of plausible perturbations in a real-world setup. Based on a proposed sensitivity measure, the experimental results indicate that all the DL models were substantially vulnerable to perturbations, compared to a traditional feature encoder. They also suggest that the experimental framework can be used to test the pre-trained DL models for measuring robustness. In the final main chapter, the explainability of black-box ML models is discussed. In particular, the chapter focuses on the evaluation of the explanation derived from model-agnostic explanation methods. With black-box ML models having become common practice, model-agnostic explanation methods have been developed to explain a prediction. However, the evaluation of such explanations is still an open problem. The work introduces an evaluation framework that measures the quality of the explanations employing fidelity and complexity. Fidelity refers to the explained mechanism's coherence to the black-box model, while complexity is the length of the explanation. Throughout the thesis, we gave special attention to the experimental design, such that robust conclusions can be reached. Furthermore, we focused on delivering machine learning framework and evaluation frameworks. This is crucial, as we intend that the experimental design and results will be reusable in general ML practice. As it implies, we also aim our findings to be applicable beyond the music applications such as computer vision or natural language processing. Trustworthiness in ML is not a domain-specific problem. Thus, it is vital for both researchers and practitioners from diverse problem spaces to increase awareness of complex ML systems' trustworthiness. We believe the research reported in this thesis provides meaningful stepping stones towards the trustworthiness of ML.
APA, Harvard, Vancouver, ISO, and other styles
50

Schober, Sebastian A., Yosra Bahri, Cecilia Carbonelli, and Robert Wille. "Neural Network Robustness Analysis Using Sensor Simulations for a Graphene-Based Semiconductor Gas Sensor." Chemosensors 10, no. 5 (April 21, 2022): 152. http://dx.doi.org/10.3390/chemosensors10050152.

Full text
Abstract:
Despite their advantages regarding production costs and flexibility, chemiresistive gas sensors often show drawbacks in reproducibility, signal drift and ageing. As pattern recognition algorithms, such as neural networks, are operating on top of raw sensor signals, assessing the impact of these technological drawbacks on the prediction performance is essential for ensuring a suitable measuring accuracy. In this work, we propose a characterization scheme to analyze the robustness of different machine learning models for a chemiresistive gas sensor based on a sensor simulation model. Our investigations are structured into four separate studies: in three studies, the impact of different sensor instabilities on the concentration prediction performance of the algorithms is investigated, including sensor-to-sensor variations, sensor drift and sensor ageing. In a further study, the explainability of the machine learning models is analyzed by applying a state-of-the-art feature ranking method called SHAP. Our results show the feasibility of model-based algorithm testing and substantiate the need for the thorough characterization of chemiresistive sensor algorithms before sensor deployment in order to ensure robust measurement performance.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography