Academic literature on the topic 'RDF dataset metrics'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'RDF dataset metrics.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "RDF dataset metrics"

1

Mountantonakis, Michalis, and Yannis Tzitzikas. "Content-based Union and Complement Metrics for Dataset Search over RDF Knowledge Graphs." Journal of Data and Information Quality 12, no. 2 (May 14, 2020): 1–31. http://dx.doi.org/10.1145/3372750.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Xia, Jianglin. "Credit Card Fraud Detection Based on Support Vector Machine." Highlights in Science, Engineering and Technology 23 (December 3, 2022): 93–97. http://dx.doi.org/10.54097/hset.v23i.3202.

Full text
Abstract:
Due to the increasing popularity cashless transactions, credit card fraud has become one of the most common frauds and caused huge harm to the financial institutions and individuals in real life. In this academic paper, the algorithm Support Vector Machine (SVM) is used to build models to deal with the credit card fraud detection problem with the performance metrics AUC and F1-score. The experiment dataset is named Credit Card Transactions Fraud Detection Dataset from the Kaggle website. After the step of preprocessing, the dataset is split into the training, testing and validation dataset with 11 numerical features and a label feature called “is_fraud”. The inner parameter “class_weight” of the SVM algorithm in Python is set as “balanced” to deal with the imbalanced datasets. The main method to find the optimized models is using the GridSearchCV function in Python library sklearn. After tuning the hyperparameters and handling the overfitting phenomenon, the optimized models for the two metrics are found. The parameter values of the best model for AUC are C=10, class_weight= “balanced”, g =0.01, kernel = “rbf”. The training AUC is 0.87 and testing AUC is 0.90. The parameter values of the final optimized model for F1-score are C=0.8, class_weight= “balanced”, g =0.06, kernel = “rbf”. The final training F-score is 0.305 and testing F-score is 0.260.
APA, Harvard, Vancouver, ISO, and other styles
3

Wang, Ke, Ligang Cheng, and Bin Yong. "Spectral-Similarity-Based Kernel of SVM for Hyperspectral Image Classification." Remote Sensing 12, no. 13 (July 6, 2020): 2154. http://dx.doi.org/10.3390/rs12132154.

Full text
Abstract:
Spectral similarity measures can be regarded as potential metrics for kernel functions, and can be used to generate spectral-similarity-based kernels. However, spectral-similarity-based kernels have not received significant attention from researchers. In this paper, we propose two novel spectral-similarity-based kernels based on spectral angle mapper (SAM) and spectral information divergence (SID) combined with the radial basis function (RBF) kernel: Power spectral angle mapper RBF (Power-SAM-RBF) and normalized spectral information divergence-based RBF (Normalized-SID-RBF) kernels. First, we prove these spectral-similarity-based kernels to be Mercer’s kernels. Second, we analyze their efficiency in terms of local and global kernels. Finally, we consider three hyperspectral datasets to analyze the effectiveness of the proposed spectral-similarity-based kernels. Experimental results demonstrate that the Power-SAM-RBF and SAM-RBF kernels can obtain an impressive performance, particularly the Power-SAM-RBF kernel. For example, when the ratio of the training set is 20 % , the kappa coefficient of Power-SAM-RBF kernel (0.8561) is 1.61 % , 1.32 % , and 1.23 % higher than that of the RBF kernel on the Indian Pines, University of Pavia, and Salinas Valley datasets, respectively. We present three conclusions. First, the superiority of the Power-SAM-RBF kernel compared to other kernels is evident. Second, the Power-SAM-RBF kernel can provide an outstanding performance when the similarity between spectral signatures in the same hyperspectral dataset is either extremely high or extremely low. Third, the Power-SAM-RBF kernel provides even greater benefits compared to other commonly used kernels when the sizes of the training sets increase. In future work, multiple kernels combining with the spectral-similarity-based kernel are expected to be provide better hyperspectral classification.
APA, Harvard, Vancouver, ISO, and other styles
4

Zhao, Qinghe, Zifang Zhang, Yuchen Huang, and Junlong Fang. "TPE-RBF-SVM Model for Soybean Categories Recognition in Selected Hyperspectral Bands Based on Extreme Gradient Boosting Feature Importance Values." Agriculture 12, no. 9 (September 13, 2022): 1452. http://dx.doi.org/10.3390/agriculture12091452.

Full text
Abstract:
Soybeans with insignificant differences in appearance have large differences in their internal physical and chemical components; therefore, follow-up storage, transportation and processing require targeted differential treatment. A fast and effective machine learning method based on hyperspectral data of soybeans for pattern recognition of categories is designed as a non-destructive testing method in this paper. A hyperspectral-image dataset with 2299 soybean seeds in four categories is collected. Ten features are selected using an extreme gradient boosting algorithm from 203 hyperspectral bands in a range of 400 to 1000 nm; a Gaussian radial basis kernel function support vector machine with optimization by the tree-structured Parzen estimator algorithm is built as the TPE-RBF-SVM model for pattern recognition of soybean categories. The metrics of TPE-RBF-SVM are significantly improved compared with other machine learning algorithms. The accuracy is 0.9165 in the independent test dataset, which is 9.786% higher for the vanilla RBF-SVM model and 10.02% higher than the extreme gradient boosting model.
APA, Harvard, Vancouver, ISO, and other styles
5

Chen, Yanji, Mieczyslaw M. Kokar, Jakub Moskal, and Kaushik R. Chowdhury. "Metrics-Based Comparison of OWL and XML for Representing and Querying Cognitive Radio Capabilities." Applied Sciences 12, no. 23 (November 23, 2022): 11946. http://dx.doi.org/10.3390/app122311946.

Full text
Abstract:
Collaborative spectrum access requires wireless devices to perform spectrum-related tasks (such as sensing) on request from other nodes. Thus, while joining the network, they need to inform neighboring devices and/or the central coordinator of their capabilities. During the operational phase, nodes may request other permissions from the the controller, like the opportunity to transmit according to the current policies and spectrum availability. To achieve such coordinated behavior, all associated devices within the network need a language for describing radio capabilities, requests, scenarios, policies, and spectrum availability. In this paper, we present a thorough comparison of the use of two candidate languages—Web Ontology Language (OWL) and eXtensible Markup Language (XML)—for such purposes. Towards this goal, we propose an evaluation method for automating quantitative comparisons with metrics such as precision, recall, device registration, and the query response time. The requests are expressed in both SPARQL Protocol and RDF Query Language (SPARQL) and XML Query Language (XQuery), whereas the device capabilities are expressed in both OWL and XML. The evaluation results demonstrate the advantages of using OWL semantics to improve the quality of matching results over XML. We also discuss how the evaluation method can be applicable to other scenarios where knowledge, datasets, and queries require richer expressiveness and semantics.
APA, Harvard, Vancouver, ISO, and other styles
6

Jerop, Brenda, and Davies Rene Segera. "An Efficient PCA-GA-HKSVM-Based Disease Diagnostic Assistant." BioMed Research International 2021 (October 20, 2021): 1–10. http://dx.doi.org/10.1155/2021/4784057.

Full text
Abstract:
Disease diagnosis faces challenges such as misdiagnosis, lack of diagnosis, and slow diagnosis. There are several machine learning techniques that have been applied to address these challenges, where a set of symptoms is applied to a classification model that predicts the presence or absence of a disease. To improve on the performance of these techniques, this paper presents a technique which involves feature selection using principal component analysis (PCA), a hybrid kernel-based support vector machine (HKSVM) classification model and hyperparameter optimization using genetic algorithm (GA). The HKSVM in this paper introduces a new way of combining three kernels: Radial basis function (RBF), linear, and polynomial. Combining local (RBF) and global (linear and polynomial) kernels has the effect of improved model performance. This is because the local kernels are better able to distinguish points closer to each other while the global kernels are more suited to distinguish points that are far away from each other. The PCA-GA-HKSVM is used on 7 different medical datasets, with two datasets being multiclass datasets and 5 datasets being binary. Performance evaluation metrics used were accuracy, precision, and recall. It was observed that the PCA-GA-HKSVM offered better performance than the single kernel support vector machines (SVMs).
APA, Harvard, Vancouver, ISO, and other styles
7

Mohammed, Yosra Abdulaziz, and Eman Gadban Saleh. "Comparative study of logistic regression and artificial neural networks on predicting breast cancer cytology." Indonesian Journal of Electrical Engineering and Computer Science 21, no. 2 (February 1, 2021): 1113. http://dx.doi.org/10.11591/ijeecs.v21.i2.pp1113-1120.

Full text
Abstract:
<p>Currently, breast cancer is one of the most common cancers and a main reason of women death worldwide particularly in<strong> </strong>developing countries such as Iraq. our work aims to predict the type of tumor whether benign or malignant through models that were built using logistic regression and neural networks and we hope it will help doctors in detecting the type of breast tumor. Four models were set using binary logistic regression and two different types of artificial neural networks namely multilayer perceptron MLP and radial basis function RBF. Evaluation of validated and trained models was done using several performance metrics like accuracy, sensitivity, specificity, and AUC (area under receiver operating characteristic ROC). Dataset was downloaded from UCI ml repository; it is composed of 9 attributes and 699 samples. The findings are clearly showing that the RBF NN classifier is the best in prediction of the type of breast tumors since it had recorded the highest performance in terms of correct classification rate (accuracy), sensitivity, specificity, and AUC (area under Receiver Operating Characteristic ROC) among all other models.</p>
APA, Harvard, Vancouver, ISO, and other styles
8

Panda, Mrutyunjaya. "Software Defect Prediction Using Hybrid Distribution Base Balance Instance Selection and Radial Basis Function Classifier." International Journal of System Dynamics Applications 8, no. 3 (July 2019): 53–75. http://dx.doi.org/10.4018/ijsda.2019070103.

Full text
Abstract:
Software is an important part of human life and with the rapid development of software engineering the demands for software to be reliable with low defects is increasingly pressing. The building of a software defect prediction model is proposed in this article by using various software metrics with publicly available historical software defect datasets collected from several projects. Such a prediction model can enable the software engineers to take proactive actions in enhancing software quality from the early stages of the software development cycle. This article introduces a hybrid classification method (DBBRBF) by combining distribution base balance (DBB) based instance selection and radial basis function (RBF) neural network classifier to obtain the best prediction compared to the existing research. The experimental results with post-hoc statistical significance tests shows the effectiveness of the proposed approach.
APA, Harvard, Vancouver, ISO, and other styles
9

Villa, Amalia, Abhijith Mundanad Narayanan, Sabine Van Huffel, Alexander Bertrand, and Carolina Varon. "Utility metric for unsupervised feature selection." PeerJ Computer Science 7 (April 21, 2021): e477. http://dx.doi.org/10.7717/peerj-cs.477.

Full text
Abstract:
Feature selection techniques are very useful approaches for dimensionality reduction in data analysis. They provide interpretable results by reducing the dimensions of the data to a subset of the original set of features. When the data lack annotations, unsupervised feature selectors are required for their analysis. Several algorithms for this aim exist in the literature, but despite their large applicability, they can be very inaccessible or cumbersome to use, mainly due to the need for tuning non-intuitive parameters and the high computational demands. In this work, a publicly available ready-to-use unsupervised feature selector is proposed, with comparable results to the state-of-the-art at a much lower computational cost. The suggested approach belongs to the methods known as spectral feature selectors. These methods generally consist of two stages: manifold learning and subset selection. In the first stage, the underlying structures in the high-dimensional data are extracted, while in the second stage a subset of the features is selected to replicate these structures. This paper suggests two contributions to this field, related to each of the stages involved. In the manifold learning stage, the effect of non-linearities in the data is explored, making use of a radial basis function (RBF) kernel, for which an alternative solution for the estimation of the kernel parameter is presented for cases with high-dimensional data. Additionally, the use of a backwards greedy approach based on the least-squares utility metric for the subset selection stage is proposed. The combination of these new ingredients results in the utility metric for unsupervised feature selection U2FS algorithm. The proposed U2FS algorithm succeeds in selecting the correct features in a simulation environment. In addition, the performance of the method on benchmark datasets is comparable to the state-of-the-art, while requiring less computational time. Moreover, unlike the state-of-the-art, U2FS does not require any tuning of parameters.
APA, Harvard, Vancouver, ISO, and other styles
10

Bashir, Kamal, Tianrui Li, and Mahama Yahaya. "A Novel Feature Selection Method Based on Maximum Likelihood Logistic Regression for Imbalanced Learning in Software Defect Prediction." International Arab Journal of Information Technology 17, no. 5 (September 1, 2020): 721–30. http://dx.doi.org/10.34028/iajit/17/5/5.

Full text
Abstract:
The most frequently used machine learning feature ranking approaches failed to present optimal feature subset for accurate prediction of defective software modules in out-of-sample data. Machine learning Feature Selection (FS) algorithms such as Chi-Square (CS), Information Gain (IG), Gain Ratio (GR), RelieF (RF) and Symmetric Uncertainty (SU) perform relatively poor at prediction, even after balancing class distribution in the training data. In this study, we propose a novel FS method based on the Maximum Likelihood Logistic Regression (MLLR). We apply this method on six software defect datasets in their sampled and unsampled forms to select useful features for classification in the context of Software Defect Prediction (SDP). The Support Vector Machine (SVM) and Random Forest (RaF) classifiers are applied on the FS subsets that are based on sampled and unsampled datasets. The performance of the models captured using Area Ander Receiver Operating Characteristics Curve (AUC) metrics are compared for all FS methods considered. The Analysis Of Variance (ANOVA) F-test results validate the superiority of the proposed method over all the FS techniques, both in sampled and unsampled data. The results confirm that the MLLR can be useful in selecting optimal feature subset for more accurate prediction of defective modules in software development process
APA, Harvard, Vancouver, ISO, and other styles

Dissertations / Theses on the topic "RDF dataset metrics"

1

Soderi, Mirco. "Semantic models for the modeling and management of big data in a smart city environment." Doctoral thesis, 2021. http://hdl.handle.net/2158/1232245.

Full text
Abstract:
The overall purpose of this research has been the building or the improve- ment of semantic models for the representation of data related to smart cities and smart industries, in such a way that it could also be possible to build context-rich, user-oriented, ecient and eective applications based on such data. In some more detail, one of the key purposes has been the modelling of structural and the functioning aspects of the urban mobility and the produc- tion of instances exploiting the Open Street Map, that once integrated with trac sensors data, it has lead to the building and displaying of real-time trac reconstructions at a city level. One second key purpose has been the modelling of the Internet of Things, that allows today to seamlessy and e- ciently identify sensing devices that are deployed in a given area or along a given path and that are of a given type, and also inspect real-time data that they produce, through a user-oriented Web application, namely the Service Map. A pragmatic approach to the modelling has been followed, always tak- ing into consideration the best practices of semantic modelling on one side for that a clean, comprehensive and understandable model could result, and the reality of the data at our hands and of the applicative requirements on the other side. As said, the identication of architectures and methods that could grant eciency and scalability in data access has also been a primary purpose of this research that has led to the denition and implementation of a federation of Service Maps, namely the Super Service Map. The archi- tecture is fully distributed: each Super Service Map has a local list of the actual Service Maps with relevant metadata, it exposes the same interface as actual Service Maps, it forwards requests and builds merged responses, also implementing security and caching mechanisms. As said, the identica- tion of technologies, tools, methods, for presenting the data in a user-friendly manner is also has been a relevant part of this research, and it has led among the other to the denition and implementation of a client-server architecture and a Web interface in the Snap4City platform for the building, manage- ment, and displaying of synoptic templates and instances thanks to which users can securely display and iteract with dierent types of data. In end, some eort has been made for the automatic classication of RDF datasets as for their structures and purposes, based on the computation of metrics through SPARQL queries and on the application of dimensionality reduc- tion and clustering techniques. A Web portal is available where directories, datasets, metrics, and computations can be inspected even at real-time.
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "RDF dataset metrics"

1

Rizvi, Syed Zeeshan, Muhammad Umar Farooq, and Rana Hammad Raza. "Performance Comparison of Deep Residual Networks-Based Super Resolution Algorithms Using Thermal Images: Case Study of Crowd Counting." In Digital Interaction and Machine Intelligence, 75–87. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-031-11432-8_7.

Full text
Abstract:
AbstractHumans are able to perceive objects only in the visible spectrum range which limits the perception abilities in poor weather or low illumination conditions. The limitations are usually handled through technological advancements in thermographic imaging. However, thermal cameras have poor spatial resolutions compared to RGB cameras. Super-resolution (SR) techniques are commonly used to improve the overall quality of low-resolution images. There has been a major shift of research among the Computer Vision researchers towards SR techniques particularly aimed for thermal images. This paper analyzes the performance of three deep learning-based state-of-the-art SR algorithms namely Enhanced Deep Super Resolution (EDSR), Residual Channel Attention Network (RCAN) and Residual Dense Network (RDN) on thermal images. The algorithms were trained from scratch for different upscaling factors of ×2 and ×4. The dataset was generated from two different thermal imaging sequences of BU-TIV benchmark. The sequences contain both sparse and highly dense type of crowds with a far field camera view. The trained models were then used to super-resolve unseen test images. The quantitative analysis of the test images was performed using common image quality metrics such as PSNR, SSIM and LPIPS, while qualitative analysis was provided by evaluating effectiveness of the algorithms for crowd counting application. After only 54 and 51 epochs of RCAN and RDN respectively, both approaches were able to output average scores of 37.878, 0.986, 0.0098 and 30.175, 0.945, 0.0636 for PSNR, SSIM and LPIPS respectively. The EDSR algorithm took the least computation time during both training and testing because of its simple architecture. This research proves that a reasonable accuracy can be achieved with fewer training epochs when an application-specific dataset is carefully selected.
APA, Harvard, Vancouver, ISO, and other styles
2

Harkat, Houda, Jose Nascimento, Alexandre Bernardino, and Hasmath Farhana Thariq Ahmed. "Fire images classification using high order statistical features." In Advances in Forest Fire Research 2022, 192–97. Imprensa da Universidade de Coimbra, 2022. http://dx.doi.org/10.14195/978-989-26-2298-9_31.

Full text
Abstract:
Wildfires and forest fires have devastated millions of hectares of forest across the world over the years. Computer vision-based fire classification, which classifies fire pixels from non-fire pixels in image or video datasets, has gained popularity as a result of recent innovations. A conventional machine learning-based approach or a deep learning-based approach can be used to distinguish fire pixels from an image or video. Deep learning is currently the most prominent method for detecting forest fires. Although deep learning algorithms can handle large volumes of data, typically ignore the differences in complexity among training samples, limiting the performance of training models. Moreover, in real-world fire scenarios, deep learning techniques with little data and features underperform. The present study utilizes a machine learning-based approach for extracting features of higher-order statistical methods from pre-processed images from publicly available datasets: Corsican and FLAME, and a private dataset: Firefront Gestosa. It should be noted that handling multidimensional data to train a classifier in machine learning applications is complex. This issue is addressed through feature selection, which eliminates duplicate or irrelevant data that has an effect on the model's performance. A greedy feature selection criterion is adopted in this study to select the most significant features for classification while reducing computational costs. The Support Vector Machine (SVM) is a conventional machine classifier that works on discriminative features input obtained using the MIFS, feature selection technique. The SVM uses a Radial Basis Function (RBF) kernel to classify fire and non-fire pixels, and the model's performance is assessed using assessment metrics like overall accuracy, sensitivity, specificity, precision, recall, F-measure, and G-mean.
APA, Harvard, Vancouver, ISO, and other styles
3

Ahmad, Mahmood, Xiaowei Tang, and Feezan Ahmad. "Evaluation of Liquefaction-Induced Settlement Using Random Forest and REP Tree Models: Taking Pohang Earthquake as a Case of Illustration." In Natural Hazards - Impacts, Adjustments and Resilience. IntechOpen, 2021. http://dx.doi.org/10.5772/intechopen.94274.

Full text
Abstract:
A liquefaction-induced settlement assessment is considered one of the major challenges in geotechnical earthquake engineering. This paper presents random forest (RF) and reduced error pruning tree (REP Tree) models for predicting settlement caused by liquefaction. Standard penetration test (SPT) data were obtained for five separate borehole sites near the Pohang Earthquake epicenter. The data used in this study comprise of four features, namely depth, unit weight, corrected SPT blow count and cyclic stress ratio. The available data is divided into two parts: training set (80%) and test set (20%). The output of the RF and REP Tree models is evaluated using statistical parameters including coefficient of correlation (r), mean absolute error (MAE), and root mean squared error (RMSE). The applications for the aforementioned approach for predicting the liquefaction-induced settlement are compared and discussed. The analysis of statistical metrics for the evaluating liquefaction-induced settlement dataset demonstrates that the RF achieved comparatively better and reliable results.
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "RDF dataset metrics"

1

Gao, Hanning, Lingfei Wu, Po Hu, and Fangli Xu. "RDF-to-Text Generation with Graph-augmented Structural Neural Encoders." In Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20}. California: International Joint Conferences on Artificial Intelligence Organization, 2020. http://dx.doi.org/10.24963/ijcai.2020/419.

Full text
Abstract:
The task of RDF-to-text generation is to generate a corresponding descriptive text given a set of RDF triples. Most of the previous approaches either cast this task as a sequence-to-sequence problem or employ graph-based encoder for modeling RDF triples and decode a text sequence. However, none of these methods can explicitly model both local and global structure information between and within the triples. To address these issues, we propose to jointly learn local and global structure information via combining two new graph-augmented structural neural encoders (i.e., a bidirectional graph encoder and a bidirectional graph-based meta-paths encoder) for the input triples. Experimental results on two different WebNLG datasets show that our proposed model outperforms the state-of-the-art baselines. Furthermore, we perform a human evaluation that demonstrates the effectiveness of the proposed method by evaluating generated text quality using various subjective metrics.
APA, Harvard, Vancouver, ISO, and other styles
2

Shi, Yu, and Rolf D. Reitz. "Assessment of Multi-Objective Genetic Algorithms With Different Niching Strategies and Regression Methods for Engine Optimization and Design." In ASME 2009 Internal Combustion Engine Division Spring Technical Conference. ASMEDC, 2009. http://dx.doi.org/10.1115/ices2009-76015.

Full text
Abstract:
In previous study [1] the Non-dominated Sorting Genetic Algorithm II (NSGA II) [2] performed better than other popular Multi-Objective Genetic Algorithms (MOGA) in engine optimization that sought optimal combinations of the piston bowl geometry, spray targeting, and swirl ratio. NSGA II is further studied in this paper using different niching strategies that are applied to the objective-space and design-space, which diversify the optimal objectives and design parameters accordingly. Convergence and diversity metrics are defined to assess the performance of NSGA II using different niching strategies. It was found that use of the design niching achieved more diversified results with respect to design parameters, as expected. Regression was then conducted on the design datasets that were obtained from the optimizations with two niching strategies. Four regression methods, including K-nearest neighbors (KN), Kriging (KR), Neural Networks (NN), and Radial Basis Functions (RBF), were compared. The results showed that the dataset obtained from optimization with objective niching provided a more fitted learning space for the regression methods. The KN, KR, outperformed the other two methods with respect to the prediction accuracy. Furthermore, a log transformation to the objective-space improved the prediction accuracy for the KN, KR, and NN methods but not the RBF method. The results indicate that it is appropriate to use a regression tool to partly replace the actual CFD evaluation tool in engine optimization designs using the genetic algorithm. This hybrid mode saves computational resources (processors) without losing optimal accuracy. A Design of Experiment (DoE) method (the Optimal Latin Hypercube method) was also used to generate a dataset for the regression processes. However, the predicted results were much less reliable than results that were learned using the dynamically increasing datasets from the NSGA II generations. Applying the dynamical learning strategy during the optimization processes allows computationally expensive CFD evaluations to be partly replaced by evaluations using the regression techniques. The present study demonstrates the feasibility of applying the hybrid mode to engine optimization problems, and the conclusions can also extend to other optimization studies (numerical or experimental) that feature time-consuming evaluations and have highly non-linear objective-spaces.
APA, Harvard, Vancouver, ISO, and other styles
3

Kornev, Denis, Roozbeh Sadeghian, Stanley Nwoji, Qinghua He, Amir Gandjbbakhche, and Siamak Aram. "Machine Learning-Based Gaming Behavior Prediction Platform." In 13th International Conference on Applied Human Factors and Ergonomics (AHFE 2022). AHFE International, 2022. http://dx.doi.org/10.54941/ahfe1001826.

Full text
Abstract:
Brain disorders caused by Gaming Addiction drastically increased due to the rise of Internet users and Internet Gaming auditory. Driven by such a tendency, in 2018, World Health Organization (WHO) and the American Medical Association (AMA) addressed this problem as a “gaming disorder” and added it to official manuals. Scientific society equipped by statistical analysis methods such as t-test, ANOVA, and neuroimaging techniques, such as functional magnetic resonance imaging (fMRI), positron emission tomography (PET), and electroencephalography (EEG), has achieved significant success in brain mapping, examining dynamics and patterns in different conditions and stages. Nevertheless, more powerful, self-learning intelligent algorithms are suitable not only to evaluate the correlation between gaming addiction patterns but also to predict behavior and prognosis brain response depending on the addiction severity. The current paper aims to enrich the knowledge base of the correlation between gaming activity, decision-making, and brain activation, using Machine Learning (ML) algorithms and advanced neuroimaging techniques. The proposed gaming behavior patterns prediction platform was built inside the experiment environment composed of a Functional Near-Infrared Spectrometer (fNIRS) and the computer version of Iowa Gambling Task (IGT). Thirty healthy participants were hired to perform 100 cards selection while equipped with fNIRS. Thus, accelerated by IGT gaming decision-making process was synchronized with changes of oxy-hemoglobin (HbO) levels in the human brain, averaged, and investigated in the left and the right brain hemispheres as well as different psychosomatic conditions, conditionally divided by blocks with 20 card trials in each: absolute unknown and uncertainty in the first block, “pre-hunch” and “hunch” in the second and third blocks, and conceptuality and risky in the fourth and fifth blocks. The features space was constructed around the HbO signal, split by training and tested in two proportions 70/30 and 80/20, and drove patterns prediction ML-based platform consisted of five mechanics, such as Multiple Regression, Classification and Regression Trees (CART), Support Vector Machine (SVM), Artificial Neural Network (ANN), and Random Forest. The algorithm prediction power was validated by the 5- and 10-fold cross-validation method and compared by Root Mean Squared Error (RMSE) and coefficient of determination (R Squared) metrics. Indicators of “the best” fit model, lowest RMSE, and highest R Squared were determined for each block and both brain hemispheres and used to make a conclusion about prediction accuracy: SVM algorithm with RBF kernel, Random Forest, and ANN demonstrated the best accuracy in most cases. Lastly, “best fit” classifiers were applied to the testing dataset and finalized the experiment. Hence, the distribution of gaming score was predicted by five blocks and both brain hemispheres that reflect the decision-making process patterns during gaming. The investigation showed increasing ML algorithm prediction power from IGT block one to five, reflecting an increasing learning effect as a behavioral pattern. Furthermore, performed inside constructed platform simulation could benefit in diagnosing gaming disorders, their patterns, mechanisms, and abnormalities.
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "RDF dataset metrics"

1

Idakwo, Gabriel, Sundar Thangapandian, Joseph Luttrell, Zhaoxian Zhou, Chaoyang Zhang, and Ping Gong. Deep learning-based structure-activity relationship modeling for multi-category toxicity classification : a case study of 10K Tox21 chemicals with high-throughput cell-based androgen receptor bioassay data. Engineer Research and Development Center (U.S.), July 2021. http://dx.doi.org/10.21079/11681/41302.

Full text
Abstract:
Deep learning (DL) has attracted the attention of computational toxicologists as it offers a potentially greater power for in silico predictive toxicology than existing shallow learning algorithms. However, contradicting reports have been documented. To further explore the advantages of DL over shallow learning, we conducted this case study using two cell-based androgen receptor (AR) activity datasets with 10K chemicals generated from the Tox21 program. A nested double-loop cross-validation approach was adopted along with a stratified sampling strategy for partitioning chemicals of multiple AR activity classes (i.e., agonist, antagonist, inactive, and inconclusive) at the same distribution rates amongst the training, validation and test subsets. Deep neural networks (DNN) and random forest (RF), representing deep and shallow learning algorithms, respectively, were chosen to carry out structure-activity relationship-based chemical toxicity prediction. Results suggest that DNN significantly outperformed RF (p < 0.001, ANOVA) by 22–27% for four metrics (precision, recall, F-measure, and AUPRC) and by 11% for another (AUROC). Further in-depth analyses of chemical scaffolding shed insights on structural alerts for AR agonists/antagonists and inactive/inconclusive compounds, which may aid in future drug discovery and improvement of toxicity prediction modeling.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography