Pour voir les autres types de publications sur ce sujet consultez le lien suivant : Deep Learning, Database.

Articles de revues sur le sujet « Deep Learning, Database »

Créez une référence correcte selon les styles APA, MLA, Chicago, Harvard et plusieurs autres

Choisissez une source :

Consultez les 50 meilleurs articles de revues pour votre recherche sur le sujet « Deep Learning, Database ».

À côté de chaque source dans la liste de références il y a un bouton « Ajouter à la bibliographie ». Cliquez sur ce bouton, et nous générerons automatiquement la référence bibliographique pour la source choisie selon votre style de citation préféré : APA, MLA, Harvard, Vancouver, Chicago, etc.

Vous pouvez aussi télécharger le texte intégral de la publication scolaire au format pdf et consulter son résumé en ligne lorsque ces informations sont inclues dans les métadonnées.

Parcourez les articles de revues sur diverses disciplines et organisez correctement votre bibliographie.

1

Karthick Chaganty, Siva. « Database Failure Prediction Based on Deep Learning Model ». International Journal of Science and Research (IJSR) 10, no 4 (27 avril 2021) : 83–86. https://doi.org/10.21275/sr21329110526.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
2

Wang, Wei, Meihui Zhang, Gang Chen, H. V. Jagadish, Beng Chin Ooi et Kian-Lee Tan. « Database Meets Deep Learning ». ACM SIGMOD Record 45, no 2 (28 septembre 2016) : 17–22. http://dx.doi.org/10.1145/3003665.3003669.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
3

Lukic, Vesna, et Marcus Brüggen. « Galaxy Classifications with Deep Learning ». Proceedings of the International Astronomical Union 12, S325 (octobre 2016) : 217–20. http://dx.doi.org/10.1017/s1743921316012771.

Texte intégral
Résumé :
AbstractMachine learning techniques have proven to be increasingly useful in astronomical applications over the last few years, for example in object classification, estimating redshifts and data mining. One example of object classification is classifying galaxy morphology. This is a tedious task to do manually, especially as the datasets become larger with surveys that have a broader and deeper search-space. The Kaggle Galaxy Zoo competition presented the challenge of writing an algorithm to find the probability that a galaxy belongs in a particular class, based on SDSS optical spectroscopy data. The use of convolutional neural networks (convnets), proved to be a popular solution to the problem, as they have also produced unprecedented classification accuracies in other image databases such as the database of handwritten digits (MNIST †) and large database of images (CIFAR ‡). We experiment with the convnets that comprised the winning solution, but using broad classifications. The effect of changing the number of layers is explored, as well as using a different activation function, to help in developing an intuition of how the networks function and to see how they can be applied to radio galaxy images.
Styles APA, Harvard, Vancouver, ISO, etc.
4

Liu, Rukun, Teng Wang, Yuxue Yang et Bingjie Yu. « Database Development Based on Deep Learning and Cloud Computing ». Mobile Information Systems 2022 (29 avril 2022) : 1–10. http://dx.doi.org/10.1155/2022/6208678.

Texte intégral
Résumé :
In this research, the author develops databases based on deep learning and cloud computing technology. On the basis of designing the overall architecture of the database with the distributed C/S mode as the core, use J2EE (Java 2 Platform, Enterprise Edition) as the development tool, apply Oracle server database, extract data features with in-depth learning technology, allocate data processing tasks based with cloud computing technology, so as to finally complete data fusion and compression. Finally, the overall development of the database is completed by designing the database backup scheme and external encryption. The test results show that the database developed by the above method has low performance loss, can quickly complete the processing of subdatabase and subtable, and can effectively support the distributed storage of data.
Styles APA, Harvard, Vancouver, ISO, etc.
5

Zhou, Lixi, Jiaqing Chen, Amitabh Das, Hong Min, Lei Yu, Ming Zhao et Jia Zou. « Serving deep learning models with deduplication from relational databases ». Proceedings of the VLDB Endowment 15, no 10 (juin 2022) : 2230–43. http://dx.doi.org/10.14778/3547305.3547325.

Texte intégral
Résumé :
Serving deep learning models from relational databases brings significant benefits. First, features extracted from databases do not need to be transferred to any decoupled deep learning systems for inferences, and thus the system management overhead can be significantly reduced. Second, in a relational database, data management along the storage hierarchy is fully integrated with query processing, and thus it can continue model serving even if the working set size exceeds the available memory. Applying model deduplication can greatly reduce the storage space, memory footprint, cache misses, and inference latency. However, existing data deduplication techniques are not applicable to the deep learning model serving applications in relational databases. They do not consider the impacts on model inference accuracy as well as the inconsistency between tensor blocks and database pages. This work proposed synergistic storage optimization techniques for duplication detection, page packing, and caching, to enhance database systems for model serving. Evaluation results show that our proposed techniques significantly improved the storage efficiency and the model inference latency, and outperformed existing deep learning frameworks in targeting scenarios.
Styles APA, Harvard, Vancouver, ISO, etc.
6

Baimakhanova, A. S., K. M. Berkimbayev, A. K. Zhumadillayeva et E. T. Abdrashova. « Technology of using deep learning algorithms ». Bulletin of the National Engineering Academy of the Republic of Kazakhstan 89, no 3 (15 septembre 2023) : 35–45. http://dx.doi.org/10.47533/2023.1606-146x.30.

Texte intégral
Résumé :
Deep learning is a branch of machine learning (machine learning-ML). Deep learning methods utilize high-level model abstraction of nonlinear transformations in large databases. In other areas, the implementation of deep learning architectures has contributed significantly to the development of artificial intelligence. This paper presents recent research on newly applied deep learning algorithms. Convolutional Neural Networks are used in deep learning. Database Management System PostgreSQL object-relational database. The implementation resulted in achieving the set goals and objectives. The method of analyzing the input data is described, the differences between machine learning and deep learning are explained, and an example of classifying an image representing a sign language image using logistic regression, one of the deep learning algorithms, is presented. Deep neural networks can work with the full set of available data better than alternative approaches. During the learning process, the neural network itself determines which features in the data are important and which are not. Artificial neural networks can predict symptoms that humans cannot. Thus, with the help of deep neural networks, we can solve problems that traditional machine learning algorithms cannot perform.
Styles APA, Harvard, Vancouver, ISO, etc.
7

Oh, Jaeho, Mincheol Kim et Sang-Woo Ban. « Deep Learning Model with Transfer Learning to Infer Personal Preferences in Images ». Applied Sciences 10, no 21 (29 octobre 2020) : 7641. http://dx.doi.org/10.3390/app10217641.

Texte intégral
Résumé :
In this paper, we propose a deep convolutional neural network model with transfer learning that reflects personal preferences from inter-domain databases of images having atypical visual characteristics. The proposed model utilized three public image databases (Fashion-MNIST, Labeled Faces in the Wild [LFW], and Indoor Scene Recognition) that include images with atypical visual characteristics in order to train and infer personal visual preferences. The effectiveness of transfer learning for incremental preference learning was verified by experiments using inter-domain visual datasets with different visual characteristics. Moreover, a gradient class activation mapping (Grad-CAM) approach was applied to the proposed model, providing explanations about personal visual preference possibilities. Experiments showed that the proposed preference-learning model using transfer learning outperformed a preference model not using transfer learning. In terms of the accuracy of preference recognition, the proposed model showed a maximum of about 7.6% improvement for the LFW database and a maximum of about 9.4% improvement for the Indoor Scene Recognition database, compared to the model that did not reflect transfer learning.
Styles APA, Harvard, Vancouver, ISO, etc.
8

Maji, Subhadip, et Smarajit Bose. « CBIR Using Features Derived by Deep Learning ». ACM/IMS Transactions on Data Science 2, no 3 (31 août 2021) : 1–24. http://dx.doi.org/10.1145/3470568.

Texte intégral
Résumé :
In a Content-based Image Retrieval (CBIR) System, the task is to retrieve similar images from a large database given a query image. The usual procedure is to extract some useful features from the query image and retrieve images that have a similar set of features. For this purpose, a suitable similarity measure is chosen, and images with high similarity scores are retrieved. Naturally, the choice of these features play a very important role in the success of this system, and high-level features are required to reduce the “semantic gap.” In this article, we propose to use features derived from pre-trained network models from a deep-learning convolution network trained for a large image classification problem. This approach appears to produce vastly superior results for a variety of databases, and it outperforms many contemporary CBIR systems. We analyse the retrieval time of the method and also propose a pre-clustering of the database based on the above-mentioned features, which yields comparable results in a much shorter time in most of the cases.
Styles APA, Harvard, Vancouver, ISO, etc.
9

Zhou, Xiaoshu, Qide Xiao et Han Wang. « Metamaterials Design Method based on Deep learning Database ». Journal of Physics : Conference Series 2185, no 1 (1 janvier 2022) : 012023. http://dx.doi.org/10.1088/1742-6596/2185/1/012023.

Texte intégral
Résumé :
Abstract In recent years, deep learning has risen to the forefront of many fields, overcoming challenges previously considered difficult to solve by traditional methods. In the field of metamaterials, there are significant challenges in the design and optimization of metamaterials, including the need for a large number of labeled data sets and one-to-many mapping when solving inverse problems. Here, we will use deep learning methods to build a metamaterial database to achieve rapid design and analysis methods of metamaterials. These technologies have significantly improved the feasibility of more complex metamaterial designs and provided new metamaterial design and analysis ideas.
Styles APA, Harvard, Vancouver, ISO, etc.
10

Liu, Yue, Rashmi Sharan Sinha, Shu-Zhi Liu et Seung-Hoon Hwang. « Side-Information-Aided Preprocessing Scheme for Deep-Learning Classifier in Fingerprint-Based Indoor Positioning ». Electronics 9, no 6 (12 juin 2020) : 982. http://dx.doi.org/10.3390/electronics9060982.

Texte intégral
Résumé :
Deep-learning classifiers can effectively improve the accuracy of fingerprint-based indoor positioning. During fingerprint database construction, all received signal strength indicators from each access point are combined without any distinction. Therefore, the database is created and utilised for deep-learning models. Meanwhile, side information regarding specific conditions may help characterise the data features for the deep-learning classifier and improve the accuracy of indoor positioning. Herein, a side-information-aided preprocessing scheme for deep-learning classifiers is proposed in a dynamic environment, where several groups of different databases are constructed for training multiple classifiers. Therefore, appropriate databases can be employed to effectively improve positioning accuracies. Specifically, two kinds of side information, namely time (morning/afternoon) and direction (forward/backward), are considered when collecting the received signal strength indicator. Simulations and experiments are performed with the deep-learning classifier trained on four different databases. Moreover, these are compared with conventional results from the combined database. The results show that the side-information-aided preprocessing scheme allows better success probability than the conventional method. With two margins, the proposed scheme has 6.55% and 5.8% improved performances for simulations and experiments compared to the conventional scheme. Additionally, the proposed scheme, with time as the side information, obtains a higher success probability when the positioning accuracy requirement is loose with larger margin. With direction as the side information, the proposed scheme shows better performance for high positioning precision requirements. Thus, side information such as time or direction is advantageous for preprocessing data in deep-learning classifiers for fingerprint-based indoor positioning.
Styles APA, Harvard, Vancouver, ISO, etc.
11

Blank, Sebastian, Florian Wilhelm, Hans-Peter Zorn et Achim Rettinger. « Querying NoSQL with Deep Learning to Answer Natural Language Questions ». Proceedings of the AAAI Conference on Artificial Intelligence 33 (17 juillet 2019) : 9416–21. http://dx.doi.org/10.1609/aaai.v33i01.33019416.

Texte intégral
Résumé :
Almost all of today’s knowledge is stored in databases and thus can only be accessed with the help of domain specific query languages, strongly limiting the number of people which can access the data. In this work, we demonstrate an end-to-end trainable question answering (QA) system that allows a user to query an external NoSQL database by using natural language. A major challenge of such a system is the non-differentiability of database operations which we overcome by applying policy-based reinforcement learning. We evaluate our approach on Facebook’s bAbI Movie Dialog dataset and achieve a competitive score of 84.2% compared to several benchmark models. We conclude that our approach excels with regard to real-world scenarios where knowledge resides in external databases and intermediate labels are too costly to gather for non-end-to-end trainable QA systems.
Styles APA, Harvard, Vancouver, ISO, etc.
12

Shen, Haijie, Yangyuan Li, Xinzhi Tian, Xiaofan Chen, Caihong Li, Qian Bian, Zhenduo Wang et Weihua Wang. « Mass data processing and multidimensional database management based on deep learning ». Open Computer Science 12, no 1 (1 janvier 2022) : 300–313. http://dx.doi.org/10.1515/comp-2022-0251.

Texte intégral
Résumé :
Abstract With the rapid development of the Internet of Things, the requirements for massive data processing technology are getting higher and higher. Traditional computer data processing capabilities can no longer deliver fast, simple, and efficient data analysis and processing for today’s massive data processing due to the real-time, massive, polymorphic, and heterogeneous characteristics of Internet of Things data. Mass heterogeneous data of different types of subsystems in the Internet of Things need to be processed and stored uniformly, so the mass data processing method is required to be able to integrate multiple different networks, multiple data sources, and heterogeneous mass data and be able to perform processing on these data. Therefore, this article proposes massive data processing and multidimensional database management based on deep learning to meet the needs of contemporary society for massive data processing. This article has deeply studied the basic technical methods of massive data processing, including MapReduce technology, parallel data technology, database technology based on distributed memory databases, and distributed real-time database technology based on cloud computing technology, and constructed a massive data fusion algorithm based on deep learning. The model and the multidimensional online analytical processing model of the multidimensional database based on deep learning analyze the performance, scalability, load balancing, data query, and other aspects of the multidimensional database based on deep learning. It is concluded that the accuracy of multidimensional database query data is as high as 100%, and the accuracy of the average data query time is only 0.0053 s, which is much lower than the general database query time.
Styles APA, Harvard, Vancouver, ISO, etc.
13

Kadhim, Ola Najah, et Mohammed Hasan Abdulameer. « A multimodal biometric database and case study for face recognition based deep learning ». Bulletin of Electrical Engineering and Informatics 13, no 1 (1 février 2024) : 677–85. http://dx.doi.org/10.11591/eei.v13i1.6605.

Texte intégral
Résumé :
Recently, multimodal biometric systems have garnered a lot of interest for the identification of human identity. The accessibility of the database is one of the contributing elements that impact biometric recognition systems. In their studies, the majority of researchers concentrate on unimodal databases. There was a need to compile a fresh, realistic multimodal biometric database, nonetheless, because there were so few comparable multimodal biometric databases that were publically accessible. This study introduces the MULBv1 multimodal biometric database, which contains homologous biometric traits. The MULBv1 database includes 20 images of each person's face in various poses, facial emotions, and accessories, 20 images of their right hand from various angles, and 20 images of their right iris from various lighting positions. The database contains real multimodal data from 174 people, and all biometrics were accurately collected using the micro camera of the iPhone 14 Pro Max. A face recognition technique is also suggested as a case study using the gathered facial features. In the case study, the deep convolutional neural network (CNN) was used, and the findings were positive. Through several trials, the accuracy was (97.41%).
Styles APA, Harvard, Vancouver, ISO, etc.
14

Li, Dana, Bolette Mikela Vilmun, Jonathan Frederik Carlsen, Elisabeth Albrecht-Beste, Carsten Ammitzbøl Lauridsen, Michael Bachmann Nielsen et Kristoffer Lindskov Hansen. « The Performance of Deep Learning Algorithms on Automatic Pulmonary Nodule Detection and Classification Tested on Different Datasets That Are Not Derived from LIDC-IDRI : A Systematic Review ». Diagnostics 9, no 4 (29 novembre 2019) : 207. http://dx.doi.org/10.3390/diagnostics9040207.

Texte intégral
Résumé :
The aim of this study was to systematically review the performance of deep learning technology in detecting and classifying pulmonary nodules on computed tomography (CT) scans that were not from the Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) database. Furthermore, we explored the difference in performance when the deep learning technology was applied to test datasets different from the training datasets. Only peer-reviewed, original research articles utilizing deep learning technology were included in this study, and only results from testing on datasets other than the LIDC-IDRI were included. We searched a total of six databases: EMBASE, PubMed, Cochrane Library, the Institute of Electrical and Electronics Engineers, Inc. (IEEE), Scopus, and Web of Science. This resulted in 1782 studies after duplicates were removed, and a total of 26 studies were included in this systematic review. Three studies explored the performance of pulmonary nodule detection only, 16 studies explored the performance of pulmonary nodule classification only, and 7 studies had reports of both pulmonary nodule detection and classification. Three different deep learning architectures were mentioned amongst the included studies: convolutional neural network (CNN), massive training artificial neural network (MTANN), and deep stacked denoising autoencoder extreme learning machine (SDAE-ELM). The studies reached a classification accuracy between 68–99.6% and a detection accuracy between 80.6–94%. Performance of deep learning technology in studies using different test and training datasets was comparable to studies using same type of test and training datasets. In conclusion, deep learning was able to achieve high levels of accuracy, sensitivity, and/or specificity in detecting and/or classifying nodules when applied to pulmonary CT scans not from the LIDC-IDRI database.
Styles APA, Harvard, Vancouver, ISO, etc.
15

Abbass, Ghida Yousif, et Ali Fadhil Marhoon. « Car license plate segmentation and recognition system based on deep learning ». Bulletin of Electrical Engineering and Informatics 11, no 4 (1 août 2022) : 1983–89. http://dx.doi.org/10.11591/eei.v11i4.3434.

Texte intégral
Résumé :
Artificial intelligence techniques and computer vision techniques dealt with the issue of automatic license plate recognition (ALPR) that has many applications in important research field. In this paper, the method of recognizing the license plates of Iraqi cars was applied based on deep learning techniques convolutional neural network (CNN). The two database built to identifying Iraqi car plates. First database includes 2000 images of Arabic numbers and Arabic letters. Second database conations 1200 images of the Arabic names for Iraqi governorates. This paper used image-processing techniques to segmenting the numbers, letters and words from the car license plate images and then convert them into two databases that used to train the two CNN. These training CNN used to recognizing the vocabulary of the car license plate. The success rate of the numbers, letters and words recognition was 98%. The overall rate of success of this proposed system in all stages was 97%.
Styles APA, Harvard, Vancouver, ISO, etc.
16

Giriprasad Gaddam, P., A. Sanjeeva reddy et R. V. Sreehari. « Automatic Classification of Cardiac Arrhythmias based on ECG Signals Using Transferred Deep Learning Convolution Neural Network ». Journal of Physics : Conference Series 2089, no 1 (1 novembre 2021) : 012058. http://dx.doi.org/10.1088/1742-6596/2089/1/012058.

Texte intégral
Résumé :
Abstract In the current article, an automatic classification of cardiac arrhythmias is presented using a transfer deep learning approach with the help of electrocardiography (ECG) signal analysis. Now a days, an ECG waveform serves as a powerful tool used for the analysis of cardiac arrhythmias (irregularities). The goal of the present work is to implement an algorithm based on deep learning for classification of different cardiac arrhythmias. Initially, the one dimensional (1-D) ECG signals are transformed to two dimensional (2-D) scalogram images with the help of Continuous Wavelet(CWT). Four different categories of ECG waveform were selected from four PhysioNet MIT-BIH databases, namely arrhythmia database, Normal Sinus Rhythm database, Malignant Ventricular Ectopy database and BIDMC Congestive heart failure database to examine the proposed technique. The major interest of the present study is to develop a transferred deep learning algorithm for automatic categorization of the mentioned four different heart diseases. Final results proved that the 2-D scalogram images trained with a deep convolutional neural network CNN with transfer learning technique (AlexNet) pepped up with a prominent accuracy of 95.67%. Hence, it is worthwhile to say the above stated algorithm demonstrates as an effective automated heart disease detection tool
Styles APA, Harvard, Vancouver, ISO, etc.
17

Shang, Xiaoran. « Database Oriented Big Data Analysis Engine Based on Deep Learning ». Computational Intelligence and Neuroscience 2022 (31 août 2022) : 1–9. http://dx.doi.org/10.1155/2022/4500684.

Texte intégral
Résumé :
In recent years, with the development of enterprises to the Internet, the demand for cloud database is also growing, especially how to capture data quickly and efficiently through the database. In order to improve the data structure at all levels in the process of database analysis engine, this paper realizes the accurate construction and rapid analysis of cloud database based on big data analysis engine technology and deep learning wolf pack greedy algorithm. Through the deep learning strategy, a big data analysis engine system based on the deep learning model is constructed. The functions of deep learning technology, wolf greedy algorithm, and data analysis strategy in the cloud database analysis engine system are analyzed, as well as the functions of the whole analysis engine system. Finally, the accuracy and response speed of the cloud database analysis engine system are tested according to the known clustering data. The results show that compared with the traditional data analysis engine system with character search as the core, the database oriented big data analysis engine system based on a deep learning model and wolf swarm greedy algorithm has faster response speed and intelligence. The research application is that the proposed engine system can significantly improve the effect of the analysis engine and greatly improve the retrieval accuracy and analysis efficiency of fixed-point data in the database.
Styles APA, Harvard, Vancouver, ISO, etc.
18

Manoj krishna, M., M. Neelima, M. Harshali et M. Venu Gopala Rao. « Image classification using Deep learning ». International Journal of Engineering & ; Technology 7, no 2.7 (18 mars 2018) : 614. http://dx.doi.org/10.14419/ijet.v7i2.7.10892.

Texte intégral
Résumé :
The image classification is a classical problem of image processing, computer vision and machine learning fields. In this paper we study the image classification using deep learning. We use AlexNet architecture with convolutional neural networks for this purpose. Four test images are selected from the ImageNet database for the classification purpose. We cropped the images for various portion areas and conducted experiments. The results show the effectiveness of deep learning based image classification using AlexNet.
Styles APA, Harvard, Vancouver, ISO, etc.
19

Yan, Yu, Shun Yao, Hongzhi Wang et Meng Gao. « Index selection for NoSQL database with deep reinforcement learning ». Information Sciences 561 (juin 2021) : 20–30. http://dx.doi.org/10.1016/j.ins.2021.01.003.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
20

Cohen, William, Fan Yang et Kathryn Rivard Mazaitis. « TensorLog : A Probabilistic Database Implemented Using Deep-Learning Infrastructure ». Journal of Artificial Intelligence Research 67 (23 février 2020) : 285–325. http://dx.doi.org/10.1613/jair.1.11944.

Texte intégral
Résumé :
We present an implementation of a probabilistic first-order logic called TensorLog, in which classes of logical queries are compiled into differentiable functions in a neural-network infrastructure such as Tensorflow or Theano. This leads to a close integration of probabilistic logical reasoning with deep-learning infrastructure: in particular, it enables high-performance deep learning frameworks to be used for tuning the parameters of a probabilistic logic. The integration with these frameworks enables use of GPU-based parallel processors for inference and learning, making TensorLog the first highly parallellizable probabilistic logic. Experimental results show that TensorLog scales to problems involving hundreds of thousands of knowledge-base triples and tens of thousands of examples.
Styles APA, Harvard, Vancouver, ISO, etc.
21

Gattan, Atif M. « Deep Learning Technique of Sentiment Analysis for Twitter Database ». International Journal of Interactive Mobile Technologies (iJIM) 16, no 01 (18 janvier 2022) : 184–93. http://dx.doi.org/10.3991/ijim.v16i01.27575.

Texte intégral
Résumé :
Sentiment investigation is the progression of calculating, recognizing as well as classifying the people mentality stated in the outline of the manuscript and the outlook of the individual concerning the particular subject matter can be investigated with the assist of that statistics. The information is accumulated by way of the assistance of API in favor of classifying of the end outcome based on the information collected as in Negative way, Neutral way and Positive way with the support of the scoring polarity allocated in favor of every statement that are collected. These data are fruitful for discovering and enhancing the consumer requirements and to acquire better tune-up. Major improvement of using the concept of sentiment analysis is to develop the user needs by straightforwardly collecting the information from the outsized set of consumers.
Styles APA, Harvard, Vancouver, ISO, etc.
22

Xu, Boyan, Ruichu Cai, Zhenjie Zhang, Xiaoyan Yang, Zhifeng Hao, Zijian Li et Zhihao Liang. « NADAQ : Natural Language Database Querying Based on Deep Learning ». IEEE Access 7 (2019) : 35012–17. http://dx.doi.org/10.1109/access.2019.2904720.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
23

Jamshed, Aatif, Bhawna Mallick et Pramod Kumar. « Deep learning-based sequential pattern mining for progressive database ». Soft Computing 24, no 22 (13 mai 2020) : 17233–46. http://dx.doi.org/10.1007/s00500-020-05015-2.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
24

Khan, Khalil, Byeong-hee Roh, Jehad Ali, Rehan Ullah Khan, Irfan Uddin, Saqlain Hassan, Rabia Riaz et Nasir Ahmad. « PHND : Pashtu Handwritten Numerals Database and deep learning benchmark ». PLOS ONE 15, no 9 (2 septembre 2020) : e0238423. http://dx.doi.org/10.1371/journal.pone.0238423.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
25

Kiran, Aqsa, Shahzad Ahmad Qureshi, Asifullah Khan, Sajid Mahmood, Muhammad Idrees, Aqsa Saeed, Muhammad Assam, Mohamad Reda A. Refaai et Abdullah Mohamed. « Reverse Image Search Using Deep Unsupervised Generative Learning and Deep Convolutional Neural Network ». Applied Sciences 12, no 10 (13 mai 2022) : 4943. http://dx.doi.org/10.3390/app12104943.

Texte intégral
Résumé :
Reverse image search has been a vital and emerging research area of information retrieval. One of the primary research foci of information retrieval is to increase the space and computational efficiency by converting a large image database into an efficiently computed feature database. This paper proposes a novel deep learning-based methodology, which captures channel-wise, low-level details of each image. In the first phase, sparse auto-encoder (SAE), a deep generative model, is applied to RGB channels of each image for unsupervised representational learning. In the second phase, transfer learning is utilized by using VGG-16, a variant of deep convolutional neural network (CNN). The output of SAE combined with the original RGB channel is forwarded to VGG-16, thereby producing a more effective feature database by the ensemble/collaboration of two effective models. The proposed method provides an information rich feature space that is a reduced dimensionality representation of the image database. Experiments are performed on a hybrid dataset that is developed by combining three standard publicly available datasets. The proposed approach has a retrieval accuracy (precision) of 98.46%, without using the metadata of images, by using a cosine similarity measure between the query image and the image database. Additionally, to further validate the proposed methodology’s effectiveness, image quality has been degraded by adding 5% noise (Speckle, Gaussian, and Salt pepper noise types) in the hybrid dataset. Retrieval accuracy has generally been found to be 97% for different variants of noise
Styles APA, Harvard, Vancouver, ISO, etc.
26

Bianchi, Alexander, Andrew Chai, Vincent Corvinelli, Parke Godfrey, Jarek Szlichta et Calisto Zuzarte. « Db2une : Tuning Under Pressure via Deep Learning ». Proceedings of the VLDB Endowment 17, no 12 (août 2024) : 3855–68. http://dx.doi.org/10.14778/3685800.3685811.

Texte intégral
Résumé :
Modern database systems including IBM Db2 have numerous parameters, "knobs," that require precise configuration to achieve optimal workload performance. Even for experts, manually "tuning" these knobs is a challenging process. We present Db2une, an automatic query-aware tuning system that leverages deep learning to maximize performance while minimizing resource usage. Via a specialized transformer-based query-embedding pipeline we name QBERT, Db2une generates context-aware representations of query workloads to feed as input to a stability-oriented, on-policy deep reinforcement learning model. In Db2une, we introduce a multi-phased, database meta-data driven training approach---which incorporates cost estimates, interpolation of these costs, and database statistics---to efficiently discover optimal tuning configurations without the need to execute queries. Thus, our model can scale to very large workloads, for which executing queries would be prohibitively expensive. Through experimental evaluation, we demonstrate Db2une's efficiency and effectiveness over a variety of workloads. We compare it against the state-of-the-art query-aware tuning systems and show that the system provides recommendations that surpass those of IBM experts.
Styles APA, Harvard, Vancouver, ISO, etc.
27

Kshirod, Kshirod Sarmah. « Speaker Diarization with Deep Learning Techniques ». Turkish Journal of Computer and Mathematics Education (TURCOMAT) 11, no 3 (15 décembre 2020) : 2570–82. http://dx.doi.org/10.61841/turcomat.v11i3.14309.

Texte intégral
Résumé :
Speaker diarization is a task to identify the speaker when different speakers spoke in an audio or video recording environment. Artificial intelligence (AI) fields have effectively used Deep Learning (DL) to solve a variety of real-world application challenges. With effective applications in a wide range of subdomains, such as natural language processing, image processing, computer vision, speech and speaker recognition, and emotion recognition, cyber security, and many others, DL, a very innovative field of Machine Learning (ML), that is quickly emerging as the most potent machine learning technique. DL techniques have outperformed conventional approaches recently in speaker diarization as well as speaker recognition. The technique of assigning classes to speech recordings that correspond to the speaker's identity is known as speaker diarization, and it allows one to determine who spoke when. A crucial step in speech processing is speaker diarization, which divides an audio recording into different speaker areas. In-depth analysis of speaker diarization utilizing a variety of deep learning algorithms that are presented in this research paper. NIST-2000 CALLHOME and our in-house database ALSD-DB are the two voice corpora we used for this study's tests. TDNN-based embeddings with x-vectors, LSTM-based embeddings with d-vectors, and lastly embeddings fusion of both x-vector and d-vector are used in the tests for the basic system. For the NIST-2000 CALLHOME database, LSTM based embeddings with d-vector and embeddings integrating both x-vector and d-vector exhibit improved performance with DER of 8.25% and 7.65%, respectively, and of 10.45% and 9.65% for the local ALSD-DB database
Styles APA, Harvard, Vancouver, ISO, etc.
28

Meng, Yang, Guoxin Liang et Mei Yue. « Deep Learning-Based Arrhythmia Detection in Electrocardiograph ». Scientific Programming 2021 (13 mai 2021) : 1–7. http://dx.doi.org/10.1155/2021/9926769.

Texte intégral
Résumé :
This study aimed to explore the application of electrocardiograph (ECG) in the diagnosis of arrhythmia based on the deep convolutional neural network (DCNN). ECG was classified and recognized with the DCNN. The specificity (Spe), sensitivity (Sen), accuracy (Acc), and area under curve (AUC) of the DCNN were evaluated in the Chinese Cardiovascular Disease Database (CCDD) and Massachusetts Institute of Technology-Beth Israel Hospital (MIT-BIH) arrhythmia database, respectively. The results showed that in the CCDD, the original model tested by the small sample set had an accuracy (Acc) of 82.78% and AUC of 0.882, while the Acc and AUC of the translated model were 85.69% and 0.893, respectively, so the difference was notable ( P < 0.05); the Acc of the original model and the translated model was 80.12% and 82.63%, respectively, in the large sample set, so the difference was obvious ( P < 0.05). In the MIT-BIH database, the Acc of normal (N) heart beat (HB) (99.38%) was higher than that of the atrial premature beat (APB) (87.45%) ( P < 0.05). In a word, applying the DCNN could improve the Acc of ECG for classification and recognition, so it could be well applied to ECG signal classification.
Styles APA, Harvard, Vancouver, ISO, etc.
29

Joshi, Vaishali M., Deepthi D. Kulkarni et Nilesh J. Uke. « Stress and anxiety detection : deep learning and higher order statistic approach ». Indonesian Journal of Electrical Engineering and Computer Science 33, no 3 (1 mars 2024) : 1567. http://dx.doi.org/10.11591/ijeecs.v33.i3.pp1567-1575.

Texte intégral
Résumé :
<p>Today's teenagers are dealing with anxiety and stress. Anxiety, depression, and suicide rates have increased in recent years because of increased social rivalry. The research is focused on detecting anxiety in students due to exam pressure to reduce the potential harm to a person's wellness. Research is performed on databases for anxious states based on psychological stimulation (DASPS) and our own database. The measured signal is divided into sub bands that correspond to the electroencephalogram (EEG) rhythms using the Butterworth sixth-order order filter. In higher dimensional space, the nonlinearities of each sub-band signal are analyzed using higher order statistics third-order cumulants (TOC). We have classified stress and anxiety using the support vector machine (SVM), K-nearest neighbor (K-NN), and deep learning bidirectional long short-term memory (BiLSTM) network. In comparison to previous techniques, the proposed system's performance using BiLSTM is quite good. The best accuracy in this analysis was 87% on the DASPS database and 98% on the own database. Finally, subjects with high stress levels had more gamma activity than subjects with little stress. This could be an important attribute in the classification of stress.</p>
Styles APA, Harvard, Vancouver, ISO, etc.
30

Sayed Farag, Mohamed, Mostafa Mohamed Mohie El Din et Hassan Ahmed Elshenbary. « Deep learning versus traditional methods for parking lots occupancy classification ». Indonesian Journal of Electrical Engineering and Computer Science 19, no 2 (1 août 2020) : 964. http://dx.doi.org/10.11591/ijeecs.v19.i2.pp964-973.

Texte intégral
Résumé :
<span>Due to the increase in number of cars and slow city developments, there is a need for smart parking system. One of the main issues in smart parking systems is parking lot occupancy status classification, so this paper introduce two methods for parking lot classification. The first method uses the mean, after converting the colored image to grayscale, then to black/white. If the mean is greater than a given threshold it is classified as occupied, otherwise it is empty. This method gave 90% correct classification rate on cnrall database. It overcome the alexnet deep learning method trained and tested on the same database (the mean method has no training time). The second method, which depends on deep learning is a deep learning neural network consists of 11 layers, trained and tested on the same database. It gave 93% correct classification rate, when trained on cnrall and tested on the same database. As shown, this method overcome the alexnet deep learning and the mean methods on the same database. On the Pklot database the alexnet and our deep learning network have a close resutls, overcome <br /> the mean method (greater than 95%).</span>
Styles APA, Harvard, Vancouver, ISO, etc.
31

Najeeb, Shaima Miqdad Mohamed, Raid Rafi Omar Al-Nima et Mohand Lokman Ahmad Al-Dabag. « Reinforced Deep Learning for Verifying Finger Veins ». International Journal of Online and Biomedical Engineering (iJOE) 17, no 07 (2 juillet 2021) : 19. http://dx.doi.org/10.3991/ijoe.v17i07.24655.

Texte intégral
Résumé :
Recently, personal verifications become crucial demands for providing securities in personal accounts and financial activities. This paper suggests a new Deep Learning (DL) model called the Re-enforced Deep Learning (RDL). This approach provides another way of personal verification by using the Finger Veins (FVs). The RDL consists of multiple layers with a feedback. Two FV fingers are employed for each person, FV of the index finger for first personal verification and FV of the middle finger for re-enforced verification. The used database is from the Hong Kong Polytechnic University Finger Image (PolyUFI) database (Version 1.0). The result shows that the proposed RDL achieved a promising performance of 91.19%. Also, other DL approaches are exploited for comparisons in this study including state-of-the-art models.
Styles APA, Harvard, Vancouver, ISO, etc.
32

Ullah, Wusat, Imran Siddique, Rana Muhammad Zulqarnain, Mohammad Mahtab Alam, Irfan Ahmad et Usman Ahmad Raza. « Classification of Arrhythmia in Heartbeat Detection Using Deep Learning ». Computational Intelligence and Neuroscience 2021 (19 octobre 2021) : 1–13. http://dx.doi.org/10.1155/2021/2195922.

Texte intégral
Résumé :
The electrocardiogram (ECG) is one of the most widely used diagnostic instruments in medicine and healthcare. Deep learning methods have shown promise in healthcare prediction challenges involving ECG data. This paper aims to apply deep learning techniques on the publicly available dataset to classify arrhythmia. We have used two kinds of the dataset in our research paper. One dataset is the MIT-BIH arrhythmia database, with a sampling frequency of 125 Hz with 1,09,446 ECG beats. The classes included in this first dataset are N, S, V, F, and Q. The second database is PTB Diagnostic ECG Database. The second database has two classes. The techniques used in these two datasets are the CNN model, CNN + LSTM, and CNN + LSTM + Attention Model. 80% of the data is used for the training, and the remaining 20% is used for testing. The result achieved by using these three techniques shows the accuracy of 99.12% for the CNN model, 99.3% for CNN + LSTM, and 99.29% for CNN + LSTM + Attention Model.
Styles APA, Harvard, Vancouver, ISO, etc.
33

Xiao, Qiao, Khuan Lee, Siti Aisah Mokhtar, Iskasymar Ismail, Ahmad Luqman bin Md Pauzi, Qiuxia Zhang et Poh Ying Lim. « Deep Learning-Based ECG Arrhythmia Classification : A Systematic Review ». Applied Sciences 13, no 8 (14 avril 2023) : 4964. http://dx.doi.org/10.3390/app13084964.

Texte intégral
Résumé :
Deep learning (DL) has been introduced in automatic heart-abnormality classification using ECG signals, while its application in practical medical procedures is limited. A systematic review is performed from perspectives of the ECG database, preprocessing, DL methodology, evaluation paradigm, performance metric, and code availability to identify research trends, challenges, and opportunities for DL-based ECG arrhythmia classification. Specifically, 368 studies meeting the eligibility criteria are included. A total of 223 (61%) studies use MIT-BIH Arrhythmia Database to design DL models. A total of 138 (38%) studies considered removing noise or artifacts in ECG signals, and 102 (28%) studies performed data augmentation to extend the minority arrhythmia categories. Convolutional neural networks are the dominant models (58.7%, 216) used in the reviewed studies while growing studies have integrated multiple DL structures in recent years. A total of 319 (86.7%) and 38 (10.3%) studies explicitly mention their evaluation paradigms, i.e., intra- and inter-patient paradigms, respectively, where notable performance degradation is observed in the inter-patient paradigm. Compared to the overall accuracy, the average F1 score, sensitivity, and precision are significantly lower in the selected studies. To implement the DL-based ECG classification in real clinical scenarios, leveraging diverse ECG databases, designing advanced denoising and data augmentation techniques, integrating novel DL models, and deeper investigation in the inter-patient paradigm could be future research opportunities.
Styles APA, Harvard, Vancouver, ISO, etc.
34

Yao, Ge. « Application of Higher Education Management in Colleges and Universities by Deep Learning ». Computational Intelligence and Neuroscience 2022 (10 août 2022) : 1–9. http://dx.doi.org/10.1155/2022/7295198.

Texte intégral
Résumé :
The development of artificial intelligence (AI) has brought great convenience to people and has been widely used in the field of education. To monitor the classroom status of college students in real time and achieve the purpose of balanced distribution of educational resources, the facial expression recognition (FER) algorithm is applied to the management of higher education in universities. Firstly, the convolutional neural network (CNN) is studied in depth, and secondly, the process and method of FER are explored in detail, and an adaptive FER algorithm based on the differential convolutional neural network (DCNN) is constructed. Finally, the algorithm is applied to the CK + database and the BU-4DFE database. The results manifest that the designed algorithm has an accuracy of 99.02% for keyframe detection in the CK + database and 98.35% for the BU-4DFE database. The algorithm has a high accuracy of keyframe detection for both expression databases. It has a good effect on the automatic detection of keyframes of expression sequences and can reach a level similar to that of manual frame selection. Compared with the existing algorithms, the proposed method still has higher advantages. It can effectively eliminate the interference of individual differences and environmental noise on FER. Experiments reveal that the proposed FER algorithm DCNN-based has a good recognition effect and is suitable for monitoring students’ classroom status. This research has certain reference significance for the application of AI in higher education management in colleges and universities.
Styles APA, Harvard, Vancouver, ISO, etc.
35

Tirado-Martin, Paloma, et Raul Sanchez-Reillo. « BioECG : Improving ECG Biometrics with Deep Learning and Enhanced Datasets ». Applied Sciences 11, no 13 (24 juin 2021) : 5880. http://dx.doi.org/10.3390/app11135880.

Texte intégral
Résumé :
Nowadays, Deep Learning tools have been widely applied in biometrics. Electrocardiogram (ECG) biometrics is not the exception. However, the algorithm performances rely heavily on a representative dataset for training. ECGs suffer constant temporal variations, and it is even more relevant to collect databases that can represent these conditions. Nonetheless, the restriction in database publications obstructs further research on this topic. This work was developed with the help of a database that represents potential scenarios in biometric recognition as data was acquired in different days, physical activities and positions. The classification was implemented with a Deep Learning network, BioECG, avoiding complex and time-consuming signal transformations. An exhaustive tuning was completed including variations in enrollment length, improving ECG verification for more complex and realistic biometric conditions. Finally, this work studied one-day and two-days enrollments and their effects. Two-days enrollments resulted in huge general improvements even when verification was accomplished with more unstable signals. EER was improved in 63% when including a change of position, up to almost 99% when visits were in a different day and up to 91% if the user experienced a heartbeat increase after exercise.
Styles APA, Harvard, Vancouver, ISO, etc.
36

Rahul et Deepika Bansal. « Object Detection Using Machine Learning and Deep Learning ». International Journal for Research in Applied Science and Engineering Technology 11, no 2 (28 février 2023) : 265–68. http://dx.doi.org/10.22214/ijraset.2023.48958.

Texte intégral
Résumé :
Abstract: An object detection system finds objects of the real-world present either in a digital image or an object detection system locates real-world items that are present in a digital image or a video. These objects can be any type of object, such as people, automobiles, or other objects. A model database, a feature detector, a hypothesis, and a hypothesis verifier are the four components that the system must have in order to successfully detect an item in an image or video. This paper provides an overview of the many methods for object detection, localization, classification, feature extraction, appearance information extraction, and many other tasks in photos and videos. The remarks are derived from the literature that has been analyzed, and important concerns are also noted
Styles APA, Harvard, Vancouver, ISO, etc.
37

Sudars, K. « Face recognition Face2vec based on deep learning : Small database case ». Automatic Control and Computer Sciences 51, no 1 (janvier 2017) : 50–54. http://dx.doi.org/10.3103/s0146411617010072.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
38

Zhang, Mingming, Zhigang Chen, Huiyu Wang, Zeng zeng et Xinwen Shan. « Research on Database Failure Prediction Based on Deep Learning Model ». IOP Conference Series : Materials Science and Engineering 452 (13 décembre 2018) : 032056. http://dx.doi.org/10.1088/1757-899x/452/3/032056.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
39

Ruescas Nicolau, A. V., E. Parrilla Bernabé, E. Medina Ripoll, D. Garrido Jaén et S. Alemany Mut. « Database generation for markerless tracking based on Deep Learning networks ». Gait & ; Posture 81 (septembre 2020) : 308–9. http://dx.doi.org/10.1016/j.gaitpost.2020.08.050.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
40

Wichmann, Andreas, Amgad Agoub, Valentina Schmidt et Martin Kada. « RoofN3D : A Database for 3D Building Reconstruction with Deep Learning ». Photogrammetric Engineering & ; Remote Sensing 85, no 6 (1 juin 2019) : 435–43. http://dx.doi.org/10.14358/pers.85.6.435.

Texte intégral
Styles APA, Harvard, Vancouver, ISO, etc.
41

Li, Zhongliang, Yaofeng Tu et Zongmin Ma. « A Sample-Aware Database Tuning System With Deep Reinforcement Learning ». Journal of Database Management 35, no 1 (9 novembre 2023) : 1–25. http://dx.doi.org/10.4018/jdm.333519.

Texte intégral
Résumé :
Based on the relationship between client load and overall system performance, the authors propose a sample-aware deep deterministic policy gradient model. Specifically, they improve sample quality by filtering out sample noise caused by the fluctuations of client load, which accelerates the model convergence speed of the intelligent tuning system and improves the tuning effect. Also, the hardware resources and client load consumed by the database in the working process are added to the model for training. This can enhance the performance characterization ability of the model and improve the recommended parameters of the algorithm. Meanwhile, they propose an improved closed-loop distributed comprehensive training architecture of online and offline training to quickly obtain high-quality samples and improve the efficiency of parameter tuning. Experimental results show that the configuration parameters can make the performance of the database system better and shorten the tuning time.
Styles APA, Harvard, Vancouver, ISO, etc.
42

Wu, Zhaohui, Lu Jiang, Qinghua Zheng et Jun Liu. « Learning to Surface Deep Web Content ». Proceedings of the AAAI Conference on Artificial Intelligence 24, no 1 (5 juillet 2010) : 1967–68. http://dx.doi.org/10.1609/aaai.v24i1.7779.

Texte intégral
Résumé :
We propose a novel deep web crawling framework based on reinforcement learning. The crawler is regarded as an agent and deep web database as the environment. The agent perceives its current state and submits a selected action (query) to the environment according to Q-value. Based on the framework we develop an adaptive crawling method. Experimental results show that it outperforms the state of art methods in crawling capability and breaks through the assumption of full-text search implied by existing methods.
Styles APA, Harvard, Vancouver, ISO, etc.
43

Rosado, Eduardo, Miguel Garcia-Remesal, Sergio Paraiso-Medina, Alejandro Pazos et Victor Maojo. « Using Machine Learning to Collect and Facilitate Remote Access to Biomedical Databases : Development of the Biomedical Database Inventory ». JMIR Medical Informatics 9, no 2 (25 février 2021) : e22976. http://dx.doi.org/10.2196/22976.

Texte intégral
Résumé :
Background Currently, existing biomedical literature repositories do not commonly provide users with specific means to locate and remotely access biomedical databases. Objective To address this issue, we developed the Biomedical Database Inventory (BiDI), a repository linking to biomedical databases automatically extracted from the scientific literature. BiDI provides an index of data resources and a path to access them seamlessly. Methods We designed an ensemble of deep learning methods to extract database mentions. To train the system, we annotated a set of 1242 articles that included mentions of database publications. Such a data set was used along with transfer learning techniques to train an ensemble of deep learning natural language processing models targeted at database publication detection. Results The system obtained an F1 score of 0.929 on database detection, showing high precision and recall values. When applying this model to the PubMed and PubMed Central databases, we identified over 10,000 unique databases. The ensemble model also extracted the weblinks to the reported databases and discarded irrelevant links. For the extraction of weblinks, the model achieved a cross-validated F1 score of 0.908. We show two use cases: one related to “omics” and the other related to the COVID-19 pandemic. Conclusions BiDI enables access to biomedical resources over the internet and facilitates data-driven research and other scientific initiatives. The repository is openly available online and will be regularly updated with an automatic text processing pipeline. The approach can be reused to create repositories of different types (ie, biomedical and others).
Styles APA, Harvard, Vancouver, ISO, etc.
44

Jirakrit, Leelarungrayub, Yankai Araya et Thipcharoen Supattanawaree. « Knowledge Discovery on Artificial Intelligence and Physical Therapy : Document Mining Analysis ». IgMin Research 2, no 11 (21 novembre 2024) : 929–37. http://dx.doi.org/10.61927/igmin270.

Texte intégral
Résumé :
Artificial intelligence (AI) is the simulation of human intelligence and benchmarks in Physical Therapy (PT). Therefore, the updated knowledge derived from large databases is highly engaging. Data Mining (DM) analysis from a big database related to “AI” and “PT” was the aim for the co-occurrence of words, network clusters, and trends under the Knowledge Discovery in Databases (KDD). The terms “AI” and “PT” were cited from a big database in the SCOPUS. The co-occurrence, network clustering, and trend were computer-analyzed with a Bibliometric tool. Between 1993 and 2024, 174 documents were published, revealing the most frequently used terms related to AI, human, PT, physical modalities, machine learning, physical treatment, deep learning, patient rehabilitation, robotics, virtual reality, algorithms, telerehabilitation, ergonomics, exercise, quality of life, and other related topics. Five network clusters were discovered as; (1) AI, decision support systems, health care, human-computer interaction, intelligent robots, learning algorithms, neuromuscular, stroke, and patient rehabilitation, PT, robotics, etc., respectively, (2) aged, algorithms, biomechanics, exercise, exercise therapy, female, humans, machine learning, middle-aged, PT modalities, rehabilitation, treatment outcome, deep learning, etc., (3) deep learning, diagnosis, and quality of life, (4) review and systematic review, and (5) clinical practice. From 2008 to 2024, a trend emerged in the fields of algorithms, computer-assisted diagnosis, treatment planning, classification, equipment design, signal processing, AI, exercise, physical and patient rehabilitation, robotics, virtual reality, machine learning, deep learning, clinical practice, etc. Discovered knowledge of AI with PT related to different machine learning for use in clinical practice.
Styles APA, Harvard, Vancouver, ISO, etc.
45

Ibrahim, Haneen Siraj, Narjis Mezaal Shati et AbdulRahman A. Alsewari. « A Transfer Learning Approach for Arabic Image Captions ». Al-Mustansiriyah Journal of Science 35, no 3 (30 septembre 2024) : 81–90. http://dx.doi.org/10.23851/mjs.v35i3.1485.

Texte intégral
Résumé :
Background: Arabic image captioning (AIC) is the automatic generation of text descriptions in the Arabic language for images. Applies a transfer learning approach in deep learning to enhance computer vision and natural language processing. There are many datasets in English reverse other languages. Instead of, the Arabs researchers unanimously agreed that there is a lack of Arabic databases available in this field. Objective: This paper presents the improvement and processing of the available Arabic textual database using Google spreadsheets for translation and creation of AR. Flicker8k2023 dataset is an extension of the Arabic Flicker8k dataset available, it was uploaded to GitHub and made public for researches. Methods: An efficient model proposed using deep learning techniques by including two pre-training models (VGG16 and VGG19), to extract features from the images and build (LSTM and GRU) models to process textual prediction sequence. In addition to the effect of pre-processing the text in Arabic. Results: The adopted model outperforms better compared to the previous study in BLEU-1 from 33 to 40. Conclusions: This paper concluded that the biggest problem is the database available in the Arabic language. This paper has worked to increase the size of the text database from 24,276 to 32,364 thousand captions, where each image contains 4 captions.
Styles APA, Harvard, Vancouver, ISO, etc.
46

Zhang, Lan, Yu Feng Nie et Zhen Hai Wang. « Image De-Noising Using Deep Learning ». Applied Mechanics and Materials 641-642 (septembre 2014) : 1287–90. http://dx.doi.org/10.4028/www.scientific.net/amm.641-642.1287.

Texte intégral
Résumé :
Deep neural network as a part of deep learning algorithm is a state-of-the-art approach to find higher level representations of input data which has been introduced to many practical and challenging learning problems successfully. The primary goal of deep learning is to use large data to help solving a given task on machine learning. We propose an methodology for image de-noising project defined by this model and conduct training a large image database to get the experimental output. The result shows the robustness and efficient our our algorithm.
Styles APA, Harvard, Vancouver, ISO, etc.
47

Syed Qamrul Kazmi, Et al. « Image Retrieval Using Auto Encoding Features In Deep Learning ». International Journal on Recent and Innovation Trends in Computing and Communication 11, no 10 (2 novembre 2023) : 155–71. http://dx.doi.org/10.17762/ijritcc.v11i10.8478.

Texte intégral
Résumé :
The latest technologies and growth in availability of image storage in day to day life has made a vast storage place for the images in the database. Several devices which help in capturing the image contribute to a huge repository of images. Keeping in mind the daily input in the database, one must think of retrieving those images according to certain criteria mentioned. Several techniques such as shape of the object, Discrete Wavelet transform (DWT), texture features etc. were used in determining the type of image and classifying them. Segmentation also plays a vital role in image retrieval but the robustness is lacking in most of the cases. The process of retrieval mainly depends on the special characteristics possessed by an image rather than the whole image. Two types of image retrieval can be seen. One with a general object and the other which may be specific to some type of application. Modern deep neural networks for unsupervised feature learning like Deep Autoencoder (AE) learn embedded representations by stacking layers on top of each other. These learnt embedded-representations, however, may degrade as the AE network deepens due to vanishing gradient, resulting in decreased performance. We have introduced here the ResNet Autoencoder (RAE) and its convolutional version (C-RAE) for unsupervised feature based learning. The proposed model is tested on three distinct databases Corel1K, Cifar-10, Cifar-100 which differ in size. The presented algorithm have significantly reduced computation time and provided very high image retrieval levels of accuracy.
Styles APA, Harvard, Vancouver, ISO, etc.
48

Zhu, Ge et Liu. « Deep Learning-Based Classification of Weld Surface Defects ». Applied Sciences 9, no 16 (12 août 2019) : 3312. http://dx.doi.org/10.3390/app9163312.

Texte intégral
Résumé :
In order to realize the non-destructive intelligent identification of weld surface defects, an intelligent recognition method based on deep learning is proposed, which is mainly formed by convolutional neural network (CNN) and forest random. First, the high-level features are automatically learned through the CNN. Random forest is trained with extracted high-level features to predict the classification results. Secondly, the weld surface defects images are collected and preprocessed by image enhancement and threshold segmentation. A database of weld surface defects is established using pre-processed images. Finally, comparative experiments are performed on the weld surface defects database. The results show that the accuracy of the method combined with CNN and random forest can reach 0.9875, and it also demonstrates the method is effective and practical.
Styles APA, Harvard, Vancouver, ISO, etc.
49

Barreto, Fabian, Jignesh Sarvaiya et Suprava Patnaik. « Learning Representations for Face Recognition : A Review from Holistic to Deep Learning ». Advances in Technology Innovation 7, no 4 (5 août 2022) : 279–94. http://dx.doi.org/10.46604/aiti.2022.8308.

Texte intégral
Résumé :
For decades, researchers have investigated how to recognize facial images. This study reviews the development of different face recognition (FR) methods, namely, holistic learning, handcrafted local feature learning, shallow learning, and deep learning (DL). With the development of methods, the accuracy of recognizing faces in the labeled faces in the wild (LFW) database has been increased. The accuracy of holistic learning is 60%, that of handcrafted local feature learning increases to 70%, and that of shallow learning is 86%. Finally, DL achieves human-level performance (97% accuracy). This enhanced accuracy is caused by large datasets and graphics processing units (GPUs) with massively parallel processing capabilities. Furthermore, FR challenges and current research studies are discussed to understand future research directions. The results of this study show that presently the database of labeled faces in the wild has reached 99.85% accuracy.
Styles APA, Harvard, Vancouver, ISO, etc.
50

Zhou, Xiao Qing, et Xiao Ping Tang. « A Kind of Web Database Classification Based on Machine Learning ». Applied Mechanics and Materials 50-51 (février 2011) : 644–48. http://dx.doi.org/10.4028/www.scientific.net/amm.50-51.644.

Texte intégral
Résumé :
The traditional search engine is unable to correct search for the magnanimous information in Deep Web hides. The Web database's classification is the key step which integrates with the Web database classification and retrieves. This article has proposed one kind of classification based on machine learning's web database. The experiment has indicated that after this taxonomic approach undergoes few sample training, it can achieve the very good classified effect, and along with training sample's increase, this classifier's performance maintains stable and the rate of accuracy and the recalling rate fluctuate in the very small scope.
Styles APA, Harvard, Vancouver, ISO, etc.
Nous offrons des réductions sur tous les plans premium pour les auteurs dont les œuvres sont incluses dans des sélections littéraires thématiques. Contactez-nous pour obtenir un code promo unique!

Vers la bibliographie