Journal articles on the topic 'Neural networks'

To see the other types of publications on this topic, follow the link: Neural networks.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 journal articles for your research on the topic 'Neural networks.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.

1

Navghare, Tukaram, Aniket Muley, and Vinayak Jadhav. "Siamese Neural Networks for Kinship Prediction: A Deep Convolutional Neural Network Approach." Indian Journal Of Science And Technology 17, no. 4 (January 26, 2024): 352–58. http://dx.doi.org/10.17485/ijst/v17i4.3018.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

N, Vikram. "Artificial Neural Networks." International Journal of Research Publication and Reviews 4, no. 4 (April 23, 2023): 4308–9. http://dx.doi.org/10.55248/gengpi.4.423.37858.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

O. H. Abdelwahed, O. H. Abdelwahed, and M. El-Sayed Wahed. "Optimizing Single Layer Cellular Neural Network Simulator using Simulated Annealing Technique with Neural Networks." Indian Journal of Applied Research 3, no. 6 (October 1, 2011): 91–94. http://dx.doi.org/10.15373/2249555x/june2013/31.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Zengguo Sun, Zengguo Sun, Guodong Zhao Zengguo Sun, Rafał Scherer Guodong Zhao, Wei Wei Rafał Scherer, and Marcin Woźniak Wei Wei. "Overview of Capsule Neural Networks." 網際網路技術學刊 23, no. 1 (January 2022): 033–44. http://dx.doi.org/10.53106/160792642022012301004.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
<p>As a vector transmission network structure, the capsule neural network has been one of the research hotspots in deep learning since it was proposed in 2017. In this paper, the latest research progress of capsule networks is analyzed and summarized. Firstly, we summarize the shortcomings of convolutional neural networks and introduce the basic concept of capsule network. Secondly, we analyze and summarize the improvements in the dynamic routing mechanism and network structure of the capsule network in recent years and the combination of the capsule network with other network structures. Finally, we compile the applications of capsule network in many fields, including computer vision, natural language, and speech processing. Our purpose in writing this article is to provide methods and means that can be used for reference in the research and practical applications of capsule networks.</p> <p>&nbsp;</p>
5

Perfetti, R. "A neural network to design neural networks." IEEE Transactions on Circuits and Systems 38, no. 9 (1991): 1099–103. http://dx.doi.org/10.1109/31.83884.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

AVeselý. "Neural networks in data mining." Agricultural Economics (Zemědělská ekonomika) 49, No. 9 (March 2, 2012): 427–31. http://dx.doi.org/10.17221/5427-agricecon.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
To posses relevant information is an inevitable condition for successful enterprising in modern business. Information could be parted to data and knowledge. How to gather, store and retrieve data is studied in database theory. In the knowledge engineering, there is in the centre of interest the knowledge and methods of its formalization and gaining are studied. Knowledge could be gained from experts, specialists in the area of interest, or it can be gained by induction from sets of data. Automatic induction of knowledge from data sets, usually stored in large databases, is called data mining. Classical methods of gaining knowledge from data sets are statistical methods. In data mining, new methods besides statistical are used. These new methods have their origin in artificial intelligence. They look for unknown and unexpected relations, which can be uncovered by exploring of data in database. In the article, a utilization of modern methods of data mining is described and especially the methods based on neural networks theory are pursued. The advantages and drawbacks of applications of multiplayer feed forward neural networks and Kohonen&rsquo;s self-organizing maps are discussed. Kohonen&rsquo;s self-organizing map is the most promising neural data-mining algorithm regarding its capability to visualize high-dimensional data.
7

J, Joselin, Dinesh T, and Ashiq M. "A Review on Neural Networks." International Journal of Trend in Scientific Research and Development Volume-2, Issue-6 (October 31, 2018): 565–69. http://dx.doi.org/10.31142/ijtsrd18461.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Ziroyan, M. A., E. A. Tusova, A. S. Hovakimian, and S. G. Sargsyan. "Neural networks apparatus in biometrics." Contemporary problems of social work 1, no. 2 (June 30, 2015): 129–37. http://dx.doi.org/10.17922/2412-5466-2015-1-2-129-137.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Marton, Sascha, Stefan Lüdtke, and Christian Bartelt. "Explanations for Neural Networks by Neural Networks." Applied Sciences 12, no. 3 (January 18, 2022): 980. http://dx.doi.org/10.3390/app12030980.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Understanding the function learned by a neural network is crucial in many domains, e.g., to detect a model’s adaption to concept drift in online learning. Existing global surrogate model approaches generate explanations by maximizing the fidelity between the neural network and a surrogate model on a sample-basis, which can be very time-consuming. Therefore, these approaches are not applicable in scenarios where timely or frequent explanations are required. In this paper, we introduce a real-time approach for generating a symbolic representation of the function learned by a neural network. Our idea is to generate explanations via another neural network (called the Interpretation Network, or I-Net), which maps network parameters to a symbolic representation of the network function. We show that the training of an I-Net for a family of functions can be performed up-front and subsequent generation of an explanation only requires querying the I-Net once, which is computationally very efficient and does not require training data. We empirically evaluate our approach for the case of low-order polynomials as explanations, and show that it achieves competitive results for various data and function complexities. To the best of our knowledge, this is the first approach that attempts to learn mapping from neural networks to symbolic representations.
10

Wang, Jun. "Artificial neural networks versus natural neural networks." Decision Support Systems 11, no. 5 (June 1994): 415–29. http://dx.doi.org/10.1016/0167-9236(94)90016-7.

Full text
APA, Harvard, Vancouver, ISO, and other styles
11

Rodriguez, Nathaniel, Eduardo Izquierdo, and Yong-Yeol Ahn. "Optimal modularity and memory capacity of neural reservoirs." Network Neuroscience 3, no. 2 (January 2019): 551–66. http://dx.doi.org/10.1162/netn_a_00082.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
The neural network is a powerful computing framework that has been exploited by biological evolution and by humans for solving diverse problems. Although the computational capabilities of neural networks are determined by their structure, the current understanding of the relationships between a neural network’s architecture and function is still primitive. Here we reveal that a neural network’s modular architecture plays a vital role in determining the neural dynamics and memory performance of the network of threshold neurons. In particular, we demonstrate that there exists an optimal modularity for memory performance, where a balance between local cohesion and global connectivity is established, allowing optimally modular networks to remember longer. Our results suggest that insights from dynamical analysis of neural networks and information-spreading processes can be leveraged to better design neural networks and may shed light on the brain’s modular organization.
12

Yashchenko, V. O. "Neural-like growing networks in the development of general intelligence. Neural-like growing networks (P. II)." Mathematical machines and systems 1 (2023): 3–29. http://dx.doi.org/10.34121/1028-9763-2023-1-3-29.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
This article is devoted to the development of general artificial intelligence (AGI) based on a new type of neural networks – “neural-like growing networks”. It consists of two parts. The first one was published in N4, 2022, and describes an artificial neural-like element (artificial neuron) in terms of its functionality, which is as close as possible to a biological neuron. An artificial neural-like element is the main element in building neural-like growing networks. The second part deals with the structures and functions of artificial and natural neural networks. The paper proposes a new approach for creating neural-like growing networks as a means of developing AGI that is as close as possible to the natural intelligence of a person. The intelligence of man and living organisms is formed by their nervous system. According to I.P. Pavlov's definition, the main mechanism of higher nervous activity is the reflex activity of the nervous system. In the nerve cell, the main storage of unconditioned reflexes is the deoxyribonucleic acid (DNA) molecule. The article describes ribosomal protein synthesis that contributes to the implementation of unconditioned reflexes and the formation of conditioned reflexes as the basis for learning biological objects. The first part of the work shows that the structure and functions of ribosomes almost completely coincide with the structure and functions of the Turing machine. Turing invented this machine to prove the fundamental (theoretical) possibility of constructing arbitrarily complex algorithms from extremely simple operations, and the operations themselves are performed automatically. Here arises a stunning analogy, nature created DNA and the ribosome to build complex algorithms for creating biological objects and their communication with each other and with the external environment, and the ribosomal protein synthesis is carried out by many ribosomes at the same time. It was concluded that the nerve cells of the brain are analog multi-machine complexes – ultra-fast molecular supercomputers with an unusually simple analog programming device.
13

JORGENSEN, THOMAS D., BARRY P. HAYNES, and CHARLOTTE C. F. NORLUND. "PRUNING ARTIFICIAL NEURAL NETWORKS USING NEURAL COMPLEXITY MEASURES." International Journal of Neural Systems 18, no. 05 (October 2008): 389–403. http://dx.doi.org/10.1142/s012906570800166x.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
This paper describes a new method for pruning artificial neural networks, using a measure of the neural complexity of the neural network. This measure is used to determine the connections that should be pruned. The measure computes the information-theoretic complexity of a neural network, which is similar to, yet different from previous research on pruning. The method proposed here shows how overly large and complex networks can be reduced in size, whilst retaining learnt behaviour and fitness. The technique proposed here helps to discover a network topology that matches the complexity of the problem it is meant to solve. This novel pruning technique is tested in a robot control domain, simulating a racecar. It is shown, that the proposed pruning method is a significant improvement over the most commonly used pruning method Magnitude Based Pruning. Furthermore, some of the pruned networks prove to be faster learners than the benchmark network that they originate from. This means that this pruning method can also help to unleash hidden potential in a network, because the learning time decreases substantially for a pruned a network, due to the reduction of dimensionality of the network.
14

Parks, Allen D. "Characterizing Computation in Artificial Neural Networks by their Diclique Covers and Forman-Ricci Curvatures." European Journal of Engineering Research and Science 5, no. 2 (February 13, 2020): 171–77. http://dx.doi.org/10.24018/ejers.2020.5.2.1689.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
The relationships between the structural topology of artificial neural networks, their computational flow, and their performance is not well understood. Consequently, a unifying mathematical framework that describes computational performance in terms of their underlying structure does not exist. This paper makes a modest contribution to understanding the structure-computational flow relationship in artificial neural networks from the perspective of the dicliques that cover the structure of an artificial neural network and the Forman-Ricci curvature of an artificial neural network’s connections. Special diclique cover digraph representations of artificial neural networks useful for network analysis are introduced and it is shown that such covers generate semigroups that provide algebraic representations of neural network connectivity.
15

Parks, Allen D. "Characterizing Computation in Artificial Neural Networks by their Diclique Covers and Forman-Ricci Curvatures." European Journal of Engineering and Technology Research 5, no. 2 (February 13, 2020): 171–77. http://dx.doi.org/10.24018/ejeng.2020.5.2.1689.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
The relationships between the structural topology of artificial neural networks, their computational flow, and their performance is not well understood. Consequently, a unifying mathematical framework that describes computational performance in terms of their underlying structure does not exist. This paper makes a modest contribution to understanding the structure-computational flow relationship in artificial neural networks from the perspective of the dicliques that cover the structure of an artificial neural network and the Forman-Ricci curvature of an artificial neural network’s connections. Special diclique cover digraph representations of artificial neural networks useful for network analysis are introduced and it is shown that such covers generate semigroups that provide algebraic representations of neural network connectivity.
16

Li, Xiao Hu, Feng Xu, Jin Hua Zhang, and Su Nan Wang. "A New Small-World Neural Network with its Performance on Fault Tolerance." Advanced Materials Research 629 (December 2012): 719–24. http://dx.doi.org/10.4028/www.scientific.net/amr.629.719.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Many artificial neural networks are the simple simulation of brain neural network’s architecture and function. However, how to rebuild new artificial neural network which architecture is similar to biological neural networks is worth studying. In this study, a new multilayer feedforward small-world neural network is presented using the results form research on complex network. Firstly, a new multilayer feedforward small-world neural network which relies on the rewiring probability heavily is built up on the basis of the construction ideology of Watts-Strogatz networks model and community structure. Secondly, fault tolerance is employed in investigating the performances of new small-world neural network. When the network with connection fault or neuron damage is used to test the fault tolerance performance under different rewiring probability, simulation results show that the fault tolerance capability of small-world neural network outmatches that of the same scale regular network when the fault probability is more than 40%, while random network has the best fault tolerance capability.
17

Tetko, Igor V. "Neural Network Studies. 4. Introduction to Associative Neural Networks." Journal of Chemical Information and Computer Sciences 42, no. 3 (March 26, 2002): 717–28. http://dx.doi.org/10.1021/ci010379o.

Full text
APA, Harvard, Vancouver, ISO, and other styles
18

Garcia, R. K., K. Moreira Gandra, J. M. Block, and D. Barrera-Arellano. "Neural networks to formulate special fats." Grasas y Aceites 63, no. 3 (July 5, 2012): 245–52. http://dx.doi.org/10.3989/gya.119011.

Full text
APA, Harvard, Vancouver, ISO, and other styles
19

Veselý, A., and D. Brechlerová. "Neural networks in intrusion detection systems." Agricultural Economics (Zemědělská ekonomika) 50, No. 1 (February 24, 2012): 35–40. http://dx.doi.org/10.17221/5164-agricecon.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Security of an information system is its very important property, especially today, when computers are interconnected via internet. Because no system can be absolutely secure, the timely and accurate detection of intrusions is necessary. For this purpose, Intrusion Detection Systems (IDS) were designed. There are two basic models of IDS: misuse IDS and anomaly IDS. Misuse systems detect intrusions by looking for activity that corresponds to the known signatures of intrusions or vulnerabilities. Anomaly systems detect intrusions by searching for an abnormal system activity. Most IDS commercial tools are misuse systems with rule-based expert system structure. However, these techniques are less successful when attack characteristics vary from built-in signatures. Artificial neural networks offer the potential to resolve these problems. As far as anomaly systems are concerned, it is very difficult to build them, because it is difficult to define the normal and abnormal behaviour of a&nbsp;system. Also for building anomaly system, neural networks can be used, because they can learn to discriminate the normal and abnormal behaviour of a&nbsp;system from examples. Therefore, they offer a&nbsp;promising technique for building anomaly systems. This paper presents an overview of the applicability of neural networks in building intrusion systems and discusses advantages and drawbacks of neural network technology.
20

Mu, Yangzi, Mengxing Huang, Chunyang Ye, and Qingzhou Wu. "Diagnosis Prediction via Recurrent Neural Networks." International Journal of Machine Learning and Computing 8, no. 2 (April 2018): 117–20. http://dx.doi.org/10.18178/ijmlc.2018.8.2.673.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Sandhiya, K., M. Vidhya, and M. Shivaranjani S. Saranya. "Smart Fruit Classification using Neural Networks." International Journal of Trend in Scientific Research and Development Volume-2, Issue-1 (December 31, 2017): 1298–303. http://dx.doi.org/10.31142/ijtsrd6986.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

S, Pothumani, and Priya N. "Analysis of RAID in Neural Networks." Journal of Advanced Research in Dynamical and Control Systems 11, no. 0009-SPECIAL ISSUE (September 25, 2019): 589–94. http://dx.doi.org/10.5373/jardcs/v11/20192609.

Full text
APA, Harvard, Vancouver, ISO, and other styles
23

TYMOSHENKO, Pavlo, Yevgen ZASOBA, Olexander KOVALCHUK, and Olexander PSHENYCHNYY. "NEUROEVOLUTIONARY ALGORITHMS FOR NEURAL NETWORKS GENERATING." Herald of Khmelnytskyi National University. Technical sciences 315, no. 6(1) (December 29, 2022): 240–44. http://dx.doi.org/10.31891/2307-5732-2022-315-6-240-244.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Solving engineering problems using conventional neural networks requires long-term research on the choice of architecture and hyperparameters. A strong artificial intelligence would be devoid of such shortcomings. Such research is carried out using a very wide range of approaches: for example, biological (attempts to grow a brain in laboratory conditions), hardware (creating neural processors) or software (using the power of ordinary CPUs and GPUs). The goal of the work is to develop such a system that would allow using evolutionary approaches to generate neural networks suitable for solving problems. This is called “neuroevolution”. The purpose of this work also includes the study of the features of possible applicable evolutionary strategies. The object of research in this work is a neuroevolutionary approach to solving problems of machine learning. The subject of research is evolutionary strategies, neural coding methods networks in the organism’s genome. The scientific novelty of the work lies in the testing of previously unused evolutionary strategies and the generalization of the obtained system to the systems of “general artificial intelligence”. A system for simulating neuroevolution was created. The specifics of implementation were considered, the choice of algorithms was justified, and their work was explained. In order to perform experiments, datasets were created and methods of applying neuroevolutionary systems were developed. It was possible to choose the most optimal training parameters, to find out the relationship between them, as well as the accuracy and speed of training. It cannot be said that the models implemented within this work directly bring us closer to strong AI. They still lack their own memory as well as a certain level of complexity. For successful use, it is necessary to configure the view of the input data or perform some calculations outside the model. However, in the future, such a system can be developed, for example, to work with SNNs, or for use on special equipment
24

Mahat, Norpah, Nor Idayunie Nording, Jasmani Bidin, Suzanawati Abu Hasan, and Teoh Yeong Kin. "Artificial Neural Network (ANN) to Predict Mathematics Students’ Performance." Journal of Computing Research and Innovation 7, no. 1 (March 30, 2022): 29–38. http://dx.doi.org/10.24191/jcrinn.v7i1.264.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Predicting students’ academic performance is very essential to produce high-quality students. The main goal is to continuously help students to increase their ability in the learning process and to help educators as well in improving their teaching skills. Therefore, this study was conducted to predict mathematics students’ performance using Artificial Neural Network (ANN). The secondary data from 382 mathematics students from UCI Machine Learning Repository Data Sets used to train the neural networks. The neural network model built using nntool. Two inputs are used which are the first and the second period grade while one target output is used which is the final grade. This study also aims to identify which training function is the best among three Feed-Forward Neural Networks known as Network1, Network2 and Network3. Three types of training functions have been selected in this study, which are Levenberg-Marquardt (TRAINLM), Gradient descent with momentum (TRAINGDM) and Gradient descent with adaptive learning rate (TRAINGDA). Each training function will be compared based on Performance value, correlation coefficient, gradient and epoch. MATLAB R2020a was used for data processing. The results show that the TRAINLM function is the most suitable function in predicting mathematics students’ performance because it has a higher correlation coefficient and a lower Performance value.
25

Boonsatit, Nattakan, Santhakumari Rajendran, Chee Peng Lim, Anuwat Jirawattanapanit, and Praneesh Mohandas. "New Adaptive Finite-Time Cluster Synchronization of Neutral-Type Complex-Valued Coupled Neural Networks with Mixed Time Delays." Fractal and Fractional 6, no. 9 (September 13, 2022): 515. http://dx.doi.org/10.3390/fractalfract6090515.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
The issue of adaptive finite-time cluster synchronization corresponding to neutral-type coupled complex-valued neural networks with mixed delays is examined in this research. A neutral-type coupled complex-valued neural network with mixed delays is more general than that of a traditional neural network, since it considers distributed delays, state delays and coupling delays. In this research, a new adaptive control technique is developed to synchronize neutral-type coupled complex-valued neural networks with mixed delays in finite time. To stabilize the resulting closed-loop system, the Lyapunov stability argument is leveraged to infer the necessary requirements on the control factors. The effectiveness of the proposed method is illustrated through simulation studies.
26

Henzinger, Thomas A., Mathias Lechner, and Đorđe Žikelić. "Scalable Verification of Quantized Neural Networks." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 5 (May 18, 2021): 3787–95. http://dx.doi.org/10.1609/aaai.v35i5.16496.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Formal verification of neural networks is an active topic of research, and recent advances have significantly increased the size of the networks that verification tools can handle. However, most methods are designed for verification of an idealized model of the actual network which works over real arithmetic and ignores rounding imprecisions. This idealization is in stark contrast to network quantization, which is a technique that trades numerical precision for computational efficiency and is, therefore, often applied in practice. Neglecting rounding errors of such low-bit quantized neural networks has been shown to lead to wrong conclusions about the network's correctness. Thus, the desired approach for verifying quantized neural networks would be one that takes these rounding errors into account. In this paper, we show that verifying the bit-exact implementation of quantized neural networks with bit-vector specifications is PSPACE-hard, even though verifying idealized real-valued networks and satisfiability of bit-vector specifications alone are each in NP. Furthermore, we explore several practical heuristics toward closing the complexity gap between idealized and bit-exact verification. In particular, we propose three techniques for making SMT-based verification of quantized neural networks more scalable. Our experiments demonstrate that our proposed methods allow a speedup of up to three orders of magnitude over existing approaches.
27

Zhang, Wei, Zhi Han, Xiai Chen, Baichen Liu, Huidi Jia, and Yandong Tang. "Fully Kernected Neural Networks." Journal of Mathematics 2023 (June 28, 2023): 1–9. http://dx.doi.org/10.1155/2023/1539436.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
In this paper, we apply kernel methods to deep convolutional neural network (DCNN) to improve its nonlinear ability. DCNNs have achieved significant improvement in many computer vision tasks. For an image classification task, the accuracy comes to saturation when the depth and width of network are enough and appropriate. The saturation accuracy will not rise even by increasing the depth and width. We find that improving nonlinear ability of DCNNs can break through the saturation accuracy. In a DCNN, the former layer is more inclined to extract features and the latter layer is more inclined to classify features. Therefore, we apply kernel methods at the last fully connected layer to implicitly map features to a higher-dimensional space to improve nonlinear ability so that the network achieves better linear separability. Also, we name the network as fully kernected neural networks (fully connected neural networks with kernel methods). Our experiment result shows that fully kernected neural networks achieve higher classification accuracy and faster convergence rate than baseline networks.
28

Simons, Robert, and J. G. Taylor. "Neural Networks." Journal of the Operational Research Society 47, no. 4 (April 1996): 596. http://dx.doi.org/10.2307/3010740.

Full text
APA, Harvard, Vancouver, ISO, and other styles
29

Schier, R. "Neural networks." Radiology 191, no. 1 (April 1994): 291. http://dx.doi.org/10.1148/radiology.191.1.8134593.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Tafti, Mohammed H. A. "Neural networks." ACM SIGMIS Database: the DATABASE for Advances in Information Systems 23, no. 1 (March 1992): 51–54. http://dx.doi.org/10.1145/134347.134361.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Turega, M. A. "Neural Networks." Computer Journal 35, no. 3 (June 1, 1992): 290. http://dx.doi.org/10.1093/comjnl/35.3.290.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Jordan, Michael I., and Christopher M. Bishop. "Neural networks." ACM Computing Surveys 28, no. 1 (March 1996): 73–75. http://dx.doi.org/10.1145/234313.234348.

Full text
APA, Harvard, Vancouver, ISO, and other styles
33

Dory, Robert A. "Neural Networks." Computers in Physics 4, no. 3 (1990): 324. http://dx.doi.org/10.1063/1.4822918.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

Ganssle, Graham. "Neural networks." Leading Edge 37, no. 8 (August 2018): 616–19. http://dx.doi.org/10.1190/tle37080616.1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
We've all heard a proselytizing hyperbolist make the artificial-intelligence-is-going-to-steal-my-job speech. If you subscribe, look at the code in the notebook accompanying this tutorial at https://github.com/seg/tutorials-2018 . It demonstrates a small neural network. You'll find a simple system composed chiefly of multiply and add operations. That's really all that happens inside a neural network. Multiply and add. There's no magic here.
35

McGourty, Christine. "Neural networks." Nature 335, no. 6186 (September 1988): 103. http://dx.doi.org/10.1038/335103b0.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Simons, Robert. "Neural Networks." Journal of the Operational Research Society 47, no. 4 (April 1996): 596–97. http://dx.doi.org/10.1057/jors.1996.70.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Beatty, P. C. W. "Neural networks." Current Anaesthesia & Critical Care 9, no. 4 (August 1998): 168–73. http://dx.doi.org/10.1016/s0953-7112(98)80050-1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
38

Cutler, Adele. "Neural Networks." Technometrics 42, no. 4 (November 2000): 432. http://dx.doi.org/10.1080/00401706.2000.10485724.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

Signorini, DavidF, JimM Slattery, S. R. Dodds, Victor Lane, and Peter Littlejohns. "Neural networks." Lancet 346, no. 8988 (December 1995): 1500–1501. http://dx.doi.org/10.1016/s0140-6736(95)92525-2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
40

Jefferson, MilesF, Neil Pendleton, Sam Lucas, MichaelA Horan, and Lionel Tarassenko. "Neural networks." Lancet 346, no. 8991-8992 (December 1995): 1712. http://dx.doi.org/10.1016/s0140-6736(95)92880-4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
41

Gutfreund, H. "NEURAL NETWORKS." International Journal of Modern Physics B 04, no. 06 (May 1990): 1223–39. http://dx.doi.org/10.1142/s0217979290000607.

Full text
APA, Harvard, Vancouver, ISO, and other styles
42

Medoff, Deborah R., and M.-A. Tagamets. "Neural Networks." American Journal of Psychiatry 157, no. 10 (October 2000): 1571. http://dx.doi.org/10.1176/appi.ajp.157.10.1571.

Full text
APA, Harvard, Vancouver, ISO, and other styles
43

Lewis, David A. "Neural Networks." American Journal of Psychiatry 157, no. 11 (November 2000): 1752. http://dx.doi.org/10.1176/appi.ajp.157.11.1752.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Graybiel, Ann M. "Neural Networks." American Journal of Psychiatry 158, no. 1 (January 2001): 21. http://dx.doi.org/10.1176/appi.ajp.158.1.21.

Full text
APA, Harvard, Vancouver, ISO, and other styles
45

Tamminga, Carol A., and Henry H. Holcomb. "Neural Networks." American Journal of Psychiatry 158, no. 2 (February 2001): 185. http://dx.doi.org/10.1176/appi.ajp.158.2.185.

Full text
APA, Harvard, Vancouver, ISO, and other styles
46

Schwindling, Jerome. "Neural Networks." EPJ Web of Conferences 4 (2010): 02002. http://dx.doi.org/10.1051/epjconf/20100402002.

Full text
APA, Harvard, Vancouver, ISO, and other styles
47

Widrow, Bernard, David E. Rumelhart, and Michael A. Lehr. "Neural networks." Communications of the ACM 37, no. 3 (March 1994): 93–105. http://dx.doi.org/10.1145/175247.175257.

Full text
APA, Harvard, Vancouver, ISO, and other styles
48

Kock, Gerd. "Neural networks." Microprocessing and Microprogramming 38, no. 1-5 (September 1993): 679. http://dx.doi.org/10.1016/0165-6074(93)90210-c.

Full text
APA, Harvard, Vancouver, ISO, and other styles
49

Titterington, Michael. "Neural networks." WIREs Computational Statistics 2, no. 1 (December 21, 2009): 1–8. http://dx.doi.org/10.1002/wics.50.

Full text
APA, Harvard, Vancouver, ISO, and other styles
50

Xu, Zenglin. "Tensor Networks Meet Neural Networks." Journal of Physics: Conference Series 2278, no. 1 (May 1, 2022): 012003. http://dx.doi.org/10.1088/1742-6596/2278/1/012003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
Abstract:
Abstract As a simulation of the human cognitive system, deep neural networks have achieved great success in many machine learning tasks and are the main driving force of the current development of artificial intelligence. On the other hand, tensor networks as an approximation of quantum many-body systems in quantum physics are applied to quantum physics, statistical physics, quantum chemistry and machine learning. This talk will first give a brief introduction to neural networks and tensor networks, and then discuss the cross-field research between deep neural networks and tensor networks, such as network compression and knowledge fusion, including our recent work on tensor neural networks. Finally, this talk will also discuss the connection to quantum machine learning.

To the bibliography