Inhaltsverzeichnis
Auswahl der wissenschaftlichen Literatur zum Thema „Compression de réseaux neuronaux“
Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an
Machen Sie sich mit den Listen der aktuellen Artikel, Bücher, Dissertationen, Berichten und anderer wissenschaftlichen Quellen zum Thema "Compression de réseaux neuronaux" bekannt.
Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.
Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.
Zeitschriftenartikel zum Thema "Compression de réseaux neuronaux"
Wendling, Fabrice. „Modélisation des réseaux neuronaux épileptogènes“. Neurophysiologie Clinique 48, Nr. 4 (September 2018): 248. http://dx.doi.org/10.1016/j.neucli.2018.06.074.
Der volle Inhalt der QuelleMeunier, Claude. „La physique des réseaux neuronaux“. Intellectica. Revue de l'Association pour la Recherche Cognitive 9, Nr. 1 (1990): 313–21. http://dx.doi.org/10.3406/intel.1990.890.
Der volle Inhalt der QuelleVenance, Laurent. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 112 (01.04.2013): 884–86. http://dx.doi.org/10.4000/annuaire-cdf.1083.
Der volle Inhalt der QuelleVenance, Laurent. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 114 (01.07.2015): 1030–32. http://dx.doi.org/10.4000/annuaire-cdf.12073.
Der volle Inhalt der QuelleVenance, Laurent. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 115 (01.11.2016): 913–16. http://dx.doi.org/10.4000/annuaire-cdf.12639.
Der volle Inhalt der QuelleVenance, Laurent. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 111 (01.04.2012): 909–11. http://dx.doi.org/10.4000/annuaire-cdf.1706.
Der volle Inhalt der QuelleDeniau, Jean-Michel. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 108 (01.12.2008): 964–69. http://dx.doi.org/10.4000/annuaire-cdf.254.
Der volle Inhalt der QuelleVenance, Laurent. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 113 (01.04.2014): 947–49. http://dx.doi.org/10.4000/annuaire-cdf.2708.
Der volle Inhalt der QuelleDeniau, Jean-Michel, und Laurent Venance. „Dynamique et physiopathologie des réseaux neuronaux“. L’annuaire du Collège de France, Nr. 109 (01.03.2010): 1082–86. http://dx.doi.org/10.4000/annuaire-cdf.456.
Der volle Inhalt der QuelleSauteur, Tania. „Comment les cerveaux humains encodent-ils leurs propres processus d'apprentissage et de mémorisation et comment la topologie du réseau social élargi d'une personne présente-t-elle des schémas neuronaux similaires à ceux de ses ami-e-s et communautés ?“ Cortica 2, Nr. 2 (19.09.2023): 157–63. http://dx.doi.org/10.26034/cortica.2023.4208.
Der volle Inhalt der QuelleDissertationen zum Thema "Compression de réseaux neuronaux"
Foucher, Christophe. „Analyse et amélioration d'algorithmes neuronaux et non neuronaux de quantification vectorielle pour la compression d'images“. Rennes 1, 2002. http://www.theses.fr/2002REN10120.
Der volle Inhalt der QuelleDupont, Robin. „Deep Neural Network Compression for Visual Recognition“. Electronic Thesis or Diss., Sorbonne université, 2023. http://www.theses.fr/2023SORUS565.
Der volle Inhalt der QuelleThanks to the miniaturisation of electronics, embedded devices have become ubiquitous since the 2010s, performing various tasks around us. As their usage expands, there's an increasing demand for efficient data processing and decision-making. Deep neural networks are apt tools for this, but they are often too large and intricate for embedded systems. Therefore, methods to compress these networks without affecting their performance are crucial. This PhD thesis introduces two methods focused on pruning to compress networks, maintaining accuracy. The thesis first details a budget-aware method for compressing large neural networks using weight reparametrisation and a budget loss, eliminating the need for fine-tuning. Traditional pruning methods often use post-training indicators to cut weights, ignoring desired pruning rates. Our method incorporates a budget loss, directing pruning during training, enabling simultaneous topology and weight optimisation. By soft-pruning smaller weights via reparametrisation, we reduce accuracy loss compared to standard pruning. We validate our method on several datasets and architectures. Later, the thesis examines extracting efficient subnetworks without weight training. We aim to discern the optimal subnetwork topology within a large network, bypassing weight optimisation yet ensuring strong performance. This is realized with our Arbitrarily Shifted Log Parametrisation, a differentiable method for discrete topology sampling, facilitating masks' training to denote weight selection probability. Additionally, a weight recalibration technique, Smart Rescale, is presented. It boosts extracted subnetworks' performance and hastens their training. Our method identifies the best pruning rate in a single training cycle, averting exhaustive hyperparameter searches and various rate training. Through extensive tests, our technique consistently surpasses similar state-of-the-art methods, creating streamlined networks that achieve high sparsity without notable accuracy drops
Leconte, Louis. „Compression and federated learning : an approach to frugal machine learning“. Electronic Thesis or Diss., Sorbonne université, 2024. http://www.theses.fr/2024SORUS107.
Der volle Inhalt der Quelle“Intelligent” devices and tools are gradually becoming the standard, as the implementation of algorithms based on artificial neural networks is experiencing widespread development. Neural networks consist of non-linear machine learning models that manipulate high-dimensional objects and obtain state-of-the-art performances in various areas, such as image recognition, speech recognition, natural language processing, and recommendation systems.However, training a neural network on a device with lower computing capacity can be challenging, as it can imply cutting back on memory, computing time or power. A natural approach to simplify this training is to use quantized neural networks, whose parameters and operations use efficient low-bit primitives. However, optimizing a function over a discrete set in high dimension is complex, and can still be prohibitively expensive in terms of computational power. For this reason, many modern applications use a network of devices to store individual data and share the computational load. A new approach, federated learning, considers a distributed environment: Data is stored on devices and a centralized server orchestrates the training process across multiple devices.In this thesis, we investigate different aspects of (stochastic) optimization with the goal of reducing energy costs for potentially very heterogeneous devices. The first two contributions of this work are dedicated to the case of quantized neural networks. Our first idea is based on an annealing strategy: we formulate the discrete optimization problem as a constrained optimization problem (where the size of the constraint is reduced over iterations). We then focus on a heuristic for training binary deep neural networks. In this particular framework, the parameters of the neural networks can only have two values. The rest of the thesis is about efficient federated learning. Following our contributions developed for training quantized neural network, we integrate them into a federated environment. Then, we propose a novel unbiased compression technique that can be used in any gradient based distributed optimization framework. Our final contributions address the particular case of asynchronous federated learning, where devices have different computational speeds and/or access to bandwidth. We first propose a contribution that reweights the contributions of distributed devices. Then, in our final work, through a detailed queuing dynamics analysis, we propose a significant improvement to the complexity bounds provided in the literature onasynchronous federated learning.In summary, this thesis presents novel contributions to the field of quantized neural networks and federated learning by addressing critical challenges and providing innovative solutions for efficient and sustainable learning in a distributed and heterogeneous environment. Although the potential benefits are promising, especially in terms of energy savings, caution is needed as a rebound effect could occur
Yvinec, Edouard. „Efficient Neural Networks : Post Training Pruning and Quantization“. Electronic Thesis or Diss., Sorbonne université, 2023. http://www.theses.fr/2023SORUS581.
Der volle Inhalt der QuelleDeep neural networks have grown to be the most widely adopted models to solve most computer vision and natural language processing tasks. Since the renewed interest, sparked in 2012, for these architectures, in machine learning, their size in terms of memory footprint and computational costs have increased tremendously, which has hindered their deployment. In particular, with the rising interest for generative ai such as large language models and diffusion models, this phenomenon has recently reached new heights, as these models can weight several billions of parameters and require multiple high-end gpus in order to infer in real-time. In response, the deep learning community has researched for methods to compress and accelerate these models. These methods are: efficient architecture design, tensor decomposition, pruning and quantization. In this manuscript, I paint a landscape of the current state-of-the art in deep neural networks compression and acceleration as well as my contributions to the field. First, I propose a general introduction to the aforementioned techniques and highlight their shortcomings and current challenges. Second, I provide a detailed discussion regarding my contributions to the field of deep neural networks pruning. These contributions led to the publication of three articles: RED, RED++ and SInGE. In RED and RED++, I introduced a novel way to perform data-free pruning and tensor decomposition based on redundancy reduction. On the flip side, in SInGE, I proposed a new importance-based criterion for data-driven pruning. This criterion was inspired by attribution techniques which consist in ranking inputs by their relative importance with respect to the final prediction. In SInGE, I adapted one of the most effective attribution technique to weight importance ranking for pruning. In the third chapter, I layout my contributions to the field of deep quantization: SPIQ, PowerQuant, REx, NUPES, and a best practice paper. Each of these methods address one of the previous limitations of post-training quantization. In SPIQ, PowerQuant and REx, I provide a solution to the granularity limitations of quantization, a novel non-uniform format which is particularly effective on transformer architectures and a technique for quantization decomposition which eliminates the need for unsupported bit-widths, respectively. In the two remaining articles, I provide significant improvements over existing gradient-based post-training quantization techniques, bridging the gap between such techniques and non-uniform quantization. In the last chapter, I propose a set of leads for future work which I believe to be the, current, most important unanswered questions in the field
Louis, Thomas. „Conventionnel ou bio-inspiré ? Stratégies d'optimisation de l'efficacité énergétique des réseaux de neurones pour environnements à ressources limitées“. Electronic Thesis or Diss., Université Côte d'Azur, 2025. http://www.theses.fr/2025COAZ4001.
Der volle Inhalt der QuelleIntegrating artificial intelligence (AI) algorithms directly into satellites presents numerous challenges. These embedded systems, which are heavily limited in energy consumption and memory footprint, must also withstand interference. This systematically requires the use of system-on-chip (SoC) solutions to combine two so-called “heterogeneous” systems: a versatile microcontroller and an energy-efficient computing accelerator (such as an FPGA or ASIC). To address the challenges related to deploying such architectures, this thesis focuses on optimizing and deploying neural networks on heterogeneous embedded architectures, aiming to balance energy consumption and AI performance.In Chapter 2 of this thesis, an in-depth study of recent compression techniques for feedforward neural networks (FNN) like MLPs or CNNs was conducted. These techniques, which reduce the computational complexity and memory footprint of these models, are essential for deployment in resource-constrained environments. Spiking neural networks (SNN) were also explored. These bio-inspired networks can indeed offer greater energy efficiency compared to FNNs.In Chapter 3, we adapted and developed innovative quantization methods to reduce the number of bits used to represent the values in a spiking network. This allowed us to compare the quantization of SNNs and FNNs, to understand and assess their respective trade-offs in terms of losses and gains. Reducing the activity of an SNN (e.g., the number of spikes generated during inference) directly improves the energy efficiency of SNNs. To this end, in Chapter 4, we leveraged knowledge distillation and regularization techniques. These methods reduce the spiking activity of the network while preserving its accuracy, ensuring effective operation of SNNs on resource-limited hardware.In the final part of this thesis, we explored the hybridization of SNNs and FNNs. These hybrid networks (HNN) aim to further optimize energy efficiency while enhancing performance. We also proposed innovative multi-timestep networks, which process information with different latencies across layers within the same SNN. Experimental results show that this approach enables a reduction in overall energy consumption while maintaining performance across a range of tasks.This thesis serves as a foundation for deploying future neural network applications in space. To validate our methods, we provide a comparative analysis on various public datasets (CIFAR-10, CIFAR-100, MNIST, Google Speech Commands) as well as on a private dataset for cloud segmentation. Our approaches are evaluated based on metrics such as accuracy, energy consumption, or SNN activity. This research extends beyond aerospace applications. We have demonstrated the potential of quantized SNNs, hybrid neural networks, and multi-timestep networks for a variety of real-world scenarios where energy efficiency is critical. This work offers promising prospects for fields such as IoT devices, autonomous vehicles, and other systems requiring efficient AI deployment
Hubens, Nathan. „Towards lighter and faster deep neural networks with parameter pruning“. Electronic Thesis or Diss., Institut polytechnique de Paris, 2022. http://www.theses.fr/2022IPPAS025.
Der volle Inhalt der QuelleSince their resurgence in 2012, Deep Neural Networks have become ubiquitous in most disciplines of Artificial Intelligence, such as image recognition, speech processing, and Natural Language Processing. However, over the last few years, neural networks have grown exponentially deeper, involving more and more parameters. Nowadays, it is not unusual to encounter architectures involving several billions of parameters, while they mostly contained thousands less than ten years ago.This generalized increase in the number of parameters makes such large models compute-intensive and essentially energy inefficient. This makes deployed models costly to maintain but also their use in resource-constrained environments very challenging.For these reasons, much research has been conducted to provide techniques reducing the amount of storage and computing required by neural networks. Among those techniques, neural network pruning, consisting in creating sparsely connected models, has been recently at the forefront of research. However, although pruning is a prevalent compression technique, there is currently no standard way of implementing or evaluating novel pruning techniques, making the comparison with previous research challenging.Our first contribution thus concerns a novel description of pruning techniques, developed according to four axes, and allowing us to unequivocally and completely define currently existing pruning techniques. Those components are: the granularity, the context, the criteria, and the schedule. Defining the pruning problem according to those components allows us to subdivide the problem into four mostly independent subproblems and also to better determine potential research lines.Moreover, pruning methods are still in an early development stage, and primarily designed for the research community. Indeed, most pruning works are usually implemented in a self-contained and sophisticated way, making it troublesome for non-researchers to apply such techniques without having to learn all the intricacies of the field. To fill this gap, we proposed FasterAI toolbox, intended to be helpful to researchers, eager to create and experiment with different compression techniques, but also to newcomers, that desire to compress their neural network for concrete applications. In particular, the sparsification capabilities of FasterAI have been built according to the previously defined pruning components, allowing for a seamless mapping between research ideas and their implementation.We then propose four theoretical contributions, each one aiming at providing new insights and improving on state-of-the-art methods in each of the four identified description axes. Also, those contributions have been realized by using the previously developed toolbox, thus validating its scientific utility.Finally, to validate the applicative character of the pruning technique, we have selected a use case: the detection of facial manipulation, also called DeepFakes Detection. The goal is to demonstrate that the developed tool, as well as the different proposed scientific contributions, can be applicable to a complex and actual problem. This last contribution is accompanied by a proof-of-concept application, providing DeepFake detection capabilities in a web-based environment, thus allowing anyone to perform detection on an image or video of their choice.This Deep Learning era has emerged thanks to the considerable improvements in high-performance hardware and access to a large amount of data. However, since the decline of Moore's Law, experts are suggesting that we might observe a shift in how we conceptualize the hardware, by going from task-agnostic to domain-specialized computations, thus leading to a new era of collaboration between software, hardware, and machine learning communities. This new quest for more efficiency will thus undeniably go through neural network compression techniques, and particularly sparse computations
Resmerita, Diana. „Compression pour l'apprentissage en profondeur“. Thesis, Université Côte d'Azur, 2022. http://www.theses.fr/2022COAZ4043.
Der volle Inhalt der QuelleAutonomous cars are complex applications that need powerful hardware machines to be able to function properly. Tasks such as staying between the white lines, reading signs, or avoiding obstacles are solved by using convolutional neural networks (CNNs) to classify or detect objects. It is highly important that all the networks work in parallel in order to transmit all the necessary information and take a common decision. Nowadays, as the networks improve, they also have become bigger and more computational expensive. Deploying even one network becomes challenging. Compressing the networks can solve this issue. Therefore, the first objective of this thesis is to find deep compression methods in order to cope with the memory and computational power limitations present on embedded systems. The compression methods need to be adapted to a specific processor, Kalray's MPPA, for short term implementations. Our contributions mainly focus on compressing the network post-training for storage purposes, which means compressing the parameters of the network without retraining or changing the original architecture and the type of the computations. In the context of our work, we decided to focus on quantization. Our first contribution consists in comparing the performances of uniform quantization and non-uniform quantization, in order to identify which of the two has a better rate-distortion trade-off and could be quickly supported in the company. The company's interest is also directed towards finding new innovative methods for future MPPA generations. Therefore, our second contribution focuses on comparing standard floating-point representations (FP32, FP16) to recently proposed alternative arithmetical representations such as BFloat16, msfp8, Posit8. The results of this analysis were in favor for Posit8. This motivated the company Kalray to conceive a decompressor from FP16 to Posit8. Finally, since many compression methods already exist, we decided to move to an adjacent topic which aims to quantify theoretically the effects of quantization error on the network's accuracy. This is the second objective of the thesis. We notice that well-known distortion measures are not adapted to predict accuracy degradation in the case of inference for compressed neural networks. We define a new distortion measure with a closed form which looks like a signal-to-noise ratio. A set of experiments were done using simulated data and small networks, which show the potential of this distortion measure
Fernandez, Brillet Lucas. „Réseaux de neurones CNN pour la vision embarquée“. Thesis, Université Grenoble Alpes, 2020. http://www.theses.fr/2020GRALM043.
Der volle Inhalt der QuelleRecently, Convolutional Neural Networks have become the state-of-the-art soluion(SOA) to most computer vision problems. In order to achieve high accuracy rates, CNNs require a high parameter count, as well as a high number of operations. This greatly complicates the deployment of such solutions in embedded systems, which strive to reduce memory size. Indeed, while most embedded systems are typically in the range of a few KBytes of memory, CNN models from the SOA usually account for multiple MBytes, or even GBytes in model size. Throughout this thesis, multiple novel ideas allowing to ease this issue are proposed. This requires to jointly design the solution across three main axes: Application, Algorithm and Hardware.In this manuscript, the main levers allowing to tailor computational complexity of a generic CNN-based object detector are identified and studied. Since object detection requires scanning every possible location and scale across an image through a fixed-input CNN classifier, the number of operations quickly grows for high-resolution images. In order to perform object detection in an efficient way, the detection process is divided into two stages. The first stage involves a region proposal network which allows to trade-off recall for the number of operations required to perform the search, as well as the number of regions passed on to the next stage. Techniques such as bounding box regression also greatly help reduce the dimension of the search space. This in turn simplifies the second stage, since it allows to reduce the task’s complexity to the set of possible proposals. Therefore, parameter counts can greatly be reduced.Furthermore, CNNs also exhibit properties that confirm their over-dimensionment. This over-dimensionement is one of the key success factors of CNNs in practice, since it eases the optimization process by allowing a large set of equivalent solutions. However, this also greatly increases computational complexity, and therefore complicates deploying the inference stage of these algorithms on embedded systems. In order to ease this problem, we propose a CNN compression method which is based on Principal Component Analysis (PCA). PCA allows to find, for each layer of the network independently, a new representation of the set of learned filters by expressing them in a more appropriate PCA basis. This PCA basis is hierarchical, meaning that basis terms are ordered by importance, and by removing the least important basis terms, it is possible to optimally trade-off approximation error for parameter count. Through this method, it is possible to compress, for example, a ResNet-32 network by a factor of ×2 both in the number of parameters and operations with a loss of accuracy <2%. It is also shown that the proposed method is compatible with other SOA methods which exploit other CNN properties in order to reduce computational complexity, mainly pruning, winograd and quantization. Through this method, we have been able to reduce the size of a ResNet-110 from 6.88Mbytes to 370kbytes, i.e. a x19 memory gain with a 3.9 % accuracy loss.All this knowledge, is applied in order to achieve an efficient CNN-based solution for a consumer face detection scenario. The proposed solution consists of just 29.3kBytes model size. This is x65 smaller than other SOA CNN face detectors, while providing equal detection performance and lower number of operations. Our face detector is also compared to a more traditional Viola-Jones face detector, exhibiting approximately an order of magnitude faster computation, as well as the ability to scale to higher detection rates by slightly increasing computational complexity.Both networks are finally implemented in a custom embedded multiprocessor, verifying that theorical and measured gains from PCA are consistent. Furthermore, parallelizing the PCA compressed network over 8 PEs achieves a x11.68 speed-up with respect to the original network running on a single PE
Verma, Sagar. „Deep Neural Network Modeling of Electric Motors“. Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPAST088.
Der volle Inhalt der QuelleThis thesis deals with the application of neural networks in solving electrical motor problems. Chapter 2 contributes to identifying a neural network that can learn the multivariate relationship between different electrical motor signals.The identified network is then used for speed-torque estimation from currents and voltages. Chapter 3 focuses on detecting and recovering from faulty measurements. Our method encompasses electrical sensor faults, mechanical faults, and temperature estimation.Chapter 4 then discusses the reliability of the speed-torque estimator in case of noisy currents and voltages. We presenta denoising method which allows our speed- torque estimator to be applicable in a realistic context. This is followed by an analysis of the adversarial robustness of the neural networks used in electrical motor tasks. The generalization capability of the speed-torque estimator is also briefly considered. In Chapter 5, we focus on the final roadblock in achieving real-world application of neural networks: computational requirements. We present the Subdifferential Inclusion for Sparsity (SIS) method to find the best sparse network from pretrained weights while maintaining original accuracy
Metz, Clément. „Codages optimisés pour la conception d'accélérateurs matériels de réseaux de neurones profonds“. Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPAST190.
Der volle Inhalt der QuelleNeural networks are an important component of machine learning tools because of their wide range of applications (health, energy, defence, finance, autonomous navigation, etc.). The performance of neural networks is greatly influenced by the complexity of their architecture in terms of the number of layers, neurons and connections. But the training and inference of ever-larger networks translates to greater demands on hardware resources and longer computing times. Conversely, their portability is limited on embedded systems with low memory and/or computing capacity.The aim of this thesis is to study and design methods for reducing the hardware footprint of neural networks while preserving their performance as much as possible. We restrict ourselves to convolution networks dedicated to computer vision by studying the possibilities offered by quantization. Quantization aims to reduce the hardware footprint, in terms of memory, bandwidth and computation operators, by reducing the number of bits in the network parameters and activations.The contributions of this thesis consist of a new post-training quantization method based on the exploitation of spatial correlations of network parameters, an approach facilitating the learning of very highly quantized networks, and a method aiming to combine mixed precision quantization and lossless entropy coding.The contents of this thesis are essentially limited to algorithmic aspects, but the research orientations were strongly influenced by the requirement for hardware feasibility of our solutions
Bücher zum Thema "Compression de réseaux neuronaux"
Kamp, Yves. Réseaux de neurones récursifs pour mémoires associatives. Lausanne: Presses polytechniques et universitaires romandes, 1990.
Den vollen Inhalt der Quelle findenMaren, Alianna. Handbook of Neural Computing Applications. San Diego: Academic Press, 1991.
Den vollen Inhalt der Quelle findenArtificial Neural Networks in Engineering Conference (1991 St. Louis, Mo.). Intelligent engineering systems through artificial neural networks: Proceedings of the Artificial Neural Networks in Engineering (ANNIE '91) Conference, held November 10-13, 1991, in St. Louis, Missouri, U.S.A. Herausgegeben von Dagli Cihan H. 1949-, Kumara Soundar T. 1952- und Shin Yung C. New York: ASME Press, 1991.
Den vollen Inhalt der Quelle findenAmat, Jean-Louis. Techniques avancées pour le traitement de l'information: Réseaux de neurones, logique floue, algorithmes génétiques. 2. Aufl. Toulouse: Cépaduès-Ed., 2002.
Den vollen Inhalt der Quelle findenNeural Information Processing Systems Conference. Proceedings of the 2003 conference. Cambridge, MA: MIT, 2004.
Den vollen Inhalt der Quelle findenHeaton, Jeff. Introduction to neural networks for C#. 2. Aufl. St. Louis: Heaton Research Inc., 2008.
Den vollen Inhalt der Quelle findenInternational Conference on Neural Information Processing (3rd 1996 Hong Kong). Progress in neural information processing: ICONIP'96 : proceedings of the International Conference on Neural Information Processing, Hong Kong, 24-27 September 1996. New York: Springer, 1996.
Den vollen Inhalt der Quelle finden1931-, Taylor John Gerald, Hrsg. Neural networks and their applications. Chichester: UNICOM, 1996.
Den vollen Inhalt der Quelle findenRojas, Raúl. Neural networks: A systematic introduction. Berlin: Springer-Verlag, 1996.
Den vollen Inhalt der Quelle findenBuchteile zum Thema "Compression de réseaux neuronaux"
ZHANG, Hanwei, Teddy FURON, Laurent AMSALEG und Yannis AVRITHIS. „Attaques et défenses de réseaux de neurones profonds : le cas de la classification d’images“. In Sécurité multimédia 1, 51–85. ISTE Group, 2021. http://dx.doi.org/10.51926/iste.9026.ch2.
Der volle Inhalt der QuelleLévy, Jean-Claude S. „4 - Complexité et désordre des structures magnétiques, application aux réseaux neuronaux“. In Complexité et désordre, 45–62. EDP Sciences, 2020. http://dx.doi.org/10.1051/978-2-7598-1961-4-005.
Der volle Inhalt der QuelleLévy, Jean-Claude S. „4 - Complexité et désordre des structures magnétiques, application aux réseaux neuronaux“. In Complexité et désordre, 45–62. EDP Sciences, 2020. http://dx.doi.org/10.1051/978-2-7598-1961-4.c005.
Der volle Inhalt der QuelleLévy, Jean-Claude S. „4 - Complexité et désordre des structures magnétiques, application aux réseaux neuronaux“. In Complexité et désordre, 45–62. EDP Sciences, 2020. https://doi.org/10.1051/978-2-7598-1777-1.c005.
Der volle Inhalt der QuelleBELMONTE, Romain, Pierre TIRILLY, Ioan Marius BILASCO, Nacim IHADDADENE und Chaabane DJERABA. „Détection de points de repères faciaux par modélisation spatio-temporelle“. In Analyse faciale en conditions non contrôlées, 105–49. ISTE Group, 2024. http://dx.doi.org/10.51926/iste.9111.ch3.
Der volle Inhalt der QuelleKonferenzberichte zum Thema "Compression de réseaux neuronaux"
Mdhaffar, Salima, Antoine Laurent und Yannick Estève. „Etude de performance des réseaux neuronaux récurrents dans le cadre de la campagne d'évaluation Multi-Genre Broadcast challenge 3 (MGB3)“. In XXXIIe Journées d’Études sur la Parole. ISCA: ISCA, 2018. http://dx.doi.org/10.21437/jep.2018-20.
Der volle Inhalt der QuelleBerichte der Organisationen zum Thema "Compression de réseaux neuronaux"
Djamai, N., R. A. Fernandes, L. Sun, F. Canisius und G. Hong. Python version of Simplified Level 2 Prototype Processor for retrieving canopy biophysical variables from Sentinel-2 multispectral data. Natural Resources Canada/CMSS/Information Management, 2024. http://dx.doi.org/10.4095/p8stuehwyc.
Der volle Inhalt der Quelle