Contents
Academic literature on the topic 'Systèmes embarqués (informatique) – Apprentissage automatique'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Systèmes embarqués (informatique) – Apprentissage automatique.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Dissertations / Theses on the topic "Systèmes embarqués (informatique) – Apprentissage automatique"
Ouairy, Léopold. "Analyse des vulnérabilités dans des systèmes embarqués face à des attaques par fuzzing." Thesis, Rennes 1, 2020. http://www.theses.fr/2020REN1S020.
Full textNowadays, smart cards are used daily. They enable users to pay or sign numeric documents for example. Because they contain sensible information about their user and secrets, attackers are interested in them. In particular, these attackers can use fuzzing. This attack consists in sending the most possible communication messages to a program in order to detect its vulnerabilities. This thesis aims at protecting smart cards against fuzzing. Two approaches for detecting implementation errors are proposed. The first one is from the state of the art, and it is adapted and improved for Java. It is based on an automated source code analysis. The second approach analyses source codes too, but it takes into account limitations of the first one. In particular, the precision and the dimension reduction is improved by using Natural Language Processing techniques. In addition, it studies plagiarsm techniques in order to reinforce its analysis against different implementations choices. An inter-procedural of the control flow graph is achieved for reducing the false positives. Both approaches are evaluated against three manually created oracles from OpenPGP and AES implementations for the neighborhood discovery and the anomaly detection. Results show that the second approach is improved in precision, recall and with less execution time than the first one. Its implementation, Confiance can be used in companies to secure source codes
Godin, Christelle. "Contributions à l'embarquabilité et à la robustesse des réseaux de neurones en environnement radiatif : apprentissage constructif : neurones à impulsions." École nationale supérieure de l'aéronautique et de l'espace (Toulouse ; 1972-2007), 2000. http://www.theses.fr/2000ESAE0013.
Full textAllenet, Thibault. "Quantization and adversarial robustness of embedded deep neural networks." Electronic Thesis or Diss., Université de Rennes (2023-....), 2023. https://ged.univ-rennes1.fr/nuxeo/site/esupversions/5f524c49-7a4a-4724-ae77-9afe383b7c3c.
Full textConvolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) have been broadly used in many fields such as computer vision, natural language processing and signal processing. Nevertheless, the computational workload and the heavy memory bandwidth involved in deep neural networks inference often prevents their deployment on low-power embedded devices. Moreover, deep neural networks vulnerability towards small input perturbations questions their deployment for applications involving high criticality decisions. This PhD research project objective is twofold. On the one hand, it proposes compression methods to make deep neural networks more suitable for embedded systems with low computing resources and memory requirements. On the other hand, it proposes a new strategy to make deep neural networks more robust towards attacks based on crafted inputs with the perspective to infer on edge. We begin by introducing common concepts for training neural networks, convolutional neural networks, recurrent neural networks and review the state of the art neural on deep neural networks compression methods. After this literature review we present two main contributions on compressing deep neural networks: an investigation of lottery tickets on RNNs and Disentangled Loss Quantization Aware Training (DL-QAT) on CNNs. The investigation of lottery tickets on RNNs analyze the convergence of RNNs and study its impact when subject to pruning on image classification and language modelling. Then we present a pre-processing method based on data sub-sampling that enables faster convergence of LSTM while preserving application performance. With the Disentangled Loss Quantization Aware Training (DL-QAT) method, we propose to further improve an advanced quantization method with quantization friendly loss functions to reach low bit settings like binary parameters where the application performance is the most impacted. Experiments on ImageNet-1k with DL-QAT show improvements by nearly 1\% on the top-1 accuracy of ResNet-18 with binary weights and 2-bit activations, and also show the best profile of memory footprint over accuracy when compared with other state-of-the art methods. This work then studies neural networks robustness toward adversarial attacks. After introducing the state of the art on adversarial attacks and defense mechanisms, we propose the Ensemble Hash Defense (EHD) defense mechanism. EHD enables better resilience to adversarial attacks based on gradient approximation while preserving application performance and only requiring a memory overhead at inference time. In the best configuration, our system achieves significant robustness gains compared to baseline models and a loss function-driven approach. Moreover, the principle of EHD makes it complementary to other robust optimization methods that would further enhance the robustness of the final system and compression methods. With the perspective of edge inference, the memory overhead introduced by EHD can be reduced with quantization or weight sharing. The contributions in this thesis have concerned optimization methods and a defense system to solve an important challenge, that is, how to make deep neural networks more robust towards adversarial attacks and easier to deployed on the resource limited platforms. This work further reduces the gap between state of the art deep neural networks and their execution on edge devices
Ramu, Jean-Philippe. "Efficience d'une documentation opérationnelle contextuelle sur la performance des pilotes de transport aérien." Toulouse, ISAE, 2008. http://www.theses.fr/2008ESAE0020.
Full textEmteu, Tchagou Serge Vladimir. "Réduction à la volée du volume des traces d'exécution pour l'analyse d'applications multimédia de systèmes embarqués." Thesis, Université Grenoble Alpes (ComUE), 2015. http://www.theses.fr/2015GREAM051/document.
Full textThe consumer electronics market is dominated by embedded systems due to their ever-increasing processing power and the large number of functionnalities they offer.To provide such features, architectures of embedded systems have increased in complexity: they rely on several heterogeneous processing units, and allow concurrent tasks execution.This complexity degrades the programmability of embedded system architectures and makes application execution difficult to understand on such systems.The most used approach for analyzing application execution on embedded systems consists in capturing execution traces (event sequences, such as system call invocations or context switch, generated during application execution).This approach is used in application testing, debugging or profiling.However in some use cases, execution traces generated can be very large, up to several hundreds of gigabytes.For example endurance tests, which are tests consisting in tracing execution of an application on an embedded system during long periods, from several hours to several days.Current tools and methods for analyzing execution traces are not designed to handle such amounts of data.We propose an approach for monitoring an application execution by analyzing traces on the fly in order to reduce the volume of recorded trace.Our approach is based on features of multimedia applications which contribute the most to the success of popular devices such as set-top boxes or smartphones.This approach consists in identifying automatically the suspicious periods of an application execution in order to record only the parts of traces which correspond to these periods.The proposed approach consists of two steps: a learning step which discovers regular behaviors of an application from its execution trace, and an anomaly detection step which identifies behaviors deviating from the regular ones.The many experiments, performed on synthetic and real-life datasets, show that our approach reduces the trace size by an order of magnitude while maintaining a good performance in detecting suspicious behaviors
Bahl, Gaétan. "Architectures deep learning pour l'analyse d'images satellite embarquée." Thesis, Université Côte d'Azur, 2022. https://tel.archives-ouvertes.fr/tel-03789667.
Full textThe recent advances in high-resolution Earth observation satellites and the reduction in revisit times introduced by the creation of constellations of satellites has led to the daily creation of large amounts of image data hundreds of TeraBytes per day). Simultaneously, the popularization of Deep Learning techniques allowed the development of architectures capable of extracting semantic content from images. While these algorithms usually require the use of powerful hardware, low-power AI inference accelerators have recently been developed and have the potential to be used in the next generations of satellites, thus opening the possibility of onboard analysis of satellite imagery. By extracting the information of interest from satellite images directly onboard, a substantial reduction in bandwidth, storage and memory usage can be achieved. Current and future applications, such as disaster response, precision agriculture and climate monitoring, would benefit from a lower processing latency and even real-time alerts.In this thesis, our goal is two-fold: On the one hand, we design efficient Deep Learning architectures that are able to run on low-power edge devices, such as satellites or drones, while retaining a sufficient accuracy. On the other hand, we design our algorithms while keeping in mind the importance of having a compact output that can be efficiently computed, stored, transmitted to the ground or other satellites within a constellation.First, by using depth-wise separable convolutions and convolutional recurrent neural networks, we design efficient semantic segmentation neural networks with a low number of parameters and a low memory usage. We apply these architectures to cloud and forest segmentation in satellite images. We also specifically design an architecture for cloud segmentation on the FPGA of OPS-SAT, a satellite launched by ESA in 2019, and perform onboard experiments remotely. Second, we develop an instance segmentation architecture for the regression of smooth contours based on the Fourier coefficient representation, which allows detected object shapes to be stored and transmitted efficiently. We evaluate the performance of our method on a variety of low-power computing devices. Finally, we propose a road graph extraction architecture based on a combination of fully convolutional and graph neural networks. We show that our method is significantly faster than competing methods, while retaining a good accuracy
Morozkin, Pavel. "Design and implementation of image processing and compression algorithms for a miniature embedded eye tracking system." Electronic Thesis or Diss., Sorbonne université, 2018. https://accesdistant.sorbonne-universite.fr/login?url=https://theses-intra.sorbonne-universite.fr/2018SORUS435.pdf.
Full textHuman-Machine Interaction (HMI) progressively becomes a part of coming future. Being an example of HMI, embedded eye tracking systems allow user to interact with objects placed in a known environment by using natural eye movements. The EyeDee™ portable eye tracking solution (developed by SuriCog) is an example of an HMI-based product, which includes Weetsy™ portable wire/wireless system (including Weetsy™ frame and Weetsy™ board), π-Box™ remote smart sensor and PC-based processing unit running SuriDev eye/head tracking and gaze estimation software, delivering its result in real time to a client’s application through SuriSDK (Software Development Kit). Due to wearable form factor developed eye-tracking system must conform to certain constraints, where the most important are low power consumption, low heat generation low electromagnetic radiation, low MIPS (Million Instructions per Second), as well as support wireless eye data transmission and be space efficient in general. Eye image acquisition, finding of the eye pupil ROI (Region Of Interest), compression of ROI and its wireless transmission in compressed form over a medium are very beginning steps of the entire eye tracking algorithm targeted on finding coordinates of human eye pupil. Therefore, it is necessary to reach the highest performance possible at each step in the entire chain. In contrast with state-of-the-art general-purpose image compression systems, it is possible to construct an entire new eye tracking application-specific image processing and compression methods, approaches and algorithms, design and implementation of which are the goal of this thesis
Yang, Wenlu. "Personalized physiological-based emotion recognition and implementation on hardware." Thesis, Sorbonne université, 2018. http://www.theses.fr/2018SORUS064.
Full textThis thesis investigates physiological-based emotion recognition in a digital game context and the feasibility of implementing the model on an embedded system. The following chanllenges are addressed: the relationship between emotional states and physiological responses in the game context, individual variabilities of the pschophysiological responses and issues of implementation on an embedded system. The major contributions of this thesis are : Firstly, we construct a multi-modal Database for Affective Gaming (DAG). This database contains multiple measurements concerning objective modalities: physiological signals (ECG, EDA, EMG, Respiration), screen recording, and player's face recording, as well as subjective assessments on both game event and match level. We presented statistics of the database and run a series of analysis on issues such as emotional moment detection and emotion classification, influencing factors of the overall game experience using various machine learning methods. Secondly, we investigate the individual variability in the collected data by creating an user-specific model and analyzing the optimal feature set for each individual. We proposed a personalized group-based model created the similar user groups by using the clustering techniques based on physiological traits deduced from optimal feature set. We showed that the proposed personalized group-based model performs better than the general model and user-specific model. Thirdly, we implemente the proposed method on an ARM A9 system and showed that the proposed method can meet the requirement of computation time
Caron, François. "Conception automatique de systèmes embarqués pour la téléphonie mobile 3G." Nice, 2006. http://www.theses.fr/2006NICE4091.
Full textDue to new request from the market of the mobile phone, the new third generation of standards challenges designers. Actually, these standards define a multitude of services using different bitrates that the final customer can use. This capability implies a very high complexity at some points of the network, in particular for base-stations. Base stations of the third generation mobile networks should be able to process a lot of data received from and sending to a multitude of users, communicating at the same time. PhD work presented in this document is about the definition of a new software/hardware co-design tool. This tool called BERLIOZ, « emBedded systEM exploRation for pipeLIned executiOn optimiZation », offers to embedded systems designers a set of architectural solutions to the problems posed by third mobile generations standards like the UMTS. So, this tool can analyse applications of signal processing which periodicity can be shorter than their deadlines. The computation of solutions is done with the help of a heuristic composed of two genetic algorithms. The use of these algorithms allows to achieve solutions matching a set of constraints like the silicium area of a system. After the deep study of third generation standard called UMTS, we have validated our approach on a real industrial application : the design of the “Symbol rate” part of a UMTS base-station
Lévy, Christophe. "Modèles acoustiques compacts pour les systèmes embarqués." Avignon, 2006. http://www.theses.fr/2006AVIG0143.
Full textThe amount of services offered by the last generation mobile phones has significantly increased compared to previous generations. Nowadays, phones offer new kinds of facilitiessuch as organizers, phone books, e-mail/fax, and games. At the same time, the size of mobile phones has steadily reduced. Both these observations raise an important question: ?How can we use the full facilities of a mobile phone without a large keyboard??. Voice based human-to-computer interfaces supply a friendly solution to this problem but require an embedded speech recognizer. Over the last decade, the performance of Automatic Speech Recognition (ASR) systems has improved and nowadays facilites the implementation of vocal human-to-computer interfaces. Moreover, even if scientific progress could be noticed, the potential gain (in performance) remains limited by computing resources: a relatively modern computer with a lot of memory is generally required. The main problem to embed ASR in a mobile phone is the low level of resources available in this context which classically consists of a 50/100 MHz processor, a 50/100 MHz DSP, and less than 100KB of memory. This thesis focuses on embedded speech recognition in the context of limited resources
Books on the topic "Systèmes embarqués (informatique) – Apprentissage automatique"
Network anomaly detection: A machine learning perspective. Boca Raton: CRC Press, Taylor & Francis Group, 2014.
Find full text(Editor), John Boose, ed. Machine Learning and Uncertain Reasoning (Knowledge-Based Systems Ser.: Vol. 3). Academic Press, 1990.
Find full textForsyth, R., and R. Rada. Machine Learning (Ellis Horwood Series Artificial Intelligence). Ellis Horwood, 1986.
Find full textMachine Learning in Cognitive Iot. Taylor & Francis Group, 2020.
Find full textKumar, Neeraj, and Aaisha Makkar. Machine Learning in Cognitive IoT. Taylor & Francis Group, 2020.
Find full textKumar, Neeraj, and Aaisha Makkar. Machine Learning in Cognitive IoT. Taylor & Francis Group, 2020.
Find full textKumar, Neeraj, and Aaisha Makkar. Machine Learning in Cognitive IoT. Taylor & Francis Group, 2020.
Find full textArslan, Hüseyin, and Ertuğrul Başar. Flexible and Cognitive Radio Access Technologies for 5G and Beyond. Institution of Engineering & Technology, 2020.
Find full textFlexible and Cognitive Radio Access Technologies for 5G and Beyond. Institution of Engineering & Technology, 2020.
Find full text