Dissertations / Theses on the topic 'Synthèse du contenu multimédia'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 45 dissertations / theses for your research on the topic 'Synthèse du contenu multimédia.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Harrando, Ismail. "Representation, information extraction, and summarization for automatic multimedia understanding." Electronic Thesis or Diss., Sorbonne université, 2022. http://www.theses.fr/2022SORUS097.
Full textWhether on TV or on the internet, video content production is seeing an unprecedented rise. Not only is video the dominant medium for entertainment purposes, but it is also reckoned to be the future of education, information and leisure. Nevertheless, the traditional paradigm for multimedia management proves to be incapable of keeping pace with the scale brought about by the sheer volume of content created every day across the disparate distribution channels. Thus, routine tasks like archiving, editing, content organization and retrieval by multimedia creators become prohibitively costly. On the user side, too, the amount of multimedia content pumped daily can be simply overwhelming; the need for shorter and more personalized content has never been more pronounced. To advance the state of the art on both fronts, a certain level of multimedia understanding has to be achieved by our computers. In this research thesis, we aim to go about the multiple challenges facing automatic media content processing and analysis, mainly gearing our exploration to three axes: 1. Representing multimedia: With all its richness and variety, modeling and representing multimedia content can be a challenge in itself. 2. Describing multimedia: The textual component of multimedia can be capitalized on to generate high-level descriptors, or annotations, for the content at hand. 3. Summarizing multimedia: we investigate the possibility of extracting highlights from media content, both for narrative-focused summarization and for maximising memorability
Turlier, Stéphane. "Accès et personnalisation du contenu multimédia dans un véhicule." Phd thesis, Télécom ParisTech, 2011. http://pastel.archives-ouvertes.fr/pastel-00683823.
Full textBenoit, Huet. "Étude de Contenus Multimédia: Apporter du Contexte au Contenu." Habilitation à diriger des recherches, Université de Nice Sophia-Antipolis, 2012. http://tel.archives-ouvertes.fr/tel-00744320.
Full textDelannoy, Pierre. "Performances des réseaux pour la diffusion de contenu multimédia." Evry, Télécom & Management SudParis, 2008. http://www.theses.fr/2008TELE0023.
Full textTurlier, Stéphane. "Accès et personnalisation de contenu multimédia à la demande dans un véhicule." Paris, Télécom ParisTech, 2011. https://pastel.hal.science/pastel-00683823.
Full textThe recent advent of connected vehicle platforms permits the distribution of infotainment assets to drivers and passengers with pulled and pushed workflows in a comparable manner to current mobile handsets. However, vehicles differ technically from mobile phones in terms of capability and in terms of usage. This thesis tackles the subject of personalised media delivery to motorists. We first study the technical characteristics of vehicle infotainment platforms, media assets and metadata in order to identify the requirements of a media delivery architecture for a vehicle. Based on those constraints, we have specified a media on-demand framework, which has been developed in a prototype. Afterwards, we tackle the topic of personalisation in light of two complementary point of views : on the one hand, the driver can process active personalisation when using a proper human machine interface. We present a music browser for online libraries that allows the creation of multicriteria playlists while driving. On the other hand, we analyse passive personalisation, which makes use of the driving context. We discuss the repartition of the functional components and build up a distributed architecture, which takes into account individual context preferences and their integration in the multimedia architecture that we have formerly presented. Eventually, the different solutions are evaluated according to experimental and expert methods
Kimiaei, Asadi Mariam. "Adaptation de Contenu Multimédia avec MPEG-21: Conversion de Ressources et Adaptation Sémantique de Scènes." Phd thesis, Télécom ParisTech, 2005. http://pastel.archives-ouvertes.fr/pastel-00001615.
Full textBenmokhtar, Rachid. "Fusion multi-niveaux pour l'indexation et la recherche multimédia par le contenu sémantique." Phd thesis, Télécom ParisTech, 2009. http://pastel.archives-ouvertes.fr/pastel-00005321.
Full textHamroun, Mohamed. "Indexation et recherche par contenu visuel, sémantique et multi-niveaux des documents multimédia." Thesis, Bordeaux, 2019. http://www.theses.fr/2019BORD0372.
Full textDue to the latest technological advances, the amount of multimedia data is constantly increasing. In this context, the problem is how to effectively use this data? it is necessary to set up tools to facilitate its access and manipulation.To achieve this goal, we first propose an indexation and retrieval model for video shots (or images) by their visual content (ISE). The innovative features of ISE are as follows: (i) definition of a new descriptor "PMC" and (ii) application of the genetic algorithm (GA) to improve the retrieval (PMGA).Then, we focus on the detection of concepts in video shots (LAMIRA approach). In the same context, we propose a semi-automatic annotation method for video shots in order to improve the quality of indexation based on the GA.Then, we provide a semantic indexation method separating the data level from a conceptual level and a more abstract, contextual level. This new system also incorporates mechanisms for expanding the request and relevance feedback. To add more fluidity to the user query, the user can perform a navigation using the three levels of abstraction. Two systems called VISEN and VINAS have been set up to validate these last positions.Finally, a SIRI Framework was proposed on the basis of a multi-level indexation combining our 3 systems: ISE, VINAS and VISEN. This Framework provides a two-dimensional representation of features (high level and low level) for each image
Moinard, Matthieu. "Codage vidéo hybride basé contenu par analyse/synthèse de données." Phd thesis, Telecom ParisTech, 2011. http://tel.archives-ouvertes.fr/tel-00830924.
Full textHarb, Hadi. "Classification du signal sonore en vue d'une indexation par le contenu des documents multimédia." Ecully, Ecole centrale de Lyon, 2003. http://bibli.ec-lyon.fr/exl-doc/hharb.pdf.
Full textHumans have a remarkable ability to categorise audio signals into classes, such as speech, music, explosion, etc. . . The thesis studies the capacity of developing audio classification algorithms inspired by the human perception of the audio semantic classes in the multimedia context. A model of short therm auditory memory is proposed in order to explain some psychoacoustic effects. The memory model is then simplified to constitute the basis of the Piecewise Gaussian Modelling (PGM) features. The PGM features are coupled to a mixture of neural networks to form a general audio signal classifier. The classifier was successfully applied to speech/music classification, gender identification, action detection and musical genre recognition. A synthesis of the classification effort was used in order to structure a video into "audio scenes" and "audio chapters". This work has permitted the development of an autoamtic audio indexer prototype, CYNDI
Layaïda, Nabil. "Représentation et analyses de contenu et de programmes Web." Habilitation à diriger des recherches, Université de Grenoble, 2013. http://tel.archives-ouvertes.fr/tel-00872752.
Full textPleşca, Cezar. "Supervision de contenus multimédia : adaptation de contenu, politiques optimales de préchargement et coordination causale de flux." Toulouse, INPT, 2007. http://ethesis.inp-toulouse.fr/archive/00000499/.
Full textDistributed systems information quality depends on service responsiveness, data consistency and its relevance according to user interests. The first part of this study deals with hypermedia content delivery and uses Markov Decision Processes (MDP) to derive aggresive optimal prefetching policies integrating both users habits and ressource availability. The second part addresses the partial observable contexts. We show how a ressource-based policy adaptation (MDP model) can be modulated according to user interest, using partially observable MDP (POMDP). Finally, the third part is placed in distributed multimedia applications context. We propose a coordination-level middleware for supporting flexible consistency. Our simulations show that its ability to handle several partial orders (e. G. Fifo, causal, total) makes it better than classic or [delta)-causality
Letessier, Pierre. "Découverte et exploitation d'objets visuels fréquents dans des collections multimédia." Thesis, Paris, ENST, 2013. http://www.theses.fr/2013ENST0014/document.
Full textThe main goal of this thesis is to discover frequent visual objects in large multimedia collections. As in many areas (finance, genetics, . . .), it consists in extracting a knowledge, using the occurence frequency of an object in a collection as a relevance criterion. A first contribution is to provide a formalism to the problems of mining and discovery of frequent visual objects. The second contribution is a generic method to solve these two problems, based on an iterative sampling process, and on an efficient and scalable rigid objects matching. The third contribution of this work focuses on building a likelihood function close to the perfect distribution. Experiments show that contrary to state-of-the-art methods, our approach allows to discover efficiently very small objects in several millions images. Finally, several applications are presented, including trademark logos discovery, transmedia events detection or visual-based query suggestion
Ben, Abdelali Abdessalem. "Etude de la conception d’architectures matérielles dédiées pour les traitements multimédia : indexation de la vidéo par le contenu." Dijon, 2007. http://www.theses.fr/2007DIJOS075.
Full textThis thesis constitutes a contribution to the study of content based automatic video indexing aiming at designing hardware architectures dedicated to this type of multimedia application. The content based video indexing represents an important domain that is in constant development for different types of applications such as the Internet, the interactive TV, the personal video recorders (PVR) and the security applications. The proposed study is done through concrete AV analysis techniques for video indexing and it is carried out according to different aspects related to application, technology and methodology. It is included in the context of dedicated hardware architectures design and exploitation of the new embedded systems technologies for the recent multimedia applications. Much more interest is given to the reconfigurable technology and to the new possibilities and means of the FPGA devices utilization. The first stage of this thesis is devoted to the study of the automatic content based video indexing domain. It is about the study of features and the new needs of indexing systems through the approaches and techniques currently used as well as the application fields of the new generations of these systems. This is in order to show the interest of using new architectures and technological solutions permitting to support the new requirements of this domain. The second stage is dedicated to the validation and the optimization of some visual descriptors of the MPEG-7 standard for the video temporal segmentation. This constitutes a case study through an important example of AV content analysis techniques. The proposed study constitutes also a stage of preparation for the hardware implementation of these techniques in the context of hardware accelerators design for real time automatic video indexing. Different Algorithm Architecture Adequacy aspects have been studied through the proposition of various algorithmic transformations that can be applied for the considered algorithms. The third stage of this thesis is devoted to study the design of dedicated hardware operators for video content analysis techniques as well as the exploitation of the new reconfigurable systems technologies for designing SORC dedicated to the automatic video indexing. Several hardware architectures have been proposed for the MPEG-7 descriptors and different concepts related to the exploitation of reconfigurable technology and SORC have been studied as well (methodologies and tools for designing such systems on chip, technology and methods for the dynamic and partial reconfiguration, FPGA based hardware platforms, SORC structure for video indexing, etc. )
Hinard, Yoann. "Sécurisation et tarification de la diffusion de contenu en multicast." Compiègne, 2008. http://www.theses.fr/2008COMP1766.
Full textIP multicast is an effective way to distribute vidéo content to large group of receivers. This technology is now widely used in the closed and private networks of the telecoms operators. However, IP multicast is not widely deployed over the Internet which is by nature an open network. In this thesis, we deal with two issues preventing wide deployment of IP multicast : ability to perform accounting and access control, and ability to secure the content distributed to large groups. We define a generic Authentication, Authorization and Accounting architecture for multicast content distribution which is based on the Diameter base protocol standardized by the IETF. We also define a new hash-code chaining scheme which allow to amortize the overhead of a digital signature on many other packets. This scheme allows data origin authentication and non-repudiation even with high packet loss ratio
Cabral, Marcio. "Remodelage et Re-eclairage Pour La Création et Manipulation de Contenu Interactif." Phd thesis, Université Nice Sophia Antipolis, 2011. http://tel.archives-ouvertes.fr/tel-01062521.
Full textDelezoide, Bertrand. "Modèles d'indéxation multimédia pour la description automatique de films de cinéma." Paris 6, 2006. http://www.theses.fr/2006PA066108.
Full textPapadopoulos, Hélène. "Estimation conjointe d'information de contenu musical d'un signal audio." Phd thesis, Université Pierre et Marie Curie - Paris VI, 2010. http://tel.archives-ouvertes.fr/tel-00548952.
Full textLetessier, Pierre. "Découverte et exploitation d'objets visuels fréquents dans des collections multimédia." Electronic Thesis or Diss., Paris, ENST, 2013. http://www.theses.fr/2013ENST0014.
Full textThe main goal of this thesis is to discover frequent visual objects in large multimedia collections. As in many areas (finance, genetics, . . .), it consists in extracting a knowledge, using the occurence frequency of an object in a collection as a relevance criterion. A first contribution is to provide a formalism to the problems of mining and discovery of frequent visual objects. The second contribution is a generic method to solve these two problems, based on an iterative sampling process, and on an efficient and scalable rigid objects matching. The third contribution of this work focuses on building a likelihood function close to the perfect distribution. Experiments show that contrary to state-of-the-art methods, our approach allows to discover efficiently very small objects in several millions images. Finally, several applications are presented, including trademark logos discovery, transmedia events detection or visual-based query suggestion
Badr, Mehdi. "Traitement de requêtes top-k multicritères et application à la recherche par le contenu dans les bases de données multimédia." Phd thesis, Université de Cergy Pontoise, 2013. http://tel.archives-ouvertes.fr/tel-00978770.
Full textJamin, Emmanuel. "La conception de documents audiovisuels : vers l'extraction sémantique et la réécriture interactive des archives multimédias." Paris 11, 2006. http://www.theses.fr/2006PA112215.
Full textDigitalization of audio-visual documents (DAV) improves storage techniques, which boosts innovating uses of DAV exploitation. Thus, our aim is enrich the activities of audio-visual writing based on the re-use of video fragments. After an analysis of documentary practices, we formalized the task of "multi-media read-writing" by adapting cognitive models of writing. This task brings into play the originator and the reader in a double narrative/discursive fitting, where everyone interprets the informational matter which is being presented. Within this relation, the document acts like a vector of communication and even of interaction. Therefore, we formalized a multi-media model for audio-visual design that supports MPEG7 standardization; we are talking about the "Interactive Scenario" (ScoI). ScoI is a virtual document and a suitable pool for the integration of heterogeneous fragments. This model integrates knowledge on the media, on the design’s process and on the contents access methods. The scenario is instrumented on a multimedia writing interactive system which is connected to a search system for contextualized multimedia information. We thus adapted a method of information search in order to extract multimedia fragments from a corpus of semi-structured documents and considered a recombination. A dynamic human-computer interaction process directs and assists the choices of the author in the construction of the document to be produced, or target document
Xie, Fuchun. "Tatouage sûr et robuste appliqué au traçage de documents multimédia." Phd thesis, Université Rennes 1, 2010. http://tel.archives-ouvertes.fr/tel-00592126.
Full textDaoudi, Imane. "Recherche par similarité dans les grandes bases de données multimédia : application à la recherche par le contenu dans les bases d'images." Lyon, INSA, 2009. http://theses.insa-lyon.fr/publication/2009ISAL0057/these.pdf.
Full text[The emergence of digital multimedia data is increasing. Access, sharing and retrieval of these data have become the real needs. This requires the use of powerful tools and search engine for fast and efficient access to data. The spectacular growth of technologies and numeric requires the use of powerful tools and search engine for fast and efficient access to data. My thesis work is in the field of multimedia data especially images. The main objectives is to develop a fast and efficient indexing and searching method of the k nearest neighbour which is adapted for applications in Content-based image retrieval (CBIR) and for properties of image descriptors (high volume, large dimension, etc. ). The main idea is on one hand, to provide answers to the problems of scalability and the curse of dimensionality and the other to deal with similarity problems that arise in indexing and CBIR. We propose in this thesis two different approaches. The first uses a multidimensional indexing structure based on approximation approach or filtering, which is an improvement in the RA-Blocks method. The proposed method is based on the proposal of an algorithm of subdividing the data space which improves the storage capacity of the index and the CPU times. In a second approach, we propose a multidimensional indexing method suitable for heterogeneous data (colour, texture, shape). The second proposed method combines a non linear dimensionality reduction technique with a multidimensional indexing approach based on approximation. This combination allows one hand to deal with the curse of dimensionality scalability problems and also to exploit the properties of the non-linear space to find suitable similarity measurement for the nature of manipulated data. ]
Gosselin, Philippe-Henri. "Apprentissage interactif pour la recherche par le contenu dans les bases multimédias." Habilitation à diriger des recherches, Université de Cergy Pontoise, 2011. http://tel.archives-ouvertes.fr/tel-00660316.
Full textKaced, Ahmed Réda. "Problèmes de sécurité posés par les proxies d'adaptation multimédia : proposition de solutions pour une sécurisation de bout-en-bout." Phd thesis, Télécom ParisTech, 2009. http://pastel.archives-ouvertes.fr/pastel-00005883.
Full textLy, Anh Tuan. "Accès et utilisation de documents multimédia complexes dans une bibliothèque numérique." Phd thesis, Université Paris Sud - Paris XI, 2013. http://tel.archives-ouvertes.fr/tel-00871651.
Full textMartin, Jean-Pascal. "Description sémiotique de contenus audiovisuels." Paris 11, 2005. http://www.theses.fr/2005PA112297.
Full textThree categories of descriptors are necessary to describe an audiovisual content: objects shown, processes used for film direction, and diegetic relations. The identification of the diegetic relations (those of the space-time continuum of the narration) cannot be automated. The formalisms used by the community exploit semantic descriptors that are difficult to select since they depend on contextual elements interpreted according to sophisticated knowledge. We choose to keep the human in the center of indexation's process. Two kinds of answers are provided. First, we propose a method of semiotic indexing based on the identification and the clarification of the signs that are reified at the time of the analysis. For that, we define the tetrahedral sign as a cognitive representation necessarily made up of one meant and of one meaning and possibly intensional and extensional referents. We define then the process of interpretation as a semiotic rewriting. We propose a formalism for graphs of signs (expressed with RDF+OWL schema integrated as extensions of MPEG-7) to represent the mental activity of interpretation. Second, we recommend a model of operative interaction between the man and the system that makes the reification of interpretation easier. A platform for the construction of graphs of signs based on the multi-agents paradigm allows dynamic and negotiated construction signs. Those signs are expressed according to provided syntax and grammar. Diagrams of interpretation provide to the agents micro-interpretations that may be activated in context
Mbarki, Mohamed. "Gestion de l'hétérogénéité documentaire : le cas d'un entrepôt de documents multimédia." Toulouse 3, 2008. http://thesesups.ups-tlse.fr/185/.
Full textThe knowledge society is based on three axes: the diffusion and use of information via new technologies, the deduction of knowledge induced by this information and the economic impacts which can result from this information. To offer to the actors and more particularly to the "decision makers" of this society some tools which enable them to produce and manage "knowledge" or at least "elements of knowledge" seem to be rather difficult to ensure. This difficulty is due to the dynamism of the environment and the diversity of factors influencing the information production, extraction and communication. Indeed, this information is included in documents which are collected from disseminated sources (Internet, Workflow, numerical libraries, etc. ). These documents are thus heterogeneous on the content and on the form (they can be related to various fields, they can be more or less structured, they can have various structures, they contain several type of media, are stored in several type of supports, etc). The current challenges are to conceive new applications to exploit this document heterogeneity. Having in mind these needs, the work presented in my thesis, aims to face these challenges and in particular at proposing solutions in order "to manage and create knowledge" starting from the integration of all information available on the heterogeneous documents. The handling of multimedia documents repositories constitutes the applicative framework of our proposals. Our approach is articulated around three complementary axes: (1) the representation, (2) storage (or integration) and (3) exploitation of the heterogeneous documents. Documents representation is related to the determination of information that must be preserved and the way according to which they must be organized to offer better apprehending and envisaging of their uses. The solution that we chose to meet these needs bases on the proposal for a documents model which integrates several overlapping and complementary levels of description (a generic layer and a specific one, a logical description and a semantic one). .
Mora, Elie-Gabriel. "Codage multi-vues multi-profondeur pour de nouveaux services multimédia." Thesis, Paris, ENST, 2014. http://www.theses.fr/2014ENST0007/document.
Full textThis PhD. thesis deals with improving the coding efficiency in 3D-HEVC. We propose both constrained approaches aimed towards standardization, and also more innovative approaches based on optical flow. In the constrained approaches category, we first propose a method that predicts the depth Intra modes using the ones of the texture. The inheritance is driven by a criterion measuring how much the two are expected to match. Second, we propose two simple ways to improve inter-view motion prediction in 3D-HEVC. The first adds an inter-view disparity vector candidate in the Merge list and the second modifies the derivation process of this disparity vector. Third, an inter-component tool is proposed where the link between the texture and depth quadtree structures is exploited to save both runtime and bits through a joint coding of the quadtrees. In the more innovative approaches category, we propose two methods that are based on a dense motion vector field estimation using optical flow. The first computes such a field on a reconstructed base view. It is then warped at the level of a dependent view where it is inserted as a dense candidate in the Merge list of prediction units in that view. The second method improves the view synthesis process: four fields are computed at the level of the left and right reference views using a past and a future temporal reference. These are then warped at the level of the synthesized view and corrected using an epipolar constraint. The four corresponding predictions are then blended together. Both methods bring significant coding gains which confirm the potential of such innovative solutions
Mora, Elie-Gabriel. "Codage multi-vues multi-profondeur pour de nouveaux services multimédia." Electronic Thesis or Diss., Paris, ENST, 2014. http://www.theses.fr/2014ENST0007.
Full textThis PhD. thesis deals with improving the coding efficiency in 3D-HEVC. We propose both constrained approaches aimed towards standardization, and also more innovative approaches based on optical flow. In the constrained approaches category, we first propose a method that predicts the depth Intra modes using the ones of the texture. The inheritance is driven by a criterion measuring how much the two are expected to match. Second, we propose two simple ways to improve inter-view motion prediction in 3D-HEVC. The first adds an inter-view disparity vector candidate in the Merge list and the second modifies the derivation process of this disparity vector. Third, an inter-component tool is proposed where the link between the texture and depth quadtree structures is exploited to save both runtime and bits through a joint coding of the quadtrees. In the more innovative approaches category, we propose two methods that are based on a dense motion vector field estimation using optical flow. The first computes such a field on a reconstructed base view. It is then warped at the level of a dependent view where it is inserted as a dense candidate in the Merge list of prediction units in that view. The second method improves the view synthesis process: four fields are computed at the level of the left and right reference views using a past and a future temporal reference. These are then warped at the level of the synthesized view and corrected using an epipolar constraint. The four corresponding predictions are then blended together. Both methods bring significant coding gains which confirm the potential of such innovative solutions
Aubry, Willy. "Etude et mise en place d’une plateforme d’adaptation multiservice embarquée pour la gestion de flux multimédia à différents niveaux logiciels et matériels." Thesis, Bordeaux 1, 2012. http://www.theses.fr/2012BOR14678/document.
Full textOn the one hand, technology advances have led to the expansion of the handheld devices market. Thanks to this expansion, people are more and more connected and more and more data are exchanged over the Internet. On the other hand, this huge amound of data imposes drastic constrains in order to achieve sufficient quality. The Internet is now showing its limits to assure such quality. To answer nowadays limitations, a next generation Internet is envisioned. This new network takes into account the content nature (video, audio, ...) and the context (network state, terminal capabilities ...) to better manage its own resources. To this extend, video manipulation is one of the key concept that is highlighted in this arising context. Video content is more and more consumed and at the same time requires more and more resources. Adapting videos to the network state (reducing its bitrate to match available bandwidth) or to the terminal capabilities (screen size, supported codecs, …) appears mandatory and is foreseen to take place in real time in networking devices such as home gateways. However, video adaptation is a resource intensive task and must be implemented using hardware accelerators to meet the desired low cost and real time constraints.In this thesis, content- and context-awareness is first analyzed to be considered at the network side. Secondly, a generic low cost video adaptation system is proposed and compared to existing solutions as a trade-off between system complexity and quality. Then, hardware conception is tackled as this system is implemented in an FPGA based architecture. Finally, this system is used to evaluate the indirect effects of video adaptation; energy consumption reduction is achieved at the terminal side by reducing video characteristics thus permitting an increased user experience for End-Users
Derbas, Nadia. "Contributions à la détection de concepts et d'événements dans les documents vidéos." Thesis, Grenoble, 2014. http://www.theses.fr/2014GRENM035/document.
Full textA consequence of the rise of digital technology is that the quantity of available collections of multimedia documents is permanently and strongly increasing. The indexing of these documents became both very costly and impossible to do manually. In order to be able to analyze, classify and search multimedia documents, indexing systems have been defined. However, most of these systems suffer quality or practicability issues. Their performance is limited and depends on the data volume and data variability. Indexing systems analyze multimedia documents, looking for static concepts (bicycle, chair,...), or events (wedding, protest,...). Therefore, the variability in shapes, positions, lighting or orientation of objects hinders the process. Another aspect is that systems must be scalable. They should be able to handle big data while using reasonable amount of computing time and memory.The aim of this thesis is to improve the general performance of content-based multimedia indexing systems. Four main contributions are brought in this thesis for improving different stages of the indexing process. The first one is an "early-early fusion method" that merges different information sources in order to extract their deep correlations. This method is used for violent scenes detection in movies. The second contribution is a weakly supervised method for basic concept (objects) localization in images. This can be used afterwards as a new descriptor to help detecting complex concepts (events). The third contribution tackles the noise reduction problem on ambiguously annotated data. Two methods are proposed: a shot annotation generator, and a shot weighing method. The last contribution is a generic descriptor optimization method, based on PCA and non-linear transforms.These four contributions are tested and evaluated using reference data collections, including TRECVid and MediaEval. These contributions helped our submissions achieving very good rankings in those evaluation campaigns
Berrani, Sid-Ahmed. "Recherche approximative de plus proches voisins avec contrôle probabiliste de la précision ; application à la recherche d'images par le contenu." Phd thesis, Université Rennes 1, 2004. http://tel.archives-ouvertes.fr/tel-00532854.
Full textLe, Guen Benjamin. "Adaptation du contenu spatio-temporel des images pour un codage par ondelettes." Phd thesis, Université Rennes 1, 2008. http://tel.archives-ouvertes.fr/tel-00355207.
Full textDans cette thèse, nous proposons d'aborder le problème d'adaptativité sous un angle différent. L'idée est de déformer le contenu d'une image pour l'adapter au noyau d'ondelette séparable standard. La déformation est modélisée par un maillage déformable et le critère d'adaptation utilisé est le coût de description de l'image déformée. Une minimisation énergétique similaire à une estimation de mouvement est mise en place pour calculer les paramètres du maillage. A l'issue de cette phase d'analyse, l'image est représentée par une image déformée de moindre coût de codage et par les paramètres de déformation. Après codage, transmission et décodage de ces inforrnations, l'image d'origine peut être synthétisée en inversant la déformation. Les performances en compression de ce schéma par analyse-synthèse spatiales sont étudiées et comparées à celles de JPEG2000. Visuellement, on observe une meilleure reconstruction des contours des images avec une atténuation significative de l'effet rebond.
Conservant l'idée d'adapter le contenu des images à un noyau de décomposition fixe, nous proposons ensuite un schéma de codage par analyse-synthèse spatio-temporelles dédié à la vidéo. L'analyse prend en entrée un groupe d'images (GOF) et génère en sortie un groupe d'images déformées dont le contenu est adapté à une décomposition 3D horizontale-verticale-temporelle fixe. Le schéma est conçu de sorte qu'une seule géométrie soit estimée et transmise pour l'ensemble du GOF. Des résultats de compression sont présentés en utilisant le maillage déformable pour modéliser la géométrie et le mouvement. Bien qu'une seule géométrie soit encodée, nous montrons que son coût est trop important pour permettre une amélioration significative de la qualité visuelle par rapport à un schéma par analyse-synthèse exploitant uniquement le mouvement.
Mora, Elie Gabriel. "Codage multi-vues multi-profondeur pour de nouveaux services multimédia." Phd thesis, Telecom ParisTech, 2014. http://tel.archives-ouvertes.fr/tel-01061005.
Full textBursuc, Andrei. "Indexation et recherche de contenus par objet visuel." Phd thesis, Ecole Nationale Supérieure des Mines de Paris, 2012. http://pastel.archives-ouvertes.fr/pastel-00873966.
Full textLivshin, Arie. "IDENTIFICATION AUTOMATIQUE DES INSTRUMENTS DE MUSIQUE." Phd thesis, Université Pierre et Marie Curie - Paris VI, 2007. http://tel.archives-ouvertes.fr/tel-00810688.
Full textLe, Huu Ton. "Improving image representation using image saliency and information gain." Thesis, Poitiers, 2015. http://www.theses.fr/2015POIT2287/document.
Full textNowadays, along with the development of multimedia technology, content based image retrieval (CBIR) has become an interesting and active research topic with an increasing number of application domains: image indexing and retrieval, face recognition, event detection, hand writing scanning, objects detection and tracking, image classification, landmark detection... One of the most popular models in CBIR is Bag of Visual Words (BoVW) which is inspired by Bag of Words model from Information Retrieval field. In BoVW model, images are represented by histograms of visual words from a visual vocabulary. By comparing the images signatures, we can tell the difference between images. Image representation plays an important role in a CBIR system as it determines the precision of the retrieval results.In this thesis, image representation problem is addressed. Our first contribution is to propose a new framework for visual vocabulary construction using information gain (IG) values. The IG values are computed by a weighting scheme combined with a visual attention model. Secondly, we propose to use visual attention model to improve the performance of the proposed BoVW model. This contribution addresses the importance of saliency key-points in the images by a study on the saliency of local feature detectors. Inspired from the results from this study, we use saliency as a weighting or an additional histogram for image representation.The last contribution of this thesis to CBIR shows how our framework enhances the BoVP model. Finally, a query expansion technique is employed to increase the retrieval scores on both BoVW and BoVP models
Gayrard, Catherine. "Problèmes rencontrés au cours de l'analyse du contenu en ADN et de la fraction de cellules en phase de synthèse d'ADN par cytométrie en flux dans le cancer du sein : à propos de 53 cas." Paris 5, 1994. http://www.theses.fr/1994PA05P028.
Full textAjili, Moez. "Reliability of voice comparison for forensic applications." Thesis, Avignon, 2017. http://www.theses.fr/2017AVIG0223/document.
Full textIt is common to see voice recordings being presented as a forensic trace in court. Generally, a forensic expert is asked to analyse both suspect and criminal’s voice samples in order to indicate whether the evidence supports the prosecution (same-speaker) or defence (different-speakers) hypotheses. This process is known as Forensic Voice Comparison (FVC). Since the emergence of the DNA typing model, the likelihood-ratio (LR) framework has become the new “golden standard” in forensic sciences. The LR not only supports one of the hypotheses but also quantifies the strength of its support. However, the LR accepts some practical limitations due to its estimation process itself. It is particularly true when Automatic Speaker Recognition (ASpR) systems are considered as they are outputting a score in all situations regardless of the case specific conditions. Indeed, several factors are not taken into account by the estimation process like the quality and quantity of information in both voice recordings, their phonological content or also the speakers intrinsic characteristics, etc. All these factors put into question the validity and reliability of FVC. In this Thesis, we wish to address these issues. First, we propose to analyse how the phonetic content of a pair of voice recordings affects the FVC accuracy. We show that oral vowels, nasal vowels and nasal consonants bring more speaker-specific information than averaged phonemic content. In contrast, plosive, liquid and fricative do not have a significant impact on the LR accuracy. This investigation demonstrates the importance of the phonemic content and highlights interesting differences between inter-speakers effects and intra-speaker’s ones. A further study is performed in order to study the individual speaker-specific information for each vowel based on formant parameters without any use of ASpR system. This study has revealed interesting differences between vowels in terms of quantity of speaker information. The results show clearly the importance of intra-speaker variability effects in FVC reliability estimation. Second, we investigate an approach to predict the LR reliability based only on the pair of voice recordings. We define a homogeneity criterion (NHM) able to measure the presence of relevant information and the homogeneity of this information between the pair of voice recordings. We are expecting that lowest values of homogeneity are correlated with the lowest LR’s accuracy measures, as well as the opposite behaviour for high values. The results showed the interest of the homogeneity measure for FVC reliability. Our studies reported also large differences of behaviour between FVC genuine and impostor trials. The results confirmed the importance of intra-speaker variability effects in FVC reliability estimation. The main takeaway of this Thesis is that averaging the system behaviour over a high number of factors (speaker, duration, content...) hides potentially many important details. For a better understanding of FVC approach and/or an ASpR system, it is mandatory to explore the behaviour of the system at an as-detailed-as-possible scale (The devil lies in the details)
Tripakis, Stavros. "L'analyse formelle des systèmes temporisés en pratique." Phd thesis, Université Joseph Fourier (Grenoble), 1998. http://tel.archives-ouvertes.fr/tel-00004907.
Full textArcila, Romain. "Séquences de maillages : classification et méthodes de segmentation." Phd thesis, Université Claude Bernard - Lyon I, 2011. http://tel.archives-ouvertes.fr/tel-00653542.
Full textPlesca, Cezar. "Supervision de contenus multimédia : adaptation de contenu, politiques optimales de préchargement et coordination causale de flux." Phd thesis, 2007. http://oatao.univ-toulouse.fr/7600/1/plesca.pdf.
Full textLEMLOUMA, Tayeb. "Architecture de Négociation et d'Adaptation de Services Multimédia dans des Environnements Hétérogènes." Phd thesis, 2004. http://tel.archives-ouvertes.fr/tel-00006253.
Full textFace à cette évolution, il est nécessaire de concevoir des systèmes qui permettent l'accès et l'utilisation de l'information sous une forme qui corresponde aux contraintes imposées par l'environnement. Cette thèse a pour objectif de contribuer à l'adaptation et à la négociation des contenus en considérant les limitations des utilisateurs et les contraintes de leur environnement. Nous présentons une architecture flexible appelée NAC qui permet de définir les composants qui interviennent dans la négociation et l'adaptation de contenu, et qui décrit comment ces composants sont organisés.
NAC permet plusieurs types d'adaptation : une adaptation structurelle, une adaptation sémantique et une adaptation des ressources médias. Ces types d'adaptation sont basés sur les différents contextes des clients (caractéristiques des terminaux, préférences des utilisateurs, etc.). Nous proposons également un modèle de description de contextes UPS (schémas universels pour la description des profils), un protocole de négociation et un ensemble de techniques d'adaptation. Les concepts de l'architecture NAC ont contribué aux travaux de standardisation du consortium W3C, en particulier au cadre de travail CC/PP et aux travaux sur l'indépendance des terminaux. Cette étude est complétée par une évaluation de performances qui démontre que le système proposé est exploitable dans la pratique.
Charhad, Mbarek. "Modèles de Documents Vidéo basés sur le Formalisme des Graphes Conceptuels pour l'Indexation et la Recherche par le Contenu Sémantique." Phd thesis, 2005. http://tel.archives-ouvertes.fr/tel-00399724.
Full textLes documents vidéo ont un caractère multimédia qui fait que la recherche par le contenu dans ceux-ci présente un certain nombre de spécificités. Par exemple, un concept donné (personne, objet...) peut être interprété de différentes manières : il peut être vu, il peut être entendu ou il peut être mentionné. Des combinaisons de ces cas peuvent également se produire. Naturellement, ces distinctions sont importantes pour l'utilisateur. Des requêtes impliquant le concept C comme par exemple : « rechercher les segments vidéos montrant une image de C » ou comme : « rechercher les segments vidéos dans lesquels on parle de C » sont susceptibles de produire des réponses tout à fait différentes. Dans le premier cas, on rechercherait C dans le contenu visuel tandis que dans le second, on rechercherait dans le contenu audio un segment dans la transcription duquel C est mentionné.
Cette étude s'inscrit dans un contexte de modélisation, indexation et recherche d'information multimédia. Au niveau théorique, notre contribution consiste à la proposition d'un modèle pour la représentation du contenu sémantique des documents vidéo. Ce modèle permet la prise en compte synthétique et intégrée des éléments d'informations issus de chacune des modalités (image, texte, son). L'instanciation de ce modèle est réalisée à l'aide du formalisme des graphes conceptuels. Le choix de ce formalisme est justifié par son expressivité et son adéquation au contexte d'indexation et de recherche d'information par le contenu.
Notre contribution au niveau expérimental consiste à l'implémentation (en partie) du prototype CLOVIS . Nous avons intégré le modèle proposé dans d'un système d'indexation et de recherche vidéo par le contenu pour évaluer ses apports en termes d'efficacité et de précision.
Mots-clés : Recherche d'information multimédia, indexation conceptuel, document vidéo, graphe conceptuel, ontologie.