Índice
Literatura científica selecionada sobre o tema "Recalage d'images à large base"
Crie uma referência precisa em APA, MLA, Chicago, Harvard, e outros estilos
Consulte a lista de atuais artigos, livros, teses, anais de congressos e outras fontes científicas relevantes para o tema "Recalage d'images à large base".
Ao lado de cada fonte na lista de referências, há um botão "Adicionar à bibliografia". Clique e geraremos automaticamente a citação bibliográfica do trabalho escolhido no estilo de citação de que você precisa: APA, MLA, Harvard, Chicago, Vancouver, etc.
Você também pode baixar o texto completo da publicação científica em formato .pdf e ler o resumo do trabalho online se estiver presente nos metadados.
Teses / dissertações sobre o assunto "Recalage d'images à large base"
Elassam, Abdelkarim. "Learning-based vanishing point detection and its application to large-baseline image registration". Electronic Thesis or Diss., Université de Lorraine, 2024. http://www.theses.fr/2024LORR0084.
Texto completo da fonteThis thesis examines the detection of vanishing points and the horizon line and their application to visual localization tasks in urban environments. Visual localization is a fundamental problem in computer vision that aims to determine the position and orientation of a camera in an environment based solely on visual information. In urban and manufactured environments, vanishing points are important visual landmarks that provide crucial information about the scene's structure, making their detection important for reconstruction and localization tasks. The thesis proposes new deep learning methods to overcome the limitations of existing approaches to vanishing point detection. The first key contribution introduces a novel approach for HL and VP detection. Unlike most existing methods, this method directly infers both the HL and an unlimited number of horizontal VPs, even those extending beyond the image frame. The second key contribution of this thesis is a structure-enhanced VP detector. This method utilizes a multi-task learning framework to estimate multiple horizontal VPs from a single image. It goes beyond simple VP detection by generating masks that identify vertical planar structures corresponding to each VP, providing valuable scene layout information. Unlike existing methods, this approach leverages contextual information and scene structures for accurate estimation without relying on detected lines. Experimental results demonstrate that this method outperforms traditional line-based methods and modern deep learning-based methods. The thesis then explores the use of vanishing points for image matching and registration, particularly in cases where images are captured from vastly different viewpoints. Despite continuous progress in feature extractors and descriptors, these methods often fail in the presence of significant scale or viewpoint variations. The proposed methods address this challenge by incorporating vanishing points and scene structures. One major challenge in using vanishing points for registration is establishing reliable correspondences, especially in large-scale scenarios. This work addresses this challenge by proposing a vanishing point detection method aided by the detection of masks of vertical scene structures corresponding to these vanishing points. To our knowledge, this is the first implementation of a method for vanishing point matching that exploits image content rather than just detected segments. This vanishing point correspondence facilitates the estimation of the camera's relative rotation, particularly in large-scale scenarios. Additionally, incorporating information from scene structures enables more reliable keypoint correspondence within these structures. Consequently, the method facilitates the estimation of relative translation, which is itself constrained by the rotation derived from the vanishing points. The quality of rotation can sometimes be impacted by the imprecision of detected vanishing points. Therefore, we propose a vanishing point-guided image matching method that is much less sensitive to the accuracy of vanishing point detection
Ben, marzouka Wissal. "Traitement possibiliste d'images, application au recalage d'images". Thesis, Ecole nationale supérieure Mines-Télécom Atlantique Bretagne Pays de la Loire, 2022. http://www.theses.fr/2022IMTA0271.
Texto completo da fonteIn this work, we propose a possibilistic geometric registration system that merges the semantic knowledge and the gray level knowledge of the images to be registered. The existing geometric registration methods are based on an analysis of the knowledge at the level of the sensors during the detection of the primitives as well as during the matching. The evaluation of the results of these geometric registration methods has limits in terms of the perfection of the precision caused by the large number of outliers. The main idea of our proposed approach is to transform the two images to be registered into a set of projections from the original images (source and target). This set is composed of images called “possibility maps”, each map of which has a single content and presents a possibilistic distribution of a semantic class of the two original images. The proposed geometric registration system based on the possibility theory presents two contexts: a supervised context and an unsupervised context. For the first case, we propose a supervised classification method based on the theory of possibilities using learning models. For the unsupervised context, we propose a possibilistic clustering method using the FCM-multicentroid method. The two proposed methods provide as a result the sets of semantic classes of the two images to be registered. We then create the knowledge bases for the proposed possibilistic registration system. We have improved the quality of the existing geometric registration in terms of precision perfection, reductionin the number of false landmarks and optimization of time complexity
Briand, Thibaud. "Image Formation from a Large Sequence of RAW Images : performance and accuracy". Thesis, Paris Est, 2018. http://www.theses.fr/2018PESC1017/document.
Texto completo da fonteThe aim of this thesis is to build a high-quality color image, containing a low level of noise and aliasing, from a large sequence (e.g. hundreds or thousands) of RAW images taken with a consumer camera. This is a challenging issue requiring to perform on the fly demosaicking, denoising and super-resolution. Existing algorithms produce high-quality images but the number of input images is limited by severe computational and memory costs. In this thesis we propose an image fusion algorithm that processes the images sequentially so that the memory cost only depends on the size of the output image. After a preprocessing step, the mosaicked (or CFA) images are aligned in a common system of coordinates using a two-step registration method that we introduce. Then, a color image is computed by accumulation of the irregularly sampled data using classical kernel regression. Finally, the blur introduced is removed by applying the inverse of the corresponding asymptotic equivalent filter (that we introduce).We evaluate the performance and the accuracy of each step of our algorithm on synthetic and real data. We find that for a large sequence of RAW images, our method successfully performs super-resolution and the residual noise decreases as expected. We obtained results similar to those obtained by slower and memory greedy methods. As generating synthetic data requires an interpolation method, we also study in detail the trigonometric polynomial and B-spline interpolation methods. We derive from this study new fine-tuned interpolation methods
Martins, Renato. "Odométrie visuelle directe et cartographie dense de grands environnements à base d'images panoramiques RGB-D". Thesis, Paris Sciences et Lettres (ComUE), 2017. http://www.theses.fr/2017PSLEM004/document.
Texto completo da fonteThis thesis is in the context of self-localization and 3D mapping from RGB-D cameras for mobile robots and autonomous systems. We present image alignment and mapping techniques to perform the camera localization (tracking) notably for large camera motions or low frame rate. Possible domains of application are localization of autonomous vehicles, 3D reconstruction of environments, security or in virtual and augmented reality. We propose a consistent localization and 3D dense mapping framework considering as input a sequence of RGB-D images acquired from a mobile platform. The core of this framework explores and extends the domain of applicability of direct/dense appearance-based image registration methods. With regard to feature-based techniques, direct/dense image registration (or image alignment) techniques are more accurate and allow us a more consistent dense representation of the scene. However, these techniques have a smaller domain of convergence and rely on the assumption that the camera motion is small.In the first part of the thesis, we propose two formulations to relax this assumption. Firstly, we describe a fast pose estimation strategy to compute a rough estimate of large motions, based on the normal vectors of the scene surfaces and on the geometric properties between the RGB-D images. This rough estimation can be used as initialization to direct registration methods for refinement. Secondly, we propose a direct RGB-D camera tracking method that exploits adaptively the photometric and geometric error properties to improve the convergence of the image alignment.In the second part of the thesis, we propose techniques of regularization and fusion to create compact and accurate representations of large scale environments. The regularization is performed from a segmentation of spherical frames in piecewise patches using simultaneously the photometric and geometric information to improve the accuracy and the consistency of the scene 3D reconstruction. This segmentation is also adapted to tackle the non-uniform resolution of panoramic images. Finally, the regularized frames are combined to build a compact keyframe-based map composed of spherical RGB-D panoramas optimally distributed in the environment. These representations are helpful for autonomous navigation and guiding tasks as they allow us an access in constant time with a limited storage which does not depend on the size of the environment
Chnafa, Christophe. "Using image-based large-eddy simulations to investigate the intracardiac flow and its turbulent nature". Thesis, Montpellier 2, 2014. http://www.theses.fr/2014MON20112/document.
Texto completo da fonteThe first objective of this thesis is to generate and analyse CFD-based databases for the intracardiac flow in realistic geometries. To this aim, an image-based CFD strategy is applied to both a pathological and a healthy human left hearts. The second objective is to illustrate how the numerical database can be analysed in order to gain insight about the intracardiac flow, mainly focusing on the unsteady and turbulent features. A numerical framework allowing insight in fluid dynamics inside patient-specific human hearts is first presented. The heart cavities and their wall dynamics are extracted from medical images, with the help of an image registration algorithm, in order to obtain a patient-specific moving numerical domain. Flow equations are written on a conformal moving computational domain, using an Arbitrary Lagrangian-Eulerian framework. Valves are modelled using immersed boundaries.Application of this framework to compute flow and turbulence statistics in both a realistic pathological and a realistic healthy human left hearts is presented. The blood flow is characterized by its transitional nature, resulting in a complex cyclic flow. Flow dynamics is analysed in order to reveal the main fluid phenomena and to obtain insights into the physiological patterns commonly detected. It is demonstrated that the flow is neither laminar nor fully turbulent, thus justifying a posteriori the use of Large Eddy Simulation.The unsteady development of turbulence is analysed from the phase averaged flow, flow statistics, the turbulent stresses, the turbulent kinetic energy, its production and through spectral analysis. A Lagrangian analysis is also presented using Lagrangian particles to gather statistical flow data. In addition to a number of classically reported features on the left heart flow, this work reveals how disturbed and transitional the flow is and describes the mechanisms of turbulence production
Yang, Liming. "Recalage robuste à base de motifs de points pseudo aléatoires pour la réalité augmentée". Thesis, Ecole centrale de Nantes, 2016. http://www.theses.fr/2016ECDN0025.
Texto completo da fonteRegistration is a very important task in Augmented Reality (AR). It provides the spatial alignment between the real environment and virtual objects. Unlike tracking (which relies on previous frame information), wide baseline localization finds the correct solution from a wide search space, so as to overcome the initialization or tracking failure problems. Nowadays, various wide baseline localization methods have been applied successfully. But for objects with no or little texture, there is still no promising method. One possible solution is to rely on the geometric information, which sometimes does not vary as much as texture or color. This dissertation focuses on new wide baseline localization methods entirely based on geometric information, and more specifically on points. I propose two novel point pattern matching algorithms, RRDM and LGC. Especially, LGC registers 2D or 3D point patterns under any known transformation type and supports multipattern recognitions. It has a linear behavior with respect to the number of points, which allows for real-time tracking. It is applied to multi targets tracking and augmentation, as well as to 3D model registration. A practical method for projector-camera system calibration based on LGC is also proposed. It can be useful for large scale Spatial Augmented Reality (SAR). Besides, I also developed a method to estimate the rotation axis of surface of revolution quickly and precisely on 3D data. It is integrated in a novel framework to reconstruct the surface of revolution on dense SLAM in real-time
Ferrante, Enzo. "Recalage déformable à base de graphes : mise en correspondance coupe-vers-volume et méthodes contextuelles". Thesis, Université Paris-Saclay (ComUE), 2016. http://www.theses.fr/2016SACLC039/document.
Texto completo da fonteImage registration methods, which aim at aligning two or more images into one coordinate system, are among the oldest and most widely used algorithms in computer vision. Registration methods serve to establish correspondence relationships among images (captured at different times, from different sensors or from different viewpoints) which are not obvious for the human eye. A particular type of registration algorithm, known as graph-based deformable registration methods, has become popular during the last decade given its robustness, scalability, efficiency and theoretical simplicity. The range of problems to which it can be adapted is particularly broad. In this thesis, we propose several extensions to the graph-based deformable registration theory, by exploring new application scenarios and developing novel methodological contributions.Our first contribution is an extension of the graph-based deformable registration framework, dealing with the challenging slice-to-volume registration problem. Slice-to-volume registration aims at registering a 2D image within a 3D volume, i.e. we seek a mapping function which optimally maps a tomographic slice to the 3D coordinate space of a given volume. We introduce a scalable, modular and flexible formulation accommodating low-rank and high order terms, which simultaneously selects the plane and estimates the in-plane deformation through a single shot optimization approach. The proposed framework is instantiated into different variants based on different graph topology, label space definition and energy construction. Simulated and real-data in the context of ultrasound and magnetic resonance registration (where both framework instantiations as well as different optimization strategies are considered) demonstrate the potentials of our method.The other two contributions included in this thesis are related to how semantic information can be encompassed within the registration process (independently of the dimensionality of the images). Currently, most of the methods rely on a single metric function explaining the similarity between the source and target images. We argue that incorporating semantic information to guide the registration process will further improve the accuracy of the results, particularly in the presence of semantic labels making the registration a domain specific problem.We consider a first scenario where we are given a classifier inferring probability maps for different anatomical structures in the input images. Our method seeks to simultaneously register and segment a set of input images, incorporating this information within the energy formulation. The main idea is to use these estimated maps of semantic labels (provided by an arbitrary classifier) as a surrogate for unlabeled data, and combine them with population deformable registration to improve both alignment and segmentation.Our last contribution also aims at incorporating semantic information to the registration process, but in a different scenario. In this case, instead of supposing that we have pre-trained arbitrary classifiers at our disposal, we are given a set of accurate ground truth annotations for a variety of anatomical structures. We present a methodological contribution that aims at learning context specific matching criteria as an aggregation of standard similarity measures from the aforementioned annotated data, using an adapted version of the latent structured support vector machine (LSSVM) framework
Monnier, Fabrice. "Amélioration de la localisation 3D de données laser terrestre à l'aide de cartes 2D ou modèles 3D". Thesis, Paris Est, 2014. http://www.theses.fr/2014PEST1114/document.
Texto completo da fonteTechnological advances in computer science (software and hardware) and particularly, GPS localization made digital models accessible to all people. In recent years, mobile mapping systems has enabled large scale mobile 3D scanning. One advantage of this technology for the urban environment is the potential ability to improve existing 2D or 3D database, especially their level of detail and variety of represented objects. Geographic database consist of a set of geometric primitives (generally 2D lines and plans or triangles in 3D) with a coarse level of detail but with the advantage of being available over wide geographical areas. They come from the fusion of various information (old campaigns performed manually, automated or hybrid design) wich may lead to manufacturing errors. The mobile mapping systems can acquire laser point clouds. These point clouds guarantee a fine level of detail up to more than one points per square centimeter. But there are some disavantages :- a large amount of data on small geographic areas that may cause problems for storage and treatment of up to several Terabyte during major acquisition,- the inherent acquisition difficulties to image the environment from the ground. In urban areas, the GPS signal required for proper georeferencing data can be disturbed by multipath or even stopped when GPS masking phenomena related to the reduction of the portion of the visible sky to capture enough satellites to find a good localization. Improve existing databases through these dataset acquired by a mobile mapping system requires alignment of these two sets. The main objective of this manuscript is to establish a pipeline of automatic processes to register these datasets together in the most reliable manner. Co-registration this data can be done in different ways. In this manuscript we have focused our work on the registration of mobile laser point cloud on geographical database by using a drift model suitable for the non rigid drift of these kind of mobile data. We have also developped a method to register geographical database containing semantics on mobile point cloud. The different optimization step performed on our methods allows to register the data fast enough for post-processing pipeline, which allows the management of large volumes of data (billions of laser points and thousands geometric primitives). We have also discussed on the problem of joint deformation. Our methods have been tested on simulated data and real data from different mission performed by IGN
González, Obando Daniel Felipe. "From digital to computational pathology for biomarker discovery". Electronic Thesis or Diss., Université Paris Cité, 2019. http://www.theses.fr/2019UNIP5185.
Texto completo da fonteHistopathology aims to analyze images of biological tissues to assess the pathologi¬cal condition of an organ and to provide a diagnosis. The advent of high-resolution slide scanners has opened the door to new possibilities for acquiring very large im¬ages (whole slide imaging), multiplexing stainings, exhaustive extraction of visual information and large scale annotations. This thesis proposes a set of algorith¬mic methods aimed at facilitating and optimizing these different aspects. First, we propose a multi-scale registration method of multi-labeled histological images based on the properties of B-splines to model, in a continuous way, a discrete image. We then propose new approaches to perform morphological analysis on weakly simple polygons generalized by straight-line graphs. They are based on the formalism of straight skeletons (an approximation of curved skeletons defined by straight segments), built with the help of motorcycle graphs. This structure makes it possible to perform mathematical morphological operations on polygons. The precision of operations on noisy polygons is obtained by refining the construction of straight skeletons. We also propose an algorithm for computing the medial axis from straight skeletons, showing it is possible to approximate the original polygonal shape. Finally, we explore weighted straight skeletons that allow directional mor¬phological operations. These morphological analysis approaches provide consistent support for improving the segmentation of objects through contextual information and performing studies related to the spatial analysis of interactions between dif¬ferent structures of interest within the tissue. All the proposed algorithms are optimized to handle gigapixel images while assuring analysis reproducibility, in particular thanks to the creation of the Icytomine plugin, an interface between Icy and Cytomine
Baumann, Michael. "SYSTEME DE SUIVI BASE SUR L'ECHOGRAPHIE 3D POUR L'ASSURANCE DE LA QUALITE DE LA DISTRIBUTION DES BIOPSIES DE LA PROSTATE ET LE GUIDAGE DU GESTE". Phd thesis, 2008. http://tel.archives-ouvertes.fr/tel-00332730.
Texto completo da fonte