Dissertations / Theses on the topic 'Stereoscopic depth'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Stereoscopic depth.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Sun, Geng. "Perceived depth control in stereoscopic cinematography." Thesis, Durham University, 2012. http://etheses.dur.ac.uk/3458/.
Full textFouche, Mark-Anthony. "Using depth information to aid stereoscopic image forensics." Diss., University of Pretoria, 2014. http://hdl.handle.net/2263/46269.
Full textDissertation (MSc)--University of Pretoria, 2014.
tm2015
Computer Science
MSc
Unrestricted
Berry, Jonathan Stuart. "Quality-controlled audio-visual depth in stereoscopic 3D media." Thesis, Durham University, 2015. http://etheses.dur.ac.uk/11286/.
Full textReinhart, William Frank. "Effects of depth cues on depth judgments using a field-sequential stereoscopic CRT display /." This resource online, 1990. http://scholar.lib.vt.edu/theses/available/etd-07132007-143145/.
Full textReinhart, William Frank. "Effects of depth cues on depth judgements using a field-sequential stereoscopic CRT display." Diss., Virginia Tech, 1990. http://hdl.handle.net/10919/38796.
Full textAdams, Daniel Lewis. "Functional organisation of the monkey visual cortex for stereoscopic depth." Thesis, University College London (University of London), 1997. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.268000.
Full textWong, Wing Shun. "The effects of matching lens focus with stereoscopic depth cues on the time taken to form a single stereoscopic image when viewing a binocular display : system prototyping and experimentation /." View abstract or full-text, 2007. http://library.ust.hk/cgi/db/thesis.pl?IELM%202007%20WONGW.
Full textAZEVEDO, ROBERTO GERSON DE ALBUQUERQUE. "SUPPORTING MULTIMEDIA APPLICATIONS IN STEREOSCOPIC AND DEPTH-BASED 3D VIDEO SYSTEMS." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2015. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=26551@1.
Full textCOORDENAÇÃO DE APERFEIÇOAMENTO DO PESSOAL DE ENSINO SUPERIOR
FUNDAÇÃO DE APOIO À PESQUISA DO ESTADO DO RIO DE JANEIRO
Tecnologias de vídeos bidimensionais (2D) têm evoluído rapidamente nos últimos anos. Apesar disso, elas não permitem uma visão realista e imersiva do mundo, pois não oferecem importantes dicas de profundidade para o sistema visual humano. Tecnologias de vídeo tridimensionais (3D) têm como objetivo preencher essa lacuna, provendo representações que permitem a reprodução de informações de profundidade em displays 3D. Embora a representação baseada em vídeos estereoscópicos ainda seja a mais utilizada até o momento, novas representações de vídeo 3D têm emergido, tais como MVV (Multi-view video), 2D plus Z (2D plus depth), MVD (Multi-view plus depth) e LDV (Layered-depth video). A integração de aplicações multimídia com mídias 3D tem o potencial de permitir novos conteúdos interativos, novas experiências com o usuário e novos modelos de negócio. Nesta tese, duas abordagens para a integração de aplicações multimídia em cadeias de transmissão de vídeo 3D fim-a-fim são propostas. Primeiro, uma abordagem que é compatível com cadeias de transmissão de vídeo 3D baseado em vídeos estereoscópicos é discutida. A proposta consiste em extensões para linguagens multimídia 2D e um processo de conversão de aplicações multimídia 2D para sua versão estereoscópica. Essa proposta não requer nenhuma alteração no exibidor de linguagens multimídia 2D para a apresentação de mídias estereoscópicas. Em uma segunda abordagem, extensões adicionais a linguagens multimídia também são propostas visando a integração de aplicações multimídia em cadeias de vídeo 3D baseado em profundidade (2D plus Z ou LDV). Além disso, uma arquitetura para a composição gráfica dessas aplicações, baseada no conceito de LDV e que permite a integração de objetos de mídia baseado em profundidade em exibidores de aplicações multimídias é apresentada. Como um exemplo de aplicação prática das proposta desta tese, ambas são implementadas e integradas em um sistema de vídeo 3D fim-a-fim baseado no Sistema Brasileiro de TV Digital.
Two-dimensional video technologies have evolved quickly in the last few years. Even so, they do not achieve a realistic and immersive view of the world since they do not offer important depth cues to the human vision system. Three-dimensional video (3DV) technologies try to fulfill this gap through video representations that enable 3D displays to provide those additional depth cues. Although CSV (Conventional Stereoscopic Video) has been the most widely-used 3DV representation, other 3DV representations have emerged during the last years. Examples of those representations include MVV (Multi-view video), 2D plus Z (2D plus depth), MVD (Multi-view plus depth), and LDV (Layered-depth Video). Although end-to-end 3DV delivery chains based on those 3DV formats have been studied, the integration of interactive multimedia applications into those 3DV delivery chains has not yet been explored enough. The integration of multimedia applications with 3D media using those new representations has the potential of allowing new rich content, user experiences and business models. In this thesis, two approaches for the integration of multimedia applications into 3DV end-to-end delivery chains are proposed. First, a backward-compatible approach for integrating CSV-based media into 2D-only multimedia languages is discussed. In this proposal, it is possible to add depth information to 2D-only media objects. The proposal consists of extensions to multimedia languages and a process for converting the original multimedia application into its stereoscopic version. It does not require any change on the language player and is ready-to-run in current CSV-based 3DV delivery chains and digital receiver s hardware. Second, extensions to multimedia languages based on layered-depth media are proposed and a software architecture for the graphics composition of multimedia applications using those extensions is presented. As an example, both proposals are implemented and integrated into an end-to-end 3DV delivery chain based on the Brazilian Digital TV System.
Froner, Barbara. "Stereoscopic 3D technologies for accurate depth tasks : a theoretical and empirical study." Thesis, Durham University, 2011. http://etheses.dur.ac.uk/3324/.
Full textHuynh, Du Quan. "Feature-based stereo vision on a mobile platform." University of Western Australia. Dept. of Computer Science, 1994. http://theses.library.uwa.edu.au/adt-WU2003.0001.
Full textTemplin, Krzysztof [Verfasser], and Hans-Peter [Akademischer Betreuer] Seidel. "Depth, shading, and stylization in stereoscopic cinematography / Krzysztof Templin. Betreuer: Hans-Peter Seidel." Saarbrücken : Saarländische Universitäts- und Landesbibliothek, 2016. http://d-nb.info/1093342951/34.
Full textIp, Ifan Betina. "Effects of visual attention on stereoscopic depth perception in the human visual cortex." Thesis, University of Oxford, 2010. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.543544.
Full textKatta, Pradeep. "Integrating depth and intensity information for vision-based head tracking." abstract and full text PDF (UNR users only), 2008. http://0-gateway.proquest.com.innopac.library.unr.edu/openurl?url_ver=Z39.88-2004&rft_val_fmt=info:ofi/fmt:kev:mtx:dissertation&res_dat=xri:pqdiss&rft_dat=xri:pqdiss:1456416.
Full textWinterbottom, Marc. "Individual Differences in the Use of Remote Vision Stereoscopic Displays." Wright State University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=wright1433453135.
Full textChang, Kam Man. "Eye fatigue when viewing stereo images presented on a binocular display : effects of matching lens focus with stereoscopic depth cues /." View abstract or full-text, 2008. http://library.ust.hk/cgi/db/thesis.pl?IELM%202008%20CHANG.
Full textGurrieri, Luis E. "The Omnidirectional Acquisition of Stereoscopic Images of Dynamic Scenes." Thèse, Université d'Ottawa / University of Ottawa, 2014. http://hdl.handle.net/10393/30923.
Full textManolas, Christos. "Sound design for stereoscopic 3D cinema : exploring current practice and the enhancement of depth perception through the use of auditory depth cues." Thesis, University of York, 2014. http://etheses.whiterose.ac.uk/13571/.
Full textTucker, Andrew James, and n/a. "Visual space attention in three-dimensional space." Swinburne University of Technology, 2006. http://adt.lib.swin.edu.au./public/adt-VSWT20070301.085637.
Full textMcIntire, John Paul. "Investigating the Relationship between Binocular Disparity, Viewer Discomfort, and Depth Task Performance on Stereoscopic 3D Displays." Wright State University / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=wright1400790668.
Full textKhaustova, Darya. "Objective assessment of stereoscopic video quality of 3DTV." Thesis, Rennes 1, 2015. http://www.theses.fr/2015REN1S021/document.
Full textThe minimum requirement for any 3D (stereoscopic images) system is to guarantee visual comfort of viewers. Visual comfort is one of the three primary perceptual attributes of 3D QoE, which can be linked directly with technical parameters of a 3D system. Therefore, the goal of this thesis is to characterize objectively the impact of these parameters on human perception for stereoscopic quality monitoring. The first part of the thesis investigates whether visual attention of the viewers should be considered when designing an objective 3D quality metrics. First, the visual attention in 2D and 3D is compared using simple test patterns. The conclusions of this first experiment are validated using complex stimuli with crossed and uncrossed disparities. In addition, we explore the impact of visual discomfort caused by excessive disparities on visual attention. The second part of the thesis is dedicated to the design of an objective model of 3D video QoE, which is based on human perceptual thresholds and acceptability level. Additionally we explore the possibility to use the proposed model as a new subjective scale. For the validation of proposed model, subjective experiments with fully controlled still and moving stereoscopic images with different types of view asymmetries are conducted. The performance is evaluated by comparing objective predictions with subjective scores for various levels of view discrepancies which might provoke visual discomfort
Einecke, Nils [Verfasser], Horst-Michael [Akademischer Betreuer] Groß, Julian P. [Akademischer Betreuer] Eggert, and Darius [Akademischer Betreuer] Burschka. "Stereoscopic depth estimation for online vision systems / Nils Einecke. Gutachter: Julian P. Eggert ; Darius Burschka. Betreuer: Horst-Michael Groß." Ilmenau : Universitätsbibliothek Ilmenau, 2013. http://d-nb.info/1031421920/34.
Full textSolh, Mashhour M. "Depth-based 3D videos: quality measurement and synthesized view enhancement." Diss., Georgia Institute of Technology, 2011. http://hdl.handle.net/1853/43743.
Full textLuzzi, Grace Maria Martins da Silva. "Narrativa tridimensional: uma investigação sobre a linguagem 3D estereoscópica." Universidade de São Paulo, 2014. http://www.teses.usp.br/teses/disponiveis/27/27161/tde-10112014-144242/.
Full textThrough an investigation of the history of techniques of illusion and immersion in popular art and entertainment, as well as theories that determine the existence of a distinct visual language, this work aims to analyse and identify the existence of a unique narrative language exclusive to 3-D stereoscopic film-making. The research takes as its point of stereoscopic analysis the stop-motion animation movie Coraline (2009, Henry Selick) designed and launched in cinemas using the stereoscopic technique.
Omori, Masako, 正子 大森, Hirofumi Nakano, 博史 中野, Masaru Miyao, and 克. 宮尾. "立体視における色彩の進出・後退効果の加齢変化について." 日本色彩学会, 2004. http://hdl.handle.net/2237/8653.
Full textLebreton, Pierre [Verfasser], Alexander [Akademischer Betreuer] Raake, Alexander [Gutachter] Raake, Ingrid E. J. [Gutachter] Heynderickx, and Marcus [Gutachter] Barkowsky. "Assessing human depth perception for 2D and 3D stereoscopic images and video and its relation with the overall 3D QoE / Pierre Lebreton ; Gutachter: Alexander Raake, Ingrid E. J. Heynderickx, Marcus Barkowsky ; Betreuer: Alexander Raake." Berlin : Technische Universität Berlin, 2016. http://d-nb.info/1156177669/34.
Full textJohansson, Anders. "Stereoscopy : Fooling the Brain into Believing There is Depth in a Flat Image." Thesis, University of Gävle, Ämnesavdelningen för datavetenskap, 2009. http://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-5082.
Full textStereoscopy is a technique that can create an illusion of depth in a flat image. There are many different methods to do this, and here is explained some of the most common and popular ways, with a bigger focus on the anaglyphic method. Since stereoscopy is an old technique, the discovery of it by Charles Wheatstone is explained briefly. In Autodesk Maya 2009, a new stereoscopic plug-in was included which makes the creation of stereoscopic imagery easier. An animated project is made during the course of this research which takes advantage of and tests the functions of the new plug-in. The main purpose of the project is to create a stereoscopic movie which utilized the anaglyph stereoscopic technique. The result is rendered stereoscopic material that is edited with Adobe Premiere Pro to create anaglyphic imagery and a full color alternative using the Infitec technique.
Slíž, Martin. "Hodnocení vlivu různých aspektů na kvalitu v 3DTV: Subjektivní testy." Master's thesis, Vysoké učení technické v Brně. Fakulta elektrotechniky a komunikačních technologií, 2014. http://www.nusl.cz/ntk/nusl-220941.
Full textSalvi, Joaquim. "An approach to coded structured light to obtain three dimensional information." Doctoral thesis, Universitat de Girona, 1998. http://hdl.handle.net/10803/7714.
Full textThe stereo vision principle is based on obtaining the three dimensional position of an object point from the position of its projective points in both camera image planes. However, before inferring 3D information, the mathematical models of both cameras have to be known. This step is known as camera calibration and is broadly describes in the thesis. Perhaps the most important problem in stereo vision is the determination of the pair of homologue points in the two images, known as the correspondence problem, and it is also one of the most difficult problems to be solved which is currently investigated by a lot of researchers. The epipolar geometry allows us to reduce the correspondence problem. An approach to the epipolar geometry is describes in the thesis. Nevertheless, it does not solve it at all as a lot of considerations have to be taken into account. As an example we have to consider points without correspondence due to a surface occlusion or simply due to a projection out of the camera scope.
The interest of the thesis is focused on structured light which has been considered as one of the most frequently used techniques in order to reduce the problems related lo stereo vision. Structured light is based on the relationship between a projected light pattern its projection and an image sensor. The deformations between the pattern projected into the scene and the one captured by the camera, permits to obtain three dimensional information of the illuminated scene. This technique has been widely used in such applications as: 3D object reconstruction, robot navigation, quality control, and so on. Although the projection of regular patterns solve the problem of points without match, it does not solve the problem of multiple matching, which leads us to use hard computing algorithms in order to search the correct matches.
In recent years, another structured light technique has increased in importance. This technique is based on the codification of the light projected on the scene in order to be used as a tool to obtain an unique match. Each token of light is imaged by the camera, we have to read the label (decode the pattern) in order to solve the correspondence problem. The advantages and disadvantages of stereo vision against structured light and a survey on coded structured light are related and discussed. The work carried out in the frame of this thesis has permitted to present a new coded structured light pattern which solves the correspondence problem uniquely and robust. Unique, as each token of light is coded by a different word which removes the problem of multiple matching. Robust, since the pattern has been coded using the position of each token of light with respect to both co-ordinate axis. Algorithms and experimental results are included in the thesis. The reader can see examples 3D measurement of static objects, and the more complicated measurement of moving objects. The technique can be used in both cases as the pattern is coded by a single projection shot. Then it can be used in several applications of robot vision.
Our interest is focused on the mathematical study of the camera and pattern projector models. We are also interested in how these models can be obtained by calibration, and how they can be used to obtained three dimensional information from two correspondence points. Furthermore, we have studied structured light and coded structured light, and we have presented a new coded structured light pattern. However, in this thesis we started from the assumption that the correspondence points could be well-segmented from the captured image. Computer vision constitutes a huge problem and a lot of work is being done at all levels of human vision modelling, starting from a)image acquisition; b) further image enhancement, filtering and processing, c) image segmentation which involves thresholding, thinning, contour detection, texture and colour analysis, and so on. The interest of this thesis starts in the next step, usually known as depth perception or 3D measurement.
Faria, José Weber Vieira de. "Criação, implementação e avaliação de um recurso didático multimídia como suporte para o ensino da neuroanatomia: realidade virtual e estereoscópica." Universidade de São Paulo, 2013. http://www.teses.usp.br/teses/disponiveis/5/5138/tde-15082014-162638/.
Full textThis thesis aims to show the process of the construction, implementation and evaluation of a tool for teaching neuroanatomy. The tool presented is accessible from personal computers, immersive, interactive, and allows photorealistic three-dimensional and stereoscopic vision. Forty fresh brains were obtained from the São Paulo Department of Death Records (SP-DDR- Serviço de Verificação de Óbitos de São Paulo (SVO-SP)) and subjected to fixation, conservation, vascular injection, staining of gray matter, white fiber dissection, turpentine and bleaching bone techniques, as needed, at the Surgical Technique and Experimental Surgery Laboratory, University of São Paulo (Laboratório de Técnica Cirúrgica e Cirurgia Experimental da Universidade de São Paulo- USP). Images of areas of interest were captured using a manual turntable built for this purpose. The images were processed with commercially available software (Photoshop CS5; Stereo Photo Maker; VRWorx2.6 for Windows) non-linear format, interactive, three-dimensional stereoscopic and stored in a database of 5337 final images. The teaching resource was applied to 84 undergraduate medical students, divided into three groups: conventional (group 1), interactive non-stereoscopic (group 2) and interactive stereoscopic (group 3). Averages on the assessment of prior knowledge did not differ significantly (P > 0.05) among groups. The tool was evaluated through a written theory test and a lab practical. Groups 2 and 3 showed the highest averages and differed significantly from Group 1 (P < 0.05), Group 2 did not differ statistically from Group 3 (p > 0.05), revealing a result of similar training on the written theory test. Observing the Effect Sizes, it was found that those were of great magnitude, indicating student training effectiveness. ANOVA results showed significant difference (P < 0.05) between group means, and the Tukey test showed statistical difference between Group 1 and the others (P < 0.05). On the lab pratical, it may be noted that similarly to the written theory test, no statistical difference between Groups 2 and 3 were found. The authors conclude that the tool presented provided a gain of knowledge for students and yielded significantly higher leaning when compared with traditional teaching resources
Hsu, Han-Shiou, and 許涵琇. "Attentional capture by stereoscopic depth." Thesis, 2014. http://ndltd.ncl.edu.tw/handle/02723447389651512148.
Full text國立交通大學
傳播研究所
102
The fundamental of stereo image is binocular disparity. By different disparity, gaze plane could be in front of screen, or behind the screen. It is a key issue that what kind of cognition process is cause. For long, whether people deal with stereo depth or not has been concerning between academic researches and industrial application. This paper examines if the visual salience theory could be manipulated by binocular disparity successfully. While depth appear with tow- dimensional image simultaneously, the section of depth will driving attention by automatic mechanism. Three experiments are examining the stereo image, position placement. Experiment 1 designed to explore whether the reaction time and eye movement could be affect by stereo disparity and position or not. Experiment 2 adopts Experiment 1, three types are classified according to disparity, and each type was added to a web AD. The results showed as following. First, depth could manipulate limited capacity model and visual salience theory, also driving attention by automatic successfully. Second, three type of depth could cause different level of resources allocation. Third, negative disparity produced much affect but not the faster one could capture attention.
Fahle, Manfred, and Tom Troscianko. "Computation of Texture and Stereoscopic Depth in Humans." 1989. http://hdl.handle.net/1721.1/6002.
Full textWu, Chao-Wei, and 吳昭威. "Stereoscopic Depth Measuring System Using Dual Structured Lights." Thesis, 2017. http://ndltd.ncl.edu.tw/handle/43833344377003212511.
Full text國立中興大學
機械工程學系所
105
This study provides a novel depth measuring system of dual structured lights. Using two computer generated holograms (CGHs) and a semiconductor laser, two independent structured light fields can be generated and projected on the reference plane. By acquiring two structured images of a test surface, we can calculate the depth information with the use of spatial frequency algorithm. First, the structured light source was designed. A symmetrical black and white line pattern was selected. It could effectively eliminate the pattern shift from the off-axis configuration. The system measuring principles were derived from the horizontal and vertical structured light fields and verified by corresponding experiments. Additionally, the fixed structured light field was changed by lens combination. The projection angle could be varied to alter the fringe pitch and further control the spatial frequency for different working distances. Finally, we combined the dual structured light fields to measure the spatial frequency on the measured plane. With a judging program, more accurate data of two systems at the same depth could be found. It is the configuration of the stereoscopic depth measuring system using dual structured lights. The experiments of horizontal and vertical structured light fields show that the formulas for off-axis and coaxial systems are consistent. The lens combination with structured light fields is able to change the magnification of the spatial frequency in a system. The linearity between two systems was finally created. The results show that this system could improve accuracy and be suitable for different working ranges.
Hu, Shang-Jen, and 胡尚仁. "Controlling depth perception for rendering stereoscopic 3D models." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/46338257466247793428.
Full text國立臺灣科技大學
色彩與照明科技研究所
101
Recently, the development of the stereo display technology progresses fast. However, less three-dimensional software provides quality stereoscopic render. In using this kind of software, users sometimes have discomfort experience when viewing stereoscopic 3D models. So, producing quality stereoscopic 3D technology becomes an important issue. Usually, user adjusts the camera position by him or herself when viewing stereoscopic models. Amateur will provoke eye fatigue and feel uncomfortable under improper camera parameters. Therefore, we test the effect of different visual stereo camera’s parameters and parallax distributions. We will propose a stereoscopic camera adjustment algorithm for improving visual comfort and preference of stereoscopic images. In this paper, we carry out a virtual stereoscopic camera system, which has different positions, convergence angles, and parallax distributions, in OpenGL. When users use this system, the vast majority of the parallax will be constrained in the safe range. It is helpful to improve 3D visual comfort and make 3D display popular.
Hsieh, Rung-Shiou, and 謝榮修. "Creating 3D Stereoscopic Image from 2D Depth Image." Thesis, 2006. http://ndltd.ncl.edu.tw/handle/09567749637530255834.
Full text中華大學
資訊工程學系碩士班
94
In this thesis, we present a method for creating stereoscopic image from original monocular image and corresponding depth image. Our method can be divided into two steps. In first step, we create two virtual views. We apply depth-image-based rendering algorithm to create two virtual view images and we solved problems of visibility and scattered hole. In second step, we deal with the newly exposed area problem generated by rendering process in step 1. We use an image inpainting method that can simultaneously propagate texture and structure from surrounding area. The newly exposed area always belongs to background area and so we use the segmentation algorithm to distinguish foreground area from background area. The concatenation of all algorithms can generate visually reasonable stereoscopic image pair for human eyes. We chose several types of image such as artificial background, natural background, occluded objects and chessboard background. Our method works well on all of these images and produces good 3D effect by displaying them on 3D equipment.
Zerebecki, Christopher Ryan. "Stereoscopic depth axis interaction: A study of performance and engagement in stereoscopic 3D games." Thesis, 2014. http://hdl.handle.net/10155/399.
Full textTai-pao, Chuang. "Use of Stereoscopic Photography to Distinguish Object Depth in Outdoor Scene." 2005. http://www.cetd.com.tw/ec/thesisdetail.aspx?etdun=U0021-2004200716265208.
Full textChau, Albert Wai Lap. "Segregation by color and stereoscopic depth in three-dimensional visual space." 1992. http://catalog.hathitrust.org/api/volumes/oclc/26813917.html.
Full textMulay, Chaitali. "A stereoscopic system using knowledge propagation to achieve accurate depth calculation." Thesis, 2008. http://library1.njit.edu/etd/fromwebvoyage.cfm?id=njit-etd2008-017.
Full textLuo, Sheng-Jie, and 羅聖傑. "Structural Stereoscopic 3D Image Editing Taking into Account the Depth Information." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/08168128192094355416.
Full text國立臺灣大學
資訊網路與多媒體研究所
101
Recently, stereoscopic 3D image editing has attracted a great deal of research interest in the computer graphics community. Different from conventional 2D image editing, it usually requires adjustment of the image content as well as the perceived depth. In this dissertation, we present a framework for structural stereoscopic 3D image editing. The framework utilizes the additional information in stereoscopic 3D image, depth structure, to manipulate the image content with respect to the user-specified editing goals. We first adapt the framework to achieve two editing operations: stereoscopic 3D image cloning and stereoscopic 3D image synthesis. The proposed stereoscopic image cloning technique performs both shape adjustment and color blending so that the stereoscopic composite is seamless in both the perceived depth and color appearance. Additionally, we introduce a versatile and robust stereoscopic image synthesis technique for synthesizing stereoscopic images with respect to editing constraints used in different editing applications (eg, image retargeting, image inpainting, texture synthesis). We demonstrate several challenging cases to show the success of our proposed techniques.
Lee, Hung-Peng, and 李宏鵬. "Hardware-aware Fast Depth Map Generation System Based on Stereoscopic Images." Thesis, 2014. http://ndltd.ncl.edu.tw/handle/69372805378123091439.
Full text國立暨南國際大學
電機工程學系
102
Capture images by stereo cameras is one way to generate 3D video. To produce virtual multi-views images for multi-view display, color image and the corresponding depth map are necessary in virtual view generation by using DIBR (Depth-Image-Based Rendering) algorithm. Therefore, this thesis proposes a hardware-aware fast depth map generation method to speed up the processing of depth map generation and maintain the quality of depth map. The depth generation is calculated by the shift position of each pixel between left view and right view in stereoscopic image (i.e. Disparity estimation). To reduce the computation complexity of disparity estimation, this thesis proposes jumping stereo matching method to obtain disparity values and uses down-sample to reduce pixels for stereo matching. To maintain the quality of depth map, this thesis proposes disparity prediction to generate continuous depth map and adopts adaptive weighted cost aggregation to elevate the best matching point. From experimental results, the adaptive weighted cost aggregation could reduce the influence of noise in stereo matching to enhance the accuracy of depth estimation. The hardware design of proposed algorithm is implemented in this thesis. The hardware architecture uses fixed matching window to calculate the disparity of image and then convert the disparity value to depth map. The experiment result reveals the proposed algorithm reduces 90% of computing time than [20] and the accuracy of depth map represents higher 0.03 in SSIM index than [21]. Finally, this thesis utilizes Modelsim to simulate the process in hardware and then apply proposed algorithm to Altera FPGA demonstration board.
Chuang, Tai-pao, and 莊臺寶. "Use of Stereoscopic Photography to Distinguish Object Depth in Outdoor Scene." Thesis, 2006. http://ndltd.ncl.edu.tw/handle/80388658829038097161.
Full text國立臺灣師範大學
資訊教育學系
94
Stereoscopic scene of the mankind is naturally caused by synthesizing two images produced by the parallax of the two eyes of human. Such being the case, mankind can distinguish the relative position of the objects. In the study of related stereovision, some persons aim at the framework of taking simulated images with two eyes using one or two cameras from front and back or simultaneously at the same time to obtain a pair of parallax mages; someone pay more attention to the theoretical analysis of the relative positions of the parallax images, and some others do the work of using parallax images as material to carry out the job of image classification, comparison and analysis. The study is mainly divided into two parts: Firstly, we used a camera to take a shot on each different position to obtain a set of parallax images and perform analysis on this set of parallax images, so as to get the calculation the intrinsic and extrinsic parameters of the camera and find the regression equation. Secondly, we use the equation to estimate the relative positions of each different object in the every set of parallax images. The study used two kinds of digital cameras, i.e. Casio Z4 and Pentax S5i to carry out the experiment to obtain individual camera’s parameter. We find the regression equation as follow, and we use it to estimate the object distance. For Casio, the regression equation is zd=24.028×b, and its focus is 24.028. For Pentax, the regression equation is zd=25.637×b, and its focus is 25.637. Of them: z: the distance between marker and camera (m), d: the disparity of the corresponding point (pixel), b: base line between two shots(cm). We took images at the Exhibition Center of Fine Arts of the Library of National Taiwan Normal University and the campus of its Branch School, and analyzed each object’s image depth.
Chang, Yu-Wei, and 張祐維. "Real-Time Stereoscopic Image Generation from Depth Map and Its Hardware Design." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/40614297081644157655.
Full text中華大學
資訊工程學系(所)
98
When we look at the object, the perceived object will cause some displacement in our eyes. This displacement is called parallax. The far the object is, the small the parallax is, and vice versa. Because of the parallax, people have feeling to stereoscopic vision. We can utilize some methods to transform the 2D image to virtual stereoscopic image. For example, DIBR (depth-image-based rendering) algorithm utilizes the depth map and virtual stereoscopic camera model of 2D image to produce the stereoscopic image with parallax. The stereoscopic image is produced by left and right shifting the original image which will cause the hole problem in the image. Therefore an image inpainting algorithm is used to fill the holes. However it is a time consuming operations. In order to achieve the real-time application, a hardware architecture is developed to generate the whole stereoscopic image. In this paper, we utilize the DIBR algorithm to generate the binocular image where a simple image inpainting algorithm is used to fill the hole. In order to increase the computing speed, the whole image is cut into pieces and each piece of image is processed separately. It is found that this kind of processing method can improve the efficiency of approximately 50% when compared with the whole image processing method. In the experiment, there are totally 2,849 logic registers in the proposed hardware design and the real-time processing in 150MHz clock rate (320 x 240 @30 fps) is achieved.
Chan, Shin-Yu, and 詹世榆. "High precision stereoscopic depth measuring system with a single shot image capture." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/28717188050075608828.
Full text國立中興大學
機械工程學系所
103
This study proposes a novel configuration for a stereoscopic depth measuring system. Structured light, which was generated by a computer generated hologram (CGH) combined with an infrared semiconductor laser, was projected onto the tested object. By means of one shot image capture and spatial frequency algorithm, the information of measurement distance can be obtained. This system is characterized by without using complex optical elements and system architecture. The measuring depth can be accurately calculated by only a single camera and a single shot. First, a CGH was adopted to replace the DOE, so as to correct the distortion of structured light field. Furthermore, by reducing its linewidth to achieve the effect of light field expansion. Based on the system architecture, a triangular relationship was established. The CGH pattern and the desired diffraction angle were designed according to the field of view (FOV) of the digital camera and distances among components. Its linewidth could be calculated by using the diffraction angle and image pixels . Then, the designed pattern was substituted into the iterative Fourier transform algorithm (IFTA) to count the phase profile of the CGH and modified for improving the asymmetric drawbacks. It was verified by using the spatial light modulation (SLM). Furthermore, elements with different depths were fabricated by the lithographic and etching processes. Their device properties were also tested. By integrating the CGH with the measuring system, correcting the digital camera position, and organizing the spatial frequency effect of various configurations, the innovative measurement system architecture for expanding the image capture area and to improving the correction accuracy was proposed. Finally, the single area depth and full range depth were respectively reconstructed by self-inductive mathematical relations and verified by experiments. The experimental results show that this system could improve the measuring accuracy with the measurement uncertainty of 1 mm in a specific setup and range. It successfully integrates the CGH design with the measuring system.
Lin, Chin-Hsin, and 林進星. "Converting 2D Video Sequences Using Object Tracking and Depth-Maps for 3D Stereoscopic Display." Thesis, 2007. http://ndltd.ncl.edu.tw/handle/62275354545388440493.
Full text中原大學
資訊工程研究所
95
A computer framework for the conversion of 2D video sequence to 3D for stereoscopic display based on inside of image frame of vanishing lines and vanishing point is presented. Given a 2D video sequence in a single-view scene, the main processes were to automatically segment and track moving objects, and to generate a depth-map with respect to a vanishing point for the scene. Depths of the motion path for the moving object could be estimated accordingly. As a result, binocular-view images were generated and recombined for stereoscopic display. The experimental results are promising, because of virtual 3D experience due to moving objects that draw attention of viewers. In addition, special 3D effects could be created with the moving objects that were further superimposed onto various backgrounds. In conclusion, our computer framework provides a systematic way of creating 3D video sequence for stereoscopic display, especially for 3D experience of moving objects in various single-view scenes.
LIN, MEI-HSIU, and 林美秀. "No-reference Stereoscopic Video Quality Assessment Using Spatial, Temporal, Depth, Transform, and Spatiotemporal Features." Thesis, 2017. http://ndltd.ncl.edu.tw/handle/v65443.
Full text國立中正大學
資訊工程研究所
105
Recently, 3D technology application is more and more widespread. Thus, humans pay more attention on stereoscopic video quality. In other words, the stereoscopic video quality assessment approaches will be widely used. No-reference stereoscopic video quality assessment technology is the most useful and effective way. Hence, no-reference stereoscopic video quality assessment technology is mainly focused in this study. First, five domain features including spatial, temporal, depth, transform, and spatiotemporal features are extracted. On the spatial domain, blurriness, blockiness, local binary pattern (LBP), and edge information are extracted. On the temporal domain, variation information of luminance, DC values, and disparity are extracted. On the depth domain, disparity and depth motion are extracted. On the transform domain, discrete wavelet transform (DWT) and discrete cosine transform (DCT) information are extracted. On the spatiotemporal domain, histogram of gradient (HOG) and 3D-DCT information are extracted. Each feature vector is obtained by using histogram statistics and normalization to the same distribution. Then, the feature vectors from the left-view and right-view videos are averaged. Here, feature selection is applied to strike out the unnecessary features to improve the performance. Here, support vector regression (SVR) is applied to estimate the stereoscopic video quality score. Finally, experimental results show that the proposed approach is better than the other NR approach on NAMA3DS1_COSPAD1 database and compares favorably with others FR and RR approaches.
CHUNG, KUO-CHUN, and 鍾國君. "No-reference Stereoscopic Video Quality Metric Computation Using Spatial, Depth, Transform, and Spatiotemporal Features." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/27s73g.
Full text國立中正大學
資訊工程研究所
106
In recent year, 3D technology application provides a new viewing experience and has become more and more widespread. Due to the reason mentioned above, humans will pay more attention on stereoscopic video quality. In other words, it is necessary to develop the stereoscopic video quality assessment approaches. Full-reference and reduced-reference stereoscopic video quality assessment methods usually obtain better performance since these approaches can make use of the information of original videos. However, it is hard to get original videos when transmitting. Hence, no-reference stereoscopic video quality assessment technology is mainly focused in this study. First, features from four domains, including spatial, depth, transform, and spatiotemporal features are extracted. On the spatial domain, blockiness, cyclopean view, binocular rivalry, cross entropy, and edge are extracted. On the depth domain, disparity saliency, depth structure, NSS, and depth entropy are extracted. On the transform domain, discrete wavelet transform (DWT) and contourlet transform information are extracted. On the spatiotemporal domain, depth motion and 3D-DCT information are extracted. The feature vectors from the left-view and right-view videos are averaged and represented as statistical feature and normalize to the same distribution. Then, support vector regression (SVR) is applied to measure the stereoscopic video quality score. Finally, experimental results show that the proposed approach is better than the other NR approach on NAMA3DS1_COSPAD1 database.
Szu-HuaWu and 吳思樺. "A Parallax Adjustable Multi-view Rendering System based on Stereoscopic Videos with Depth Information." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/61927844729721099451.
Full text國立成功大學
電腦與通信工程研究所
101
With the development of 3D techniques, the related products are now available in recent years. However, the traditional 3D television should watch the TV to feel the stereo perception with the equipment of polarized glasses, shutter glasses or red-cyan glasses. Nevertheless, it is inconvenient to wear a pair of glasses while watching 3D television. Therefore, in order to increase the application area of 3D techniques, the naked-eye stereoscopic display must be the trend of the future development. On the other hand, the multiview display system can provide the different viewing angles to perceive stereo visions. The goal of this thesis focuses on a parallax adjustable multiview rendering system. With the rich view information, we can provide better rendering results than the system with one view plus on depth. The proposed stereo-based direct single-image multiview rendering algorithm is based on the traditional depth-image-based rendering (DIBR) algorithm. It can directly render the output image with multiview information. The proposed stereo-based parallax adjustable multiview rendering system is implemented with GPU to reduce the rendering time.
Hsuan-ChihChen and 陳炫志. "Real-time Stereoscopic Image Generation Using Depth Image Based Rendering with Virtual View Point Estimation." Thesis, 2010. http://ndltd.ncl.edu.tw/handle/80039425762676350782.
Full text國立成功大學
電腦與通信工程研究所
98
Three dimensional applications are widely spread all over the world. In this thesis, we proposed a stereoscopic image generation system using Depth Image Based Rendering (DIBR) method. Due to the unfavorable artifacts produced by the DIBR algorithm, researchers have developed various algorithms to handle the problem. The most common one is to smooth the depth map before rendering. However, pre-processing of the depth map usually generates other artifacts, such as rubber sheet artifact, geometry distortion and even the degradation of depth perception. In order to avoid these defects, we proposed a solution with another approach based on virtual view point estimation. We first analyze the feature of the holes produced by the DIBR method which tends to generate big holes in the regions with large variation of depth. We develop a virtual view estimation method similar to motion estimation to fill these holes, that is, to search the most similar block and fill the blank pixels by the corresponding pixels of the best candidate block. This simple and regular filling method is suitable for hardware implementation and is able to achieve real-time applications. The results show that our method works well on images with monotonous background and the strong depth perception of the main object in the scene can be preserved.
Wang, Rou-Wen, and 王柔雯. "Depth perception of exocentric distances and movement in stereoscopic display with real and virtual targets." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/73828312747502940916.
Full text中原大學
工業與系統工程研究所
104
With the development of technology, hardware equipment and devices become more and more advanced. People now expect that virtual reality technology not only provide the users with virtual world display but also with a real-time interaction. In this case, the applications of augmented reality emerge. When the real object and the virtual object coexist, the perception of reality and its interactivity are the key. As a result, the aim of this thesis is to investigate the effect that the display of real and virtual targets, parallax and the exocentric distances have on the depth perception. The results of the study show that, in exocentric distance judgment, a Virtual-Virtual environment has the highest signed error rate, the signed error rate of Real-Virtual is higher than that of Virtual-Real, and the level of underestimation and the signed error rate of Virtual-Virtual, Real-Virtual, Virtual-Real increase with the enlargement of the exocentric distances of two objects. As for the error of position coordinates, Z-axis has the highest error. The error of first target’s Z-axis coordinate shows that Real-Virtual has the lowest error and the Virtual-Real the highest. Also, Virtual-Virtual, Real-Virtual, Virtual-Real overestimate the actual position coordinates. And the error of Virtual-Virtual and Virtual-Real on the first target’s Z-axis coordinate decreases with the enlargement of distance between the subjects and the object. What’s more, the error of the end of Z-axis shows that Virtual-Real has the lowest error and Virtual-Virtual the highest. Virtual-Virtual, Real-Virtual, Virtual-Real on the error of second target’s Z-axis coordinate will increase with the enlargement of the parallax. The error will also elevate with the enlargement of the exocentric distances between two objects. The hand movement velocity, the linear acceleration and the deceleration of Real-Virtual are larger than those in Virtual-Virtual. The linear acceleration will increase with the enlargement of the exocentric distances. On the other hand, the linear deceleration will increase with the enlargement of the distance of the object.
Liu, Cheng-Hsin, and 劉政鑫. "3D OBJECT MODELING BY INTEGRATING DEPTH INFORMATION OBTAINED FROM STEREOSCOPIC IMAGE PAIRS TAKEN FROM MULTIPLE VIEW ANGLES." Thesis, 2016. http://ndltd.ncl.edu.tw/handle/59627229848327927644.
Full text大同大學
通訊工程研究所
104
The 3D modeling of real world objects is important to the applications associated with virtual reality and 3D printing. There are many existing techniques that are able to obtain the satisfactory modeling results by scanning stationary objects in a sequential process. However, the modeling of moving objects is highly demanded in many practical applications that target on living creatures that are not likely to be stationary during a finite lapse of time. Henceforth, the modeling of moving object is most likely to be solved with the information gathered instantaneously from view angles around the object to be modeled. In this project, the stereoscopic images of the object taken simultaneously at a number of view angles are utilized to construct the 3D model of the moving object. Given conditions on the arrangement of the object to be modeled and the cameras in the real world space, the depth information of the object can be obtained from pairs of stereoscopic images taken from different view angles. The depth information is calculated based on corresponding pixels on the pair of stereoscopic images, which is then converted to the distance away from a reference axis. The reference axis is assumed passing through the center of the object, and passing the plane parallel to the image planes of the pair of cameras. The radial distance and the position of the pixel on the camera's image plane is then converted to the three-dimensional cylindrical coordinates of a voxel corresponding to this pixel. In the project, the angle that separates each set of camera pair is so chosen such that the depth fields estimated by the associated image pair are partially overlapped. The depth fields from all angles are then stitched together to form the outer surface of the object located inside a cylindrical space. With the surface information of the object, the object is represented by non-overlapping cubes or cubic voxels in the three dimensional space. The cubic voxels that constitute the surface of the object are further smoothed into surface formed by triangular planes. The modeled object is finally represented and stored in the commonly used OBJ format. The experimental results of modeling some real sample objects are acceptable in general, and prove that the proposed 3D modeling method is feasible.