Dissertations / Theses on the topic 'Video frame'

To see the other types of publications on this topic, follow the link: Video frame.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Video frame.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Chau, Wing San. "Key frame selection for video transcoding /." View abstract or full-text, 2005. http://library.ust.hk/cgi/db/thesis.pl?ELEC%202005%20CHAU.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Yoon, Kyongil. "Key-frame appearance analysis for video surveillance." College Park, Md. : University of Maryland, 2005. http://hdl.handle.net/1903/2818.

Full text
Abstract:
Thesis (Ph. D.) -- University of Maryland, College Park, 2005.
Thesis research directed by: Computer Science. Title from t.p. of PDF. Includes bibliographical references. Published by UMI Dissertation Services, Ann Arbor, Mich. Also available in paper.
APA, Harvard, Vancouver, ISO, and other styles
3

Czerepinski, Przemyslaw Jan. "Displaced frame difference coding for video compression." Thesis, University of Bristol, 1998. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.267009.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

SCHAPHORST, RICHARD A., and ALAN R. DEUTERMANN. "FRAME RATE REDUCTION IN VIDEO TELEMETRY SYSTEMS." International Foundation for Telemetering, 1989. http://hdl.handle.net/10150/614503.

Full text
Abstract:
International Telemetering Conference Proceedings / October 30-November 02, 1989 / Town & Country Hotel & Convention Center, San Diego, California
In video telemetry systems the transmitted picture rate, or temporal resolution, is a critical parameter in determining system performance as well as the transmitted bit rate. In many applications it is important to transmit every TV frame because the maximum temporal resolution must be maintained to analyze critical events such as an encounter between a missile and a target. Typical transmission bit rates for operation at these picture rates are 5.0 to 10.0 mbps. In other cases the frame rate can be reduced slightly to 15 or 7.5 frames/sec. without significantly reducing the value of the output video. At these frame rates it is still possible to sense the continuity of motion although some jerkiness may appear on rapidly moving objects. At these reduced frame rates the transmitted bit rate can go as low as 1.0 mbps. There is a third class of video telemetry applications where the scene is changing very slowly, and it is permissible to transmit a series of still pictures at very reduced rates. For example one picture can be transmitted every second at a transmission bit rate of 100 Kbps. The purpose of this paper is to examine operation of the standard video coding system (Range Commander Council Standard RCC 209) at conventional frame rates as well as a wide range of reduced frame rates. The following section describes the basic digital TV system which employs the standard codec. Two particular modes of operation are discussed: (1) those which reduce the frame rate by a fixed amount and vary the spatial resolution according to the complexity of the TV image; (2) those which maintain the spatial resolution at a fixed level and automatically vary the temporal resolution according to the complexity of the image. A tradeoff analysis is presented illustrating the interaction of spatial resolution, temporal resolution, and transmission bit rate. A video tape is described and presented illustrating system operation at a wide range of frame rates. Finally, conclusions are drawn.
APA, Harvard, Vancouver, ISO, and other styles
5

Mackin, Alex. "High frame rate formats for immersive video." Thesis, University of Bristol, 2017. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.730841.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Arici, Tarik. "Single and multi-frame video quality enhancement." Diss., Atlanta, Ga. : Georgia Institute of Technology, 2009. http://hdl.handle.net/1853/29722.

Full text
Abstract:
Thesis (Ph.D)--Electrical and Computer Engineering, Georgia Institute of Technology, 2009.
Committee Chair: Yucel Altunbasak; Committee Member: Brani Vidakovic; Committee Member: Ghassan AlRegib; Committee Member: James Hamblen; Committee Member: Russ Mersereau. Part of the SMARTech Electronic Thesis and Dissertation Collection.
APA, Harvard, Vancouver, ISO, and other styles
7

Levy, Alfred K. "Object tracking in low frame-rate video sequences." Honors in the Major Thesis, University of Central Florida, 2004. http://digital.library.ucf.edu/cdm/ref/collection/ETH/id/339.

Full text
Abstract:
This item is only available in print in the UCF Libraries. If this is your Honors Thesis, you can help us make it available online for use by researchers around the world by following the instructions on the distribution consent form at http://library.ucf.edu/Systems/DigitalInitiatives/DigitalCollections/InternetDistributionConsentAgreementForm.pdf You may also contact the project coordinator, Kerri Bottorff, at kerri.bottorff@ucf.edu for more information.
Bachelors
Engineering
Computer Science
APA, Harvard, Vancouver, ISO, and other styles
8

Sharon, C. M. (Colin Michael) Carleton University Dissertation Engineering Systems and Computer. "Compressed video in integrated services frame relay networks." Ottawa, 1994.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
9

King, Donald V. "(Frame) /-bridge-\ !bang! ((spill)) *sparkle* (mapping Mogadore) /." [Kent, Ohio] : Kent State University, 2008. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=kent1216759724.

Full text
Abstract:
Thesis (M.F.A.)--Kent State University, 2008.
Title from PDF t.p. (viewed Oct. 19, 2009). Advisor: Paul O'Keeffe. Keywords: Sculpture, Installation Art, Video Art. Includes bibliographical references (p. 25).
APA, Harvard, Vancouver, ISO, and other styles
10

Amin, A. M. "Geometrical analysis and rectification of thermal infrared video frame scanner imagery video frame scanner imagery and its potential applications to topographic mapping." Thesis, University of Glasgow, 1986. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.375444.

Full text
APA, Harvard, Vancouver, ISO, and other styles
11

Tu, Shing Fat. "Video interpolation for deinterlacing and frame rate up conversion /." View abstract or full-text, 2009. http://library.ust.hk/cgi/db/thesis.pl?ECED%202009%20TU.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Chang, Andy. "Fast multi-frame and multi-block selection for H.264 video coding standard /." View Abstract or Full-Text, 2003. http://library.ust.hk/cgi/db/thesis.pl?ELEC%202003%20CHANG.

Full text
Abstract:
Thesis (M. Phil.)--Hong Kong University of Science and Technology, 2003.
Includes bibliographical references (leaves 57-58). Also available in electronic version. Access restricted to campus users.
APA, Harvard, Vancouver, ISO, and other styles
13

Banelis, Justas, and Arunas Proscevicius. "Streaming Video Based on an Intelligent Frame Skipping Technique." Thesis, Blekinge Tekniska Högskola, Sektionen för datavetenskap och kommunikation, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-3476.

Full text
Abstract:
Video streaming is an important field of global communications and data processing. It is divided into server and client sides connected via network. Video streaming is concerned with delivering video data from server to client over the network as fast and with as little loss as possible. In this study the possibilities to minimize the amount of data transferred over the network in video streaming are investigated and a video streaming technique comprised of server and client sides is proposed. To expand the flexibility and adaptability of the proposed video streaming technique an operational parameter system was constructed and the parameter value ranges were defined. The proposed video streaming technique was then applied to three sample videos. Before streaming the server side of the proposed technique reduced the frame count of input videos based on operational parameter values while the client side reconstructed the skipped frames. Then the quality of the resulting videos was measured and evaluated. To evaluate the reconstructed frames and videos the PSNR measurement method was used. The study concludes that by using the proposed video streaming technique it is possible to reduce the amount of transfer data by dropping frames on the server side and reconstructing them on the client side.
APA, Harvard, Vancouver, ISO, and other styles
14

Gunturk, Bahadir K. "Multi-frame information fusion for image and video enhancement." Diss., Available online, Georgia Institute of Technology, 2004:, 2003. http://etd.gatech.edu/theses/available/etd-04072004-180015/unrestricted/gunturk%5Fbahadir%5Fk%5F200312%5Fphd.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

Brown, Kyle Russell. "Packet level frame discard for MPEG-2 video in an active network." [Gainesville, Fla.] : University of Florida, 2002. http://purl.fcla.edu/fcla/etd/UFE1000109.

Full text
Abstract:
Thesis (M.S.)--University of Florida, 2002.
Title from title page of source document. Document formatted into pages; contains vii, 67 p.; also contains graphics. Includes vita. Includes bibliographical references.
APA, Harvard, Vancouver, ISO, and other styles
16

Tiwari, Mayank. "Bit-rate allocation for multiple video streams dual-frame video coding and competitive equilibrium methods /." Diss., [La Jolla] : University of California, San Diego, 2010. http://wwwlib.umi.com/cr/ucsd/fullcit?p3402999.

Full text
APA, Harvard, Vancouver, ISO, and other styles
17

Kempster, Kurt A. "Frame rate effects on human spatial perception in video intelligence." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2000. http://handle.dtic.mil/100.2/ADA382287.

Full text
Abstract:
Thesis (M.S. in Information Technology Management) Naval Postgraduate School, Sept. 2000.
"September 2000." Thesis advisor(s): Darken, Rudolph P.; Brady, Terrance C. Includes bibliographical references (p. 77-78). Also available online.
APA, Harvard, Vancouver, ISO, and other styles
18

Stodart, NP. "The development of a video frame grabber for a PC." Thesis, Cape Technikon, 1993. http://hdl.handle.net/20.500.11838/1159.

Full text
Abstract:
Thesis (Masters Diploma (Electrical Engineering)--Cape Technikon, Cape Town, 1993
This thesis describes the design and development of a computer vision system. The system (Video Frame Grabber) will give PCUsers the potential to capture any visual image into the memory of a computer. This computer intelligible image opens the way for new development in computer photography, Image recognition and . Desktop Publishing.
APA, Harvard, Vancouver, ISO, and other styles
19

Bilen, Cagdas. "A Hybrid Approach For Full Frame Loss Concealment Of Multiview Video." Master's thesis, METU, 2007. http://etd.lib.metu.edu.tr/upload/12608559/index.pdf.

Full text
Abstract:
Multiview video is one of the emerging research areas especially among the video coding community. Transmission of multiview video over an error prone network is possible with efficient compression of these videos. But along with the studies for efficiently compressing the multiview video, new error concealment and error protection methods are also necessary to overcome the problems due to erroneous channel conditions in practical applications. In packet switching networks, packet losses may lead to block losses in a frame or the loss of an entire frame in an encoded video sequence. In recent years several algorithms are proposed to handle the loss of an entire frame efficiently. However methods for full frame losses in stereoscopic or multiview videos are limited in the literature. In this thesis a stereoscopic approach for full frame loss concealment of multiview video is proposed. In the proposed methods, the redundancy and disparity between the views and motion information between the previously decoded frames are used to estimate the lost frame. Even though multiview video can be composed of more than two views, at most three view are utilized for concealment. The performance of the proposed algorithms are tested against monoscopic methods and the conditions under which the proposed methods are superior are investigated. The proposed algorithms are applied to both stereoscopic and multiview video.
APA, Harvard, Vancouver, ISO, and other styles
20

Boyer, Eva. "Expert video modeling with video feedback to enhance gymnastics skills." [Tampa, Fla] : University of South Florida, 2008. http://purl.fcla.edu/usf/dc/et/SFE0002467.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Dikbas, Salih. "A low-complexity approach for motion-compensated video frame rate up-conversion." Diss., Georgia Institute of Technology, 2011. http://hdl.handle.net/1853/42730.

Full text
Abstract:
Video frame rate up-conversion is an important issue for multimedia systems in achieving better video quality and motion portrayal. Motion-compensated methods offer better quality interpolated frames since the interpolation is performed along the motion trajectory. In addition, computational complexity, regularity, and memory bandwidth are important for a real-time implementation. Motion-compensated frame rate up-conversion (MC-FRC) is composed of two main parts: motion estimation (ME) and motion-compensated frame interpolation (MCFI). Since ME is an essential part of MC-FRC, a new fast motion estimation (FME) algorithm capable of producing sub-sample motion vectors at low computational-complexity has been developed. Unlike existing FME algorithms, the developed algorithm considers the low complexity sub-sample accuracy in designing the search pattern for FME. The developed FME algorithm is designed in such a way that the block distortion measure (BDM) is modeled as a parametric surface in the vicinity of the integer-sample motion vector; this modeling enables low computational-complexity sub-sample motion estimation without pixel interpolation. MC-FRC needs more accurate motion trajectories for better video quality; hence, a novel true-motion estimation (TME) algorithm targeting to track the projected object motion has been developed for video processing applications, such as motion-compensated frame interpolation (MCFI), deinterlacing, and denoising. Developed TME algorithm considers not only the computational complexity and regularity but also memory bandwidth. TME is obtained by imposing implicit and explicit smoothness constraints on block matching algorithm (BMA). In addition, it employs a novel adaptive clustering algorithm to keep the low-complexity at reasonable levels yet enable exploiting more spatiotemporal neighbors. To produce better quality interpolated frames, dense motion field at the interpolation instants are obtained for both forward and backward motion vectors (MVs); then, bidirectional motion compensation using forward and backward MVs is applied by mixing both elegantly.
APA, Harvard, Vancouver, ISO, and other styles
22

Leontaris, Athanasios. "Quality, drift, and delay issues in multiple reference frame video coding." Connect to a 24 p. preview or request complete full text in PDF format. Access restricted to UC campuses, 2006. http://wwwlib.umi.com/cr/ucsd/fullcit?p3215293.

Full text
Abstract:
Thesis (Ph. D.)--University of California, San Diego, 2006.
Title from first page of PDF file (viewed July 24, 2006). Available via ProQuest Digital Dissertations. Vita. Includes bibliographical references (p. 133-138).
APA, Harvard, Vancouver, ISO, and other styles
23

McDonald, Eric Lawrence. "A video controller and distributed frame buffer for the J-machine." Thesis, Massachusetts Institute of Technology, 1995. http://hdl.handle.net/1721.1/35021.

Full text
Abstract:
Thesis (M.S.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1995.
Includes bibliographical references (p. 201-202).
by Eric Lawrence McDonald.
M.S.
APA, Harvard, Vancouver, ISO, and other styles
24

Nevits, Jeffrey A. "Design and implementation of a reconfigurable FPGA-based video frame grabber board." Thesis, This resource online, 1996. http://scholar.lib.vt.edu/theses/available/etd-10022008-063139/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
25

Aziz, Hussein. "Streaming Video over Unreliable and Bandwidth Limited Networks." Doctoral thesis, Blekinge Tekniska Högskola, Sektionen för datavetenskap och kommunikation, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-00560.

Full text
Abstract:
The main objective of this thesis is to provide a smooth video playout on the mobile device over wireless networks. The parameters that specify the wireless channel include: bandwidth variation, frame losses, and outage time. These parameters may affect the quality of the video negatively, and the mobile users may notice sudden stops during the playout video, i.e., the picture is momentarily frozen, followed by a jump from one scene to a different one. This thesis focuses on eliminating frozen pictures and reducing the amount of video data that need to be transmitted. In order to eliminate frozen scenes on the mobile screen, we propose three different techniques. In the first technique, the video frames are split into sub-frames; these sub-frames are streamed over different channels. In the second technique the sub-frames will be “crossed” and sent together with other sub-frames that are from different positions in the streaming video sequence. If some sub-frames are lost during the transmission a reconstruction mechanism will be applied on the mobile device to recreate the missing sub-frames. In the third technique, we propose a Time Interleaving Robust Streaming (TIRS) technique to stream the video frames in different order. The benefit of that is to avoid losing a sequence of neighbouring frames. A missing frame from the streaming video will be reconstructed based on the surrounding frames on the mobile device. In order to reduce the amount of video data that are streamed over limited bandwidth channels, we propose two different techniques. These two techniques are based on identifying and extracting a high motion region of the video frames. We call this the Region Of Interest (ROI); the other parts of the video frames are called the non-Region Of Interest (non-ROI). The ROI is transmitted with high quality, whereas the non-ROI is interpolated from a number of references frames. In the first technique the ROI is a fixed size region; we considered four different types of ROI and three different scenarios. The scenarios are based on the position of the reference frames in the streaming frame sequence. In the second technique the ROI is identified based on the motion in the video frames, therefore the size, position, and shape of the ROI will be different from one video to another according to the video characteristic. The videos are coded using ffmpeg to study the effect of the proposed techniques on the encoding size. Subjective and objective metrics are used to measure the quality level of the reconstructed videos that are obtained from the proposed techniques. Mean Opinion Score (MOS) measurements are used as a subjective metric based on human opinions, while for objective metric the Structural Similarity (SSIM) index is used to compare the similarity between the original frames and the reconstructed frames.
APA, Harvard, Vancouver, ISO, and other styles
26

Milburn, Paul Spencer. "Multi-variable block transforms for motion compensated digital video compression." Thesis, University of Salford, 1995. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.261859.

Full text
APA, Harvard, Vancouver, ISO, and other styles
27

Rapoport, Robert S. "The iterative frame : algorithmic video editing, participant observation & the black box." Thesis, University of Oxford, 2016. https://ora.ox.ac.uk/objects/uuid:8339bcb5-79f2-44d1-b78d-7bd28aa1956e.

Full text
Abstract:
Machine learning is increasingly involved in both our production and consumption of video. One symptom of this is the appearance of automated video editing applications. As this technology spreads rapidly to consumers, the need for substantive research about its social impact grows. To this end, this project maintains a focus on video editing as a microcosm of larger shifts in cultural objects co-authored by artificial intelligence. The window in which this research occurred (2010-2015) saw machine learning move increasingly into the public eye, and with it ethical concerns. What follows is, on the most abstract level, a discussion of why these ethical concerns are particularly urgent in the realm of the moving image. Algorithmic editing consists of software instructions to automate the creation of timelines of moving images. The criteria that this software uses to query a database is variable. Algorithmic authorship already exists in other media, but I will argue that the moving image is a separate case insofar as the raw material of text and music software can develop on its own. The performance of a trained actor can still not be generated by software. Thus, my focus is on the relationship between live embodied performance, and the subsequent algorithmic editing of that footage. This is a process that can employ other software like computer vision (to analyze the content of video) and predictive analytics (to guess what kind of automated film to make for a given user). How is performance altered when it has to communicate to human and non-human alike? The ritual of the iterative frame gives literal form to something that throughout human history has been a projection: the omniscient participant observer, more commonly known as the Divine. We experience black boxed software (AI's, specifically neural networks, which are intrinsically opaque) as functionally omniscient and tacitly allow it to edit more and more of life (e.g. filtering articles, playlists and even potential spouses). As long as it remains disembodied, we will continue to project the Divine on to the black box, causing cultural anxiety. In other words, predictive analytics alienate us from the source code of our cultural texts. The iterative frame then is a space in which these forces can be inscribed on the body, and hence narrated. The algorithmic editing of content is already taken for granted. The editing of moving images, in contrast, still requires a human hand. We need to understand the social power of moving image editing before it is delegated to automation. Practice Section: This project is practice-led, meaning that the portfolio of work was produced as it was being theorized. To underscore this, the portfolio comes at the end of the document. Video editors use artificial intelligence (AI) in a number of different applications, from deciding the sequencing of timelines to using facial and language detection to find actors in archives. This changes traditional production workflows on a number of levels. How can the single decision cut a between two frames of video speak to the larger epistemological shifts brought on by predictive analytics and Big Data (upon which they rely)? When predictive analytics begin modeling the world of moving images, how will our own understanding of the world change? In the practice-based section of this thesis, I explore how these shifts will change the way in which actors might approach performance. What does a gesture mean to AI and how will the editor decontextualize it? The set of a video shoot that will employ an element of AI in editing represents a move towards ritualization of production, summarized in the term the 'iterative frame'. The portfolio contains eight works that treat the set was taken as a microcosm of larger shifts in the production of culture. There is, I argue, metaphorical significance in the changing understanding of terms like 'continuity' and 'sync' on the AI-watched set. Theory Section In the theoretical section, the approach is broadly comparative. I contextualize the current dynamic by looking at previous shifts in technology that changed the relationship between production and post-production, notably the lightweight recording technology of the 1960s. This section also draws on debates in ethnographic filmmaking about the matching of film and ritual. In this body of literature, there is a focus on how participant observation can be formalized in film. Triangulating between event, participant observer and edit grammar in ethnographic filmmaking provides a useful analogy in understanding how AI as film editor might function in relation to contemporary production. Rituals occur in a frame that is dependent on a spatially/temporally separate observer. This dynamic also exists on sets bound for post-production involving AI, The convergence of film grammar and ritual grammar occurred in the 1960s under the banner of cinéma vérité in which the relationship between participant observer/ethnographer and the subject became most transparent. In Rouch and Morin's Chronicle of a Summer (1961), reflexivity became ritualized in the form of on-screen feedback sessions. The edit became transparent-the black box of cinema disappeared. Today as artificial intelligence enters the film production process this relationship begins to reverse-feedback, while it exists, becomes less transparent. The weight of the feedback ritual gets gradually shifted from presence and production to montage and post-production. Put differently, in cinéma vérité, the participant observer was most present in the frame. As participant observation gradually becomes shared with code it becomes more difficult to give it an embodied representation and thus its presence is felt more in the edit of the film. The relationship between the ritual actor and the participant observer (the algorithm) is completely mediated by the edit, a reassertion of the black box, where once it had been transparent. The crucible for looking at the relationship between algorithmic editing, participant observation and the black box is the subject in trance. In ritual trance the individual is subsumed by collective codes. Long before the advent of automated editing trance was an epistemological problem posed to film editing. In the iterative frame, for the first time, film grammar can echo ritual grammar and indeed become continuous with it. This occurs through removing the act of cutting from the causal world, and projecting this logic of post-production onto performance. Why does this occur? Ritual and specifically ritual trance is the moment when a culture gives embodied form to what it could not otherwise articulate. The trance of predictive analytics-the AI that increasingly choreographs our relationship to information-is the ineffable that finds form in the iterative frame. In the iterative frame a gesture never exists in a single instance, but in a potential state. The performers in this frame begin to understand themselves in terms of how automated indexing processes reconfigure their performance. To the extent that gestures are complicit with this mode of databasing they can be seen as votive toward the algorithmic. The practice section focuses on the poetics of this position. Chapter One focuses on cinéma vérité as a moment in which the relationship between production and post-production shifted as a function of more agile recording technology, allowing the participant observer to enter the frame. This shift becomes a lens to look at changes that AI might bring. Chapter Two treats the work of Pierre Huyghe as a 'liminal phase' in which a new relationship between production and post-production is explored. Finally, Chapter Three looks at a film in which actors perform with awareness that footage will be processed by an algorithmic edit.
The conclusion looks at the implications this way of relating to AI-especially commercial AI-through embodied performance could foster a more critical relationship to the proliferating black-boxed modes of production.
APA, Harvard, Vancouver, ISO, and other styles
28

Le, Xuesong. "Automatic Video Object Segmentation and Tracking." Thesis, Griffith University, 2018. http://hdl.handle.net/10072/384281.

Full text
Abstract:
The segmentation of moving objects with unknown motion under a non-stationary camera is a difficult problem because the camera and object motions are disparate. In addition, the uncertain motion of typical surveillance targets, e.g. slow moving or stopped, abrupt acceleration, and uniform motion makes a single salient motion detection algorithm unsuitable for accurate tracking. This thesis solved this problem by blending the information from the image registration, the frame differences, motion-based segmentation and the spatial segmentation in a non-declarative approach. The image registration is used to generate a motion compensated current frame. Next, temporal differencing and adaptive Kalman filter motion detections are applied to detect the changes in the compensated frames. Finally, detected changes from two motion detection algorithms and the spatial segmentation are combined to identify the moving regions. Experimental results comparing the proposed and other competing methods are evaluated objectively in various accuracy metrics and show that the proposed method achieves promising motion results for a variety of real environments.
Thesis (PhD Doctorate)
Doctor of Philosophy (PhD)
School of Info & Comm Tech
Science, Environment, Engineering and Technology
Full Text
APA, Harvard, Vancouver, ISO, and other styles
29

Mopidevi, Vydik. "Video Quality Evaluation using NR metric with Detection of Frame Dips and Drops." Thesis, Blekinge Tekniska Högskola, Sektionen för ingenjörsvetenskap, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-4153.

Full text
Abstract:
Wireless technology has become the most interesting field in the area of communications and networking. Video transmission plays a major role in this area. Data transmission via wireless communications is a major task, especially to send the data without loss of packets. The digital video decoder produces dropped or repeated frames from the encoded video due to packet loss. Brief frame drops of one frame duration are referred as dips. This research work investigates how to identify these frame dips and drops and also to measure the quality of the video which contain frame dips and frame drops. A dynamic threshold estimator is implemented to identify these frame dips. The qualitative analysis of a video is performed using a No Reference metric. In this metric the quality assessment can be achieved by identification of frame dips and drops and calculating Modified Fraction of Dropped Frames. Dips and abrupt temporal variation occurring at the end of frame freezing event of a video fluidity impairment caused by frame dropping will be taken into account using this metric. The parameters which are generated from Structural Similarity Index Metric (SSIM) analysis are efficient enough to compare the video quality and estimate the metric performance objectively. Our proposed metric has predicted the SSIM with a Pearson correlation coefficient of 0.99 in test1, 0.92 in test2 and 0.99 in test3. Therefore in terms of prediction accuracy the proposed metric has a good correlation with the deployed metric and obtained robustness to our approach.
Trådlös teknik har blivit den mest intressanta fältet inom kommunikation och nätverk. Videoöverföring spelar en viktig roll på detta område. Dataöverföring via trådlös kommunikation är en stor uppgift, särskilt för att skicka data utan förlust av paket. Den digitala videoavkodare producerar tappade eller upprepade bildrutor från den kodade videon grund paketförluster. Kort ram droppar en ram varaktighet kallat dips. Denna forskning undersöker hur man identifierar dessa ram dips och droppar och även att mäta kvaliteten på video som innehåller ram dips och droppar ram. En dynamisk tröskel estimatorn genomförs för att identifiera dessa ram dips. Den kvalitativa analysen av ett videoklipp utförs med hjälp av en No Reference metrisk. I denna statistik kvalitetsbedömningen kan uppnås genom identifiering av ram nedgångar och droppar och beräkning modifierad Andel bildrutor. Dips och abrupt temporal variation inträffar i slutet av ramen frysning händelse av en video flytbarhet orsakats av ram släppa kommer att beaktas med denna metrik. De parametrar som genereras från strukturell likhet Index Metric (SSIM) analys är effektiva nog att jämföra videokvalitet och uppskatta det metriska prestanda objektivt. Vår föreslagna värdet har förutspådde SSIM med Pearson korrelationskoefficient av 0,99 i test1, 0,92 i test2 och 0,99 i test3. Därför gäller förutsägelse noggrannhet föreslagna måttet har en god korrelation med utplacerade metriska och fick robusthet till vår strategi.
Vydik Mopidevi, D.No: 5-162, Sri Vijetha Public School, Tallapudi, West Godavari District, Pin Code: 534341 Phone Number: Home: 00813282979 Work: 0046734784445 Email: vydik.mopidevi@gmail.com
APA, Harvard, Vancouver, ISO, and other styles
30

Sasongko, Johannes. "Automatic generation of effective video summaries." Thesis, Queensland University of Technology, 2011. https://eprints.qut.edu.au/45473/1/Johannes_Sasongko_Thesis.pdf.

Full text
Abstract:
As the popularity of video as an information medium rises, the amount of video content that we produce and archive keeps growing. This creates a demand for shorter representations of videos in order to assist the task of video retrieval. The traditional solution is to let humans watch these videos and write textual summaries based on what they saw. This summarisation process, however, is time-consuming. Moreover, a lot of useful audio-visual information contained in the original video can be lost. Video summarisation aims to turn a full-length video into a more concise version that preserves as much information as possible. The problem of video summarisation is to minimise the trade-off between how concise and how representative a summary is. There are also usability concerns that need to be addressed in a video summarisation scheme. To solve these problems, this research aims to create an automatic video summarisation framework that combines and improves on existing video summarisation techniques, with the focus on practicality and user satisfaction. We also investigate the need for different summarisation strategies in different kinds of videos, for example news, sports, or TV series. Finally, we develop a video summarisation system based on the framework, which is validated by subjective and objective evaluation. The evaluation results shows that the proposed framework is effective for creating video skims, producing high user satisfaction rate and having reasonably low computing requirement. We also demonstrate that the techniques presented in this research can be used for visualising video summaries in the form web pages showing various useful information, both from the video itself and from external sources.
APA, Harvard, Vancouver, ISO, and other styles
31

Gupta, P., and V. B. Gadicha. "Enhancement of low quality degraded video using haar wavelet decomposition technique." Thesis, Sumy State University, 2017. http://essuir.sumdu.edu.ua/handle/123456789/55756.

Full text
Abstract:
The point recommends the general system for video improvement, object location and following continue at the same time in an aggressive and helpful way .Video is a gathering of successive pictures with a consistent time interim. So video can give more data about the article when situations are changing regarding time. Thusly, physically taking care of recordings are entirely unimaginable. So there is need of a computerized gadget to handle these recordings. Numerous calculations and innovation have been created to robotize video improvement and checking the article in a video document. Video improvement with item identification and following is a one of the testing errand in PC vision. Additionally, following of an article essentially includes two going before steps object identification and item representation. Object discovery is performed to check presence of items in video and to correctly find that article .Object following is a procedure of dividing a district of enthusiasm from a video scene and monitoring its movement, position and impediment. The following is performed by observing articles' spatial and transient changes amid a video grouping, including its nearness, position, size, shape, and so forth. Video improvement with article following is utilized as a part of a few applications, for example, video reconnaissance, robot vision, movement observing.
APA, Harvard, Vancouver, ISO, and other styles
32

Aydogmus, Sercan. "Error Concealment In 3d Video." Master's thesis, METU, 2011. http://etd.lib.metu.edu.tr/upload/12613914/index.pdf.

Full text
Abstract:
The advances in multimedia technologies increased the interest in utilizing three dimensional (3D) video applications in mobile devices. However, wireless transmission is significantly prone to errors. Typically, packets may be corrupted or lost due to transmission errors, causing blocking artifacts. Furthermore, because of compression and coding, the error propagates through the sequence and salient features of the video cannot be recovered until a key-frame or synchronization-frame is correctly received. Without the use of concealment and enhancement techniques, visible artifacts would inevitably and regularly appear in the decoded stream. In this thesis, error concealment techniques for full frame losses in depth plus video and stereo video structures are implemented and compared. Temporal and interview correlations are utilized to predict the lost frames while considering the memory usage and computational complexity.The concealment methods are implemented on jm17.2 decoder which is based on H.264/AVC specifications [1]. The simulation results are compared with the simple frame copy (FC) method for different sequences having different characteristics.
APA, Harvard, Vancouver, ISO, and other styles
33

Tseng, Chih-feng, and 曾志峰. "Frame-size Scaling in Video Decoder." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/74870716521154185571.

Full text
Abstract:
碩士
國立中正大學
電機工程所
97
Nowadays, many handheld multimedia devices have the capability of watching TV programs. However, such handheld devices are restricted to low-resolution display screens. Accordingly, the frame size of videos must be downscaled before displaying. The conventional methods that include the inverse transform and spatial-domain frame-size downscaling consume large computational complexities. Hence, frequency-domain frame-size downscaling method is devised to reduce the computational complexity of the inverse transform and frame-size downscaling. In this thesis, we propose a kind of frequency-domain frame-size downscaling method which integrates the Inverse Discrete Cosine Transform (IDCT) and downscaling. Additionally, the integrated method is optimized by using two coefficient compensation manners to preserve good picture quality. The proposed integrated method based on coefficient compensation can reduce about 55% complexity of the conventional IDCT and downscaling under the same picture quality. Therefore, the proposed method can be widely used in various multimedia devices to enjoy high-quality video displaying under a low computational complexity.
APA, Harvard, Vancouver, ISO, and other styles
34

Dhara, Sobhan Kanti. "Some Intra-Frame and Inter-Frame Processing Schemes for Efficient Video Compression." Thesis, 2015. http://ethesis.nitrkl.ac.in/7073/1/Some_Dhara_2015.pdf.

Full text
Abstract:
Rapid increase in digital applications due to recent advances in digital communication and devices needs significant video information storing, processing and transmitting. But the amount of original captured video data is huge and thus makes the system complex in all kind of video processing.But applications demand a faster transmission in different sized electronic devices with good quality.Along with, limited bandwidth and memory for storage makes it challenging. These practical constraints for processing a huge amount of video data, makes video compression as active and challenging field of research. The aim of video compression is to remove redundancy of raw video while maintaining the quality and fidelity. For inter frame processing, motion estimation technique is significantly used to reduce temporal redundancy in almost all the video coding standards e.g. MPEG2, MPEG4, H264/AVC which uses state-of-art algorithm to provide higher compression with a perceptual quality.Though motion estimation is main contributor for higher compression, this is the most computationally complex part of video coding tools. So, it is always a requirement to design an algorithm that is both faster and accurate and provides higher compression but good quality output. The goal of this project is to propose an algorithm for motion estimation which will meet all the requirements and overcome all the practical limitations. In this thesis we analyze the motion of video sequences and some novel block matching based motion estimation algorithms are proposed to improve video coding efficiency in inter frame processing. Particle Swarm Optimization technique and Differential Evolutionary model is used for fast and accurate motion estimation and compensation. Spatial and temporal correlation is adapted for initial population. We followed some strategy for adaptive generations, particle population, particle location history preservation and exploitation. The experimental result shows that our proposed algorithm is efficient to maintain the accuracy. There is significant reduction of search points and thus computational complexity while achieving comparable performance in video coding. Spatial domain redundancy is reduced skipping the irrelevant or spatially co-related data by different sub-sampling algorithm.The sub-sampled intra-frame is up-sampled at the receiver side. The up-sampled high resolution frame requires to have good quality . The existing up-sampling or interpolation techniques produce undesirable blurring and ringing artifacts. To alleviate this problem, a novel spatio-temporal pre-processing approach is proposed to improve the quality. The proposed method use low frequency DCT (Discrete cosine transform) component to sub-sample the frame at the transmitter side. In transmitter side a preprocessing method is proposed where the received subsampled frame is passed through a Wiener filter which uses its local statistics in 3×3 neighborhood to modify pixel values. The output of Wiener filter is added with optimized multiple of high frequency component. The output is then passed through a DCT block to up-sample. Result shows that the proposed method outperforms popularly used interpolation techniques in terms of quality measures
APA, Harvard, Vancouver, ISO, and other styles
35

Ho, Yu-Hsuan, and 何宥萱. "Key-Frame Extraction for Video Summarization and Shot-Based Video Retrieval." Thesis, 2004. http://ndltd.ncl.edu.tw/handle/73943136238044640470.

Full text
Abstract:
碩士
國立中正大學
資訊工程研究所
92
In this paper, we present an adaptive rate-constrained key-frame selection scheme for channel-aware realtime video streaming and shot-based video retrieval. First, the streaming server dynamically determines the target number of key-frames by estimating the channel conditions according to the feedback information. Under the constraint of the target key-frame number, a two-step sequential key-frame selection scheme is adopted to select the target number of key-frames by first finding the optimal allocation among the video shots in a video clip, and then selecting most representative key-frames in each shot according to the allocation to guide the temporal-downscaling transcoding. After extracting the key-frames, we propose a multi-pass video retrieval using spatio-temporal statistics information. In the first-pass, the probability distributions of object motion for each shot of the query video clip are extracted and then are compared with the probability of the shots in the database by using the Bhattacharyya distance. In the second-pass, two consecutive shots are employed to the introduction of the “causality” effect. Finally, in the refinement-pass, we extract one key-frame from each shot using our key-frame selection method, and calculate the color histogram of each key-frame. Then we use the Bhattacharya distance to compare the similarity of the two color histograms of key-frames and cumulate the second-stage distance to be the similarity of two video shots. Without respect to the two-step key-frame selection or multi-pass video retrieval, our experimental results show that the proposed methods are efficient and satisfactory.
APA, Harvard, Vancouver, ISO, and other styles
36

Liao, Wei-Hao, and 廖偉豪. "Dynamic Frame Rate Control in Video Transcoding for Multipoint Video Conferencing." Thesis, 2000. http://ndltd.ncl.edu.tw/handle/29195596208006916607.

Full text
Abstract:
碩士
國立清華大學
電機工程學系
88
Video conference is an Internet application which provides an environment for people to see and talk to other persons who are far away. A video transcoder is required in a multipoint video conference to convert input streams into different format, especially different bit-rate. Ordinary transcoder degrades temporal quality by setting a fixed frame-rate, but the requirement of motion smoothness changes with video properties. It’s necessary to adjust the frame-rate to a proper value to achieve better spatial and temporal quality. We propose a dynamic frame rate control scheme to adjust frame-rate to fit the property of video. By using frame-rate and motion level of object as measurement of temporal quality, spatial and temporal quality can be considered at the same time. When bit-rate decreases, frame-rate will be decreased to improve frame quality. When motion level increases, frame-rate will be increased to provide better continuity. Since the adaptive control method is used and all parameters are extracted from original transcoder processing, our proposed scheme is suitable for real-time applications.
APA, Harvard, Vancouver, ISO, and other styles
37

Chi, Haohan, and 紀皓瀚. "Key Frame Based Asymmetry Stereo Video Coding." Thesis, 2012. http://ndltd.ncl.edu.tw/handle/93130858612012561847.

Full text
Abstract:
碩士
國立中正大學
電機工程研究所
100
In recently years, because of the rise of the 3D video technology and movie market, more and more 3D products have drawn public’s attention, and one of the most popular product is Free-viewpoint TV (FTV), it allows viewers can change any view point they want while watching video, so that can get highly entertainment quality. However, FTV requires a lot of coding data rates to provide the good visual quality, and may increases the cost of the equipments. So, in this work, we propose the asymmetric coding scheme, by using the mix resolution formats at encoder, the data rates can effectively reduce, but to maintain the visual quality with fewer bit-rates, we proposed the super-resolution method to enhance the visual quality by using the image information which we already have. Results show the coding performance is improved by asymmetric coding scheme, and the quality of distortion resolution image is also enhanced with our super-resolution method.
APA, Harvard, Vancouver, ISO, and other styles
38

Chen, Siao-Wei, and 陳孝畏. "HVS-Based Frame Compatible Stereo Video Coding." Thesis, 2013. http://ndltd.ncl.edu.tw/handle/54809730288792290548.

Full text
Abstract:
碩士
國立中正大學
電機工程研究所
101
Recently, the advance in 3D display technology, enabling the depth perception, provides humans with better visual experiences. However, this technique is not yet widespread into our daily life due to the lack of 3D content, as well as the increased data for storage and transmission. To solve these problems, Motion Picture Expert Group (MPEG), Advanced Television Systems Committee (ATSC) and Digital Video Broadcasting (DVB) are currently standardizing the 3D video coding technique and related video broadcast specifications. Since the 3D Coding and transmission standards are not available nowadays, to facilitate the introduction of stereoscopic services through existing infrastructure and equipment, we need to transfer the 3D contents into another format, so-called frame-compatible stereo format, where the left and right view frames are combined into one single frame. Then we can use the existing broadcasting equipment systems to transmit such kind of stereo video. However, the reconstructed stereo video quality is degraded due to the reduced-resolution before encoding. Besides, the image quality depends on the packing type. In this work, a novel frame-compatible stereo video coding technique based on human visual system is proposed where the packing format is adaptively selected by considering video characteristics. Usually, uniform down-sampling is used to form the frame-compatible stereo video, we propose to sub-sample the stereo video in an unequal way, by taking the human visual system into account. The regions the people pay a higher attention will be less distorted during the down-sampling and up-sampling processes by assigned a down-sampling factor smaller the 2. On the other hand, the regions with minor interest will be assigned a higher down-sampling factor. The experimental results show that the proposed technique outperforms the existing techniques in subjective experiments. In the future, this work will be extended to the case of high quality full resolution stereo where how to maximize the inter-layer prediction between different image layers will be one of the research works.
APA, Harvard, Vancouver, ISO, and other styles
39

Chao, Hsiang-Ya, and 趙祥雅. "Video Anomaly Detection via Multi-frame Prediction." Thesis, 2019. http://ndltd.ncl.edu.tw/handle/sq8zdz.

Full text
Abstract:
碩士
國立臺灣大學
電信工程學研究所
107
Video anomaly detection which intents to identify rarely-happened or unexpected events is a worthy and developmental problem in video understanding tasks. Most of the previous works deal with the problem in an unsupervised way by learning normal representations of training data and identified the outliers as anomalies. Common deep learning-based methods are reconstruction-based. They train an autoencoder by minimizing the reconstruction errors of regular videos. Nevertheless, abnormal events don''t always lead to larger reconstruction errors. To address this issue, We propose using multi-frame prediction framework to enlarge the unexpected change and overcome the generalization property which stems from the use of an autoencoder. We use ConvLSTM model as the multi-frame predictor and show the effectivenes of utilizing latter frames for computing the frame anomaly scores. Experimental results show that our model leads to better performance on motion and appearance deformation irregularities. In addition, we collect a new car crash dataset which contains various car accidents as abnormal events from YouTube for evaluation. Compared to existing anomaly detection datasets, it is a more challenging and practical dataset due to the diversity of events and its different environmental conditions. Our model achieves comparable results in popular existing anomaly detection datasets and outperforms the state-of-the-art on the new proposed dataset.
APA, Harvard, Vancouver, ISO, and other styles
40

Sandeep, Manthi Venkat. "Full Frame Video Stabilization Using Motion Inpainting." Thesis, 2015. http://ethesis.nitrkl.ac.in/7776/1/2015_Full_Frame_Video_Sandeep.pdf.

Full text
Abstract:
The amount of video data has increased dramatically with the advent of digital imaging. Most of the video captured these days originates from a mobile phones and handheld video cameras. Such videos are shaky compared to videos that are shot with a tripod mounted camera. Stabilizing this video to remove the shaky effect using software is called Digital video stabilization which results in a stable and visually pleasant video. In order digitally stabilize the image, we need to (1) Estimate the motion of camera, (2) Regenerate the motion of camera without the undesirable artifacts and (3) Synthesize new video frames. This dissertation is targeted at improving the last two steps of stabilizing the video. Most of the previous techniques of video stabilization produce a lower resolution stabilized video output and clip portions of frames to remove the empty area formed by transformation of the video frames. We use a Gaussian averaging filter to smoother the global motion in the video. Then the frames are transformed using the new transformation matrices obtained by subtracting the original transformation chain from the modified transformation chain. For the last step of synthesizing new video frames, we introduce an improved completion technique which can produce full frame video by using the pixel information from nearby frames to estimate the intensity of the missing pixels. This technique uses motion inpainting to ensure that the video frames are filled in both the static image area and dynamic image area with the same consistency. Additionally, the quality of the video is improved by using a deblurring algorithm which further improves the smoothness of video by eliminating undesirable motion blur. We do not estimate the PSF, in its place, we transfer and interpolate the sharper pixels from nearby frames to improve the sharpness and deblur current frame. Completing the video with motion inpainting and deblurring technique allow us to construct a full frame video stabilization system with good image quality. This is verified by implementing the technique on different video sequences.
APA, Harvard, Vancouver, ISO, and other styles
41

Shen, Chyang, and 沈強. "MPEG-2 Video Enhancement for Interlaced Frame Sequences." Thesis, 2000. http://ndltd.ncl.edu.tw/handle/19871565762589651074.

Full text
Abstract:
碩士
國立交通大學
資訊工程系
88
Compared to MPEG-I video standard using frame picture as picture mode, MPEG-II video standard has adopted field picture as its alternate picture mode besides frame picture. Furthermore, progressive frame and interlaced frame are two kinds of frame pictures in MPEG-2 video standard. Interlaced frame combines with two field pictures at difference time instant. Frame picture is used in progressive monitor, and field picture is used in interlaced monitor; but sawtooth effect sometime will arise when display interlaced frame in progressive monitor. Our thesis proposes a new video enhancement method being used to reduce sawtooth effect due to displaying interlaced frame in progressive monitor. Our video enhancement method combine itself with MPEG-2 video decoding process. Also, our method utilizes temporal and spatial information of video decoding information to have lower computational complexity and higher visual quality.
APA, Harvard, Vancouver, ISO, and other styles
42

Lin, Vincent, and 林子平. "On Video Frame Rate and Sampling Rate Conversions." Thesis, 1995. http://ndltd.ncl.edu.tw/handle/65983245470786821496.

Full text
Abstract:
碩士
國立交通大學
電子研究所
83
In the thesis, we propose several efficient frame rate conversion algorithms which can improve image quality with low computational complexity. For frame rate conversion, we propose two algorithms, one is based on MCI algorithm with pixel-level motion vector preselection, the other is besed on block MCI. The first algorithm acieves a better image quality at a reduced computation load, while the second algorithm costs even less computation and still maintains a comparable visual quality. For sampling rate conversion, we propose a transform-domain based conversion algorithm, which does not have to design a strict and complicated filter. Compared with the time-domain interpolaion filtering approach, not only the algorithm gives better performance, but it may also reduce computation load in some cases. Finally, we implement a video output interface unit for MPEG2 decoder. Its function is to synchronize the output picture to the peripheral devices and position the picture on the screen, to interpolate chrominance samples, to convert the color space from YCrCb to RGB and the aspect ratio from 4:3 to 16:9.
APA, Harvard, Vancouver, ISO, and other styles
43

Liang, Yang, and 梁洋. "Frame Rate Control for Constant Bit Rate Video." Thesis, 2003. http://ndltd.ncl.edu.tw/handle/33317929075882486255.

Full text
Abstract:
碩士
國立臺北科技大學
電機工程系碩士班
91
This thesis proposed a variable frame skip (VFS) mechanism to improve the video quality in low bit rate channel. The basic idea of VFS mechanism is to decide and skip a suitable, non-fixed number of frames in temporal domain to reduce bit usage. The saved bits can be allocated to enhance the spatial quality of the video. In literature, several methods of frame skip decision have proposed, but most of them only consider the similarity between neighboring coded frames as the decision criterion. Our proposed method takes into account the reconstruction of the skipped frames using motion-compensated frame interpolation at decoder. The proposed VFS models the reconstructed quality of the skipped frame and, therefore, can provide a fast estimate to the frame skipped at encoder. The proposed VFS can decide the frame skip in real time, and its encoded video has better spatial-temporal bit allocation.
APA, Harvard, Vancouver, ISO, and other styles
44

"Multiplexing video traffic using frame-skipping aggregation technique." 1998. http://library.cuhk.edu.hk/record=b5889564.

Full text
Abstract:
by Alan Yeung.
Thesis (M.Phil.)--Chinese University of Hong Kong, 1998.
Includes bibliographical references (leaves 53-[56]).
Abstract also in Chinese.
Chapter 1 --- Introduction --- p.1
Chapter 2 --- MPEG Overview --- p.5
Chapter 3 --- Framework of Frame-Skipping Lossy Aggregation --- p.10
Chapter 3.1 --- Video Frames Delivery using Round-Robin Scheduling --- p.10
Chapter 3.2 --- Underflow Safety Margin on Receiver Buffers --- p.12
Chapter 3.3 --- Algorithm in Frame-Skipping Aggregation Controller --- p.13
Chapter 4 --- Replacement of Skipped Frames in MPEG Sequence --- p.17
Chapter 5 --- Subjective Assessment Test on Frame-Skipped Video --- p.21
Chapter 5.1 --- Test Settings and Material --- p.22
Chapter 5.2 --- Choice of Test Methods --- p.23
Chapter 5.3 --- Test Procedures --- p.25
Chapter 5.4 --- Test Results --- p.26
Chapter 6 --- Performance Study --- p.29
Chapter 6.1 --- Experiment 1: Number of Supportable Streams --- p.31
Chapter 6.2 --- Experiment 2: Frame-Skipping Rate When Multiplexing on a Leased T3 Link --- p.33
Chapter 6.3 --- Experiment 3: Bandwidth Usage --- p.35
Chapter 6.4 --- Experiment 4: Optimal USMT --- p.38
Chapter 7 --- Implementation Considerations --- p.41
Chapter 8 --- Conclusions --- p.45
Chapter A --- The Construction of Stuffed Artificial B Frame --- p.48
Bibliography --- p.53
APA, Harvard, Vancouver, ISO, and other styles
45

Haung, Cheng-wei, and 黃正維. "Optimal Tiling Algorithms for Intra-frame Video Coding." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/9wzh8u.

Full text
Abstract:
碩士
國立臺灣科技大學
資訊工程系
99
Algorithms for best basis search in tree-structured dictionaries have been effectively used for many signal processing problems, including noise removal, image compression, audio compression, and image segmentation. An important class of best basis algorithms are methods that search for an optimal rectangular tiling of an image. Such methods can significantly improve the performance of image coding strategies based on block or lapped transforms through adapting the sizes and shapes of transform blocks to the structure of an image. H.264/AVC employs variable-block-size for motion compensation and transform stages that can significantly improve the coding performance compared with previous video coding standards. In this paper, we propose the use of large dictionaries of tilings in order to further improve the coding efficiency of H.264/AVC intraframe.We designed a fast search algorithm to select the optimal tiling for the transform stage of a video coder. In addition, we will have implemented an efficient arithmetic coding algorithms to encode the selected tiling. The experimental results show that the proposed method of intraframe coding results in up to 15% savings in bit rate as compared to H.264/AVC.
APA, Harvard, Vancouver, ISO, and other styles
46

Lin, Jung-Shan, and 林蓉珊. "Full-Frame Video Stabilization by Considering Capturing Intention." Thesis, 2007. http://ndltd.ncl.edu.tw/handle/81707872936691623184.

Full text
Abstract:
碩士
國立臺灣大學
資訊工程學研究所
95
Annoying shaky motion is one of the significant problems in home videos, since hand shake is an unavoidable effect when capturing by using a hand-held camcorder. Video stabilization is an important technique to solve this problem. However, the stabilized videos resulted by current methods usually have decreased resolution and are still not so stable. In this thesis, we propose a novel, robust, and practical method of video stabilization while considering users'' capturing intention. Our method can produce full-frame stabilized videos, and not only the high frequency shaky motions but also the low frequency unexpected movements are removed. To guess the user''s capturing intention, we first consider the regions of interest in the video to estimate which regions or objects the user wants to capture, and then use a polyline to estimate a new stable camcorder motion path while avoiding the user''s interested regions being cut out. Then, we fill the dynamic and static missing areas caused by frame alignment from other frames to keep the same resolution and quality as the original video. Furthermore, we smooth the discontinuous regions by using a three-dimensional Poisson-based method. After the above automatic operations, a full-frame stabilized video can be achieved and the important regions can also be preserved.
APA, Harvard, Vancouver, ISO, and other styles
47

Tsai, Han-Yu, and 蔡涵宇. "Unequal Frame Protection Policy in Video Streaming System." Thesis, 2011. http://ndltd.ncl.edu.tw/handle/39466701526336196411.

Full text
Abstract:
碩士
國立雲林科技大學
資訊工程研究所
99
In recent decades, with the development of Internet, video streaming on the Internet have widely grown, such as high-definition video on demand, surveillance monitoring, and video conferencing. These multimedia applications change our lifestyle. However, packet loss in Internet is still an unavoidable problem. In order to meet the constraint of maximum transfer unit (MTU) in network transmission, each frame is fragmented into multiple correlated packets during transmission. In video streaming, the packets fragmented from video frames may incur packet error or packet loss. Therefore, in the IP-based video streaming, once an IP packet incurs errors, the packet will be discard by the IP protocol in the receiver side. Thus the associated video frame incurs frame error which impacts the playback quality of video streaming. Additionally, each frame in the video streaming service is associated with its previous frames. Losing an important frame may cause a set of frames loss or damage as well. To mitigate the impact of the packet loss, numbers of schemes have been published in the literature. One of major frame protection schemes is UEP (unequal video frame protection) which considers frame important level to provide different frame protection level for providing high playback quality in limited resource condition. Thus, how to precisely determine the frame important level and how to base on the frame important level to allocate suitable resources to maximum the playback quality are two key issues in the UEP scheme. In the literatures, most of publishes use error propagation level (EPL) to decide the frame important level. The frame important level in EPL only considers the location of frame in GOP (group of picture). In EPL, the frame in more front of GOP has more important level. In our opinion, only consider the EPL value is difficult to identify the frame important value. The video playback distortion is not only affected by the dependent frames, but also affected by the variation of lost frame. Herein, the frame variation is defined as the difference between the lost frame and prior frame of lost frame. If a frame with large frame variation is lost, the simple frame copy will get large distortion in frame playback. On the other hand, it will get good playback quality when the lost frame is much similar to its previous frame. Based on these concepts, in this thesis, we propose different PSNR-based important frame decision schemes and based on the frame important level an optimal UEP scheme to find the optimal network resource assignment for each frame for maximizing the playback quality in client. The simulation results reveal that the proposed resource allocation policy with nonlinear important frame decision scheme has great performance in term of playback quality than other video frame protection schemes.
APA, Harvard, Vancouver, ISO, and other styles
48

Huang, Wei-Ting, and 黃惟婷. "Full-Frame Video Stabilization with Large Moving Object." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/27680320781860145678.

Full text
Abstract:
碩士
國立臺灣大學
資訊工程學研究所
96
This thesis presents an approach to post-processing casually captured videos to improve apparent camera movement. A lot of home videos have some problems about the artifacts like hand shaking when capturing without tripod. Video stabilization is an important technique to solve this problem. However, the technique does not work very well in some situations such as the larger foreground, incomplete background or other situations,etc. In this thesis,we propose a novel method of video stabilization to overcome the situations with the larger foreground, or the feature points amount of the background is less then of foreground and some camera motion such as zoom in and zoom out. This system applies a method, optical flow, to estimate the motion vector of all pixels between each pair frames. Then, we use K-means clustering to group the similar motion vectors of each frame. To select an adequate segment to estimate the global camera path of video could obtain one more correct global camera path. After motion vector segmentation and camera path estimation, we could stitch all of video frames to get a panorama and estimate the range of moving object we can recover by neighbor frames. Based on the background panorama and moving object recovered range, we could find some new paths which would lose information less. After the above operations, a full-stabilized video could be achieved.
APA, Harvard, Vancouver, ISO, and other styles
49

Wu, Jian-Rong, and 吳健榮. "Asymmetric Frame Compatible Depth Video Coding Considering Occulusion." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/64585478527320328212.

Full text
Abstract:
碩士
國立中正大學
電機工程研究所
103
Stereo video offers favorable reception in 3D entertainment. It supplies more immersive perception for audience than conventional video does. In order to expand stereo video application scenarios, depth video is used to render virtual view in conjunction with texture video. It not only provides much fun for watching video, but supports next generation TV, FTV (free-view TV). To reduce the stereo data and also reuse the existing infrastructure and equipment for 2D video, a frame-compatible stereo video format is usually used. For the frame-compatible stereo coding, only stereo texture video is considered. In this work, we propose frame-compatible stereo depth coding. Depth map plays a important role in view synthesis and the reliability of depth map is very important in rendering good quality virtual view. Hence, how to perform the downsampling while maintaining the preciseness of depth video is the challenge in this work. We propose an asymmetric frame-compatible depth video coding considering occulusion region. The content in both views is analyzed after performing view warping from the primary view to the secondary view. The occlusion region is defined as the missing region where no corresponding information can be founded in the primary view. The idea is to avoid the sending of the same information in both views. Thus, the downsized primary view and the (downsized) occlusion region in the secondary view is packed into a frame. Then the frame-compatible stereo depth can be encoded by video coding standards. The experiment results show that our proposed technique achieves better coding performance in terms of objective and subjective ways, compared to conventional frame-compatible technology. In the future, we will extend this technique to three views case. It will provide expanded view angle for audiences
APA, Harvard, Vancouver, ISO, and other styles
50

Lin, Jung-Shan. "Full-Frame Video Stabilization by Considering Capturing Intention." 2007. http://www.cetd.com.tw/ec/thesisdetail.aspx?etdun=U0001-1807200719054900.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography