To see the other types of publications on this topic, follow the link: Digital video – Data processing.

Dissertations / Theses on the topic 'Digital video – Data processing'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Digital video – Data processing.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Altilar, Deniz Turgay. "Data partitioning and scheduling for parallel digital video processing." Thesis, Queen Mary, University of London, 2002. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.399730.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Gu, Lifang. "Video analysis in MPEG compressed domain." University of Western Australia. School of Computer Science and Software Engineering, 2003. http://theses.library.uwa.edu.au/adt-WU2003.0016.

Full text
Abstract:
The amount of digital video has been increasing dramatically due to the technology advances in video capturing, storage, and compression. The usefulness of vast repositories of digital information is limited by the effectiveness of the access methods, as shown by the Web explosion. The key issues in addressing the access methods are those of content description and of information space navigation. While textual documents in digital form are somewhat self-describing (i.e., they provide explicit indices, such as words and sentences that can be directly used to categorise and access them), digital video does not provide such an explicit content description. In order to access video material in an effective way, without looking at the material in its entirety, it is therefore necessary to analyse and annotate video sequences, and provide an explicit content description targeted to the user needs. Digital video is a very rich medium, and the characteristics in which users may be interested are quite diverse, ranging from the structure of the video to the identity of the people who appear in it, their movements and dialogues and the accompanying music and audio effects. Indexing digital video, based on its content, can be carried out at several levels of abstraction, beginning with indices like the video program name and name of subject, to much lower level aspects of video like the location of edits and motion properties of video. Manual video indexing requires the sequential examination of the entire video clip. This is a time-consuming, subjective, and expensive process. As a result, there is an urgent need for tools to automate the indexing process. In response to such needs, various video analysis techniques from the research fields of image processing and computer vision have been proposed to parse, index and annotate the massive amount of digital video data. However, most of these video analysis techniques have been developed for uncompressed video. Since most video data are stored in compressed formats for efficiency of storage and transmission, it is necessary to perform decompression on compressed video before such analysis techniques can be applied. Two consequences of having to first decompress before processing are incurring computation time for decompression and requiring extra auxiliary storage.To save on the computational cost of decompression and lower the overall size of the data which must be processed, this study attempts to make use of features available in compressed video data and proposes several video processing techniques operating directly on compressed video data. Specifically, techniques of processing MPEG-1 and MPEG-2 compressed data have been developed to help automate the video indexing process. This includes the tasks of video segmentation (shot boundary detection), camera motion characterisation, and highlights extraction (detection of skin-colour regions, text regions, moving objects and replays) in MPEG compressed video sequences. The approach of performing analysis on the compressed data has the advantages of dealing with a much reduced data size and is therefore suitable for computationally-intensive low-level operations. Experimental results show that most analysis tasks for video indexing can be carried out efficiently in the compressed domain. Once intermediate results, which are dramatically reduced in size, are obtained from the compressed domain analysis, partial decompression can be applied to enable high resolution processing to extract high level semantic information.
APA, Harvard, Vancouver, ISO, and other styles
3

Wang, Yang, and 王揚. "Digital video segmentation and annotation in news programs." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2001. http://hub.hku.hk/bib/B31225305.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Chen, Wu-hsiang Jonas. "Error-resilient techniques for packet video communications /." Thesis, Connect to this title online; UW restricted, 1999. http://hdl.handle.net/1773/5850.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Bayrakeri, Sadik. "Scalable video coding using spatio-temporal interpolation." Diss., Georgia Institute of Technology, 1997. http://hdl.handle.net/1853/15385.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Wong, Chi Wah. "Studying real-time rate control in perceptual, modeling and efficient aspects /." View abstract or full-text, 2004. http://library.ust.hk/cgi/db/thesis.pl?ELEC%202004%20WONGC.

Full text
Abstract:
Thesis (M. Phil.)--Hong Kong University of Science and Technology, 2004.
Includes bibliographical references (leaves 205-212). Also available in electronic version. Access restricted to campus users.
APA, Harvard, Vancouver, ISO, and other styles
7

Lanciani, Christopher A. "Compressed-domain processing of MPEG audio signals." Diss., Georgia Institute of Technology, 1999. http://hdl.handle.net/1853/13760.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Guo, Liwei. "Restoration and modeling for multimedia compression /." View abstract or full-text, 2008. http://library.ust.hk/cgi/db/thesis.pl?ECED%202008%20GUOL.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Ye, Mang. "Open-world person re-identification." HKBU Institutional Repository, 2019. https://repository.hkbu.edu.hk/etd_oa/688.

Full text
Abstract:
With the increasing demand of intelligent video surveillance systems, person re-identification (re-ID) plays an important role in intelligent video analysis, which aims at matching person images across non-overlapping camera views. It has gained increasing attention in computer vision community. With the advanced deep neural networks, existing methods have achieved promising performance on the widely-used re-ID benchmarks, even outperform the human-level rank-1 matching accuracy. However, most of the research efforts are conducted on the closed-world settings, with large-scale well annotated training data and all the person images are from the same visible modality. As a prerequisite in practical video surveillance application, there is still a large gap between the closed-world research-oriented setting and the practical open-world settings. In this thesis, we try to narrow the gap by studying three important issues in open-world person re-identification, including 1) unsupervised learning with large-scale unlabelled training data; 2) learning robust re-ID model with label corrupted training data and 3) cross-modality visible-thermal person re-identification with multi-modality data. For unsupervised learning with unlabelled training data, we mainly focus on video-based person re-identification, since the video data is usually easily obtained by tracking algorithms and the video sequence provides rich weakly labelled samples by assuming the image frames within the tracked sequence belonging to the same person identity. Following the cross-camera label estimation approach, we formulate the cross-camera label estimation as a one-to-one graph matching problem, and then propose a novel dynamic graph matching framework to estimate cross-camera labels. However, in a practical wild scenario, the unlabelled training data usually cannot satisfy the one-to-one matching constraint, which would result in a large proportion of false positives. To address this issue, we further propose a novel robust anchor embedding method for unsupervised video re-ID. In the proposed method, some anchor sequences are firstly selected to initialize the CNN feature representation. Then a robust anchor embedding method is proposed to measure the relationship between the unlabelled sequences and anchor sequences, which considers both the scalability and efficiency. After that, a top-{dollar}k{dollar} counts label prediction strategy is proposed to predict the labels of unlabelled sequences. With the newly estimated sequences, the CNN representation could be further updated. For robust re-ID model learning with label corrupted training data, we propose a two-stage learning method to handle the label noise. Rather than simply filtering the falsely annotated samples, we propose a joint learning method by simultaneously refining the falsely annotated labels and optimizing the neural networks. To address the limited training samples for each identity, we further propose a novel hard-aware instance re-weighting strategy to fine-tune the learned model, which assigns larger weights to hard samples with correct labels. For cross-modality visible-thermal person re-identification, it addresses an important issue in night-time surveillance applications by matching person images across different modalities. We propose a dual-path network to learn the cross-modality feature representations, which learns the multi-modality sharable feature representations by simultaneously considering the modality discrepancy and commonness. To guide the feature representation learning process, we propose a dual-constrained top-ranking loss, which contains both cross-modality and intra-modality top-ranking constraints to reduce the large cross-modality and intra-modality variations. Besides the open-world person re-identification, we have also studied the unsupervised embedding learning problem for general image classification and retrieval. Motivated by supervised embedding learning, we propose a data augmentation invariant and instance spread-out feature. To learn the feature embedding, we propose a instance feature-based softmax embedding, which optimizes the embedding directly on top of the real-time instance features. It achieves much faster learning speed and better accuracy than existing methods. In short, the major contributions of this thesis are summarized as follows. l A dynamic graph matching framework is proposed to estimate cross-camera labels for unsupervised video-based person re-identification. l A robust anchor embedding method with top-{dollar}k{dollar} counts label prediction is proposed to efficiently estimate the cross-camera labels for unsupervised video-based person re-identification under wild settings. l A two-stage PurifyNet is introduced to handle the label noise problem in person re-identification, which jointly refines the falsely annotated labels and mines hard samples with correct labels. l A dual-constrained top-ranking loss with a dual-path network is proposed for cross-modality visible-thermal person re-identification, which simultaneously addresses the cross-modality and intra-modality variations. l A data augmentation invariant and instance spread-out feature is proposed for unsupervised embedding learning, which directly optimizes the learned embedding on top of real-time instance features with softmax function
APA, Harvard, Vancouver, ISO, and other styles
10

Lee, Yen-Chi. "Error resilient video streaming over lossy networks." Diss., Available online, Georgia Institute of Technology, 2004:, 2003. http://etd.gatech.edu/theses/available/etd-04082004-180302/unrestricted/lee%5fyen-chi%5f200312%5fphd.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
11

Armenio, Christopher. "A communication framework for distributed computer vision on stationary and mobile platforms /." Online version of thesis, 2009. http://hdl.handle.net/1850/10753.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Wise, Richard J. Jr. "A DESIGN FOR A 10.4 GIGABIT/SECOND SOLID-STATE DATA RECORDER." International Foundation for Telemetering, 1999. http://hdl.handle.net/10150/607335.

Full text
Abstract:
International Telemetering Conference Proceedings / October 25-28, 1999 / Riviera Hotel and Convention Center, Las Vegas, Nevada
A need has been identified in the Test and Evaluation (T&E) and tactical aircraft communities for a ruggedized high-speed instrumentation data recorder to complement the ever-increasing number of high frame-rate digital cameras and sensors. High-speed digital camera manufacturers are entering this market in order to provide adequate recording capability for their own cameras. This paper discusses a Solid-State Data Recorder (SSDR) for use in Imaging and High-Speed Sensor Data Aquisition applications. The SSDR is capable of a 10.4 Gb/sec sustained, 16Gb/sec burst, input data rate via a proprietary 32-channel-by-10-bit generic high-speed parallel interface, a massively-parallel 256-bit bus architecture, and unique memory packaging design. A 32-bit PCIbus control/archive and dedicated DCRsi™ interface are also employed, allowing data archiving to standard high-speed interfaces (SCSI, Fiber-Channel, USB, etc.) and DCRsi™-compatible tape recorders.
APA, Harvard, Vancouver, ISO, and other styles
13

Silva, Fernando Silvestre da. "Procedimentos para tratamento e compressão de imagens e video utilizando tecnologia fractal e transformadas wavelet." [s.n.], 2005. http://repositorio.unicamp.br/jspui/handle/REPOSIP/260581.

Full text
Abstract:
Orientador: Yuzo Iano
Tese (doutorado) - Universidade Estadual de Campinas, Faculdade de Engenharia Eletrica e de Computação
Made available in DSpace on 2018-08-05T13:46:30Z (GMT). No. of bitstreams: 1 Silva_FernandoSilvestreda_D.pdf: 35017484 bytes, checksum: fb460a6a42e44fe0a50e94599ac027fc (MD5) Previous issue date: 2005
Resumo: A excelente qualidade visual e taxa de compressão dos codificadores fractais de imagem tem aplicações limitadas devido ao exaustivo tempo de codificação inerente. Esta pesquisa apresenta um novo codificador híbrido de imagens que aplica a velocidade da transformada wavelet à qualidade da compressão fractal. Neste esquema, uma codificação fractal acelerada usando a classificação de domínios de Fisher é aplicada na sub-banda passa-baixas de uma imagem transformada por wavelet e uma versão modificada do SPIHT (Set Partitioned in Hierarchical Trees) é aplicada nos coeficientes remanescentes. Os detalhes de imagem e as características de transmissão progressiva da transformada wavelet são mantidas; nenhum efeito de bloco comuns às técnicas fractais é introduzido, e o problema de fidelidade de codificação comuns aos codificadores híbridos fractal-wavelet é resolvido. O sistema proposto reduz o tempo médio de processamento das imagens em 94% comparado com o codificador fractal tradicional e um ganho de 0 a 2,4dB em PSNR sobre o algoritmo SPIHT puro. Em ambos os casos, o novo esquema proposto apresentou melhorias em temos de qualidade subjetiva das imagens para altas, médias e baixas taxas de compressão
Abstract: The excellent visual quality and compression rate of fractal image coding have limited applications due to exhaustive inherent encoding time. This research presents a new fast and efficient image coder that applies the speed of the wavelet transform to the image quality of the fractal compression. In this scheme, a fast fractal encoding using Fisher¿s domain classification is applied to the lowpass subband of wavelet transformed image and a modified SPIHT coding (Set Partitioning in Hierarchical Trees), on the remaining coefficients. The image details and wavelet progressive transmission characteristics are maintained; no blocking effects from fractal techniques are introduced; and the encoding fidelity problem common in fractal-wavelet hybrid coders is solved. The proposed scheme provides an average of 94% reduction in encoding-decoding time compared to the pure accelerated Fractal coding results, and a 0-2,4dB gain in PSNR over the pure SPIHT coding. In both cases, the new scheme improves the subjective quality of pictures for high, medium and low bit rates
Doutorado
Telecomunicações e Telemática
Doutor em Engenharia Elétrica
APA, Harvard, Vancouver, ISO, and other styles
14

Hanson, Anthony. "General Purpose Computing in Gpu - a Watermarking Case Study." Thesis, University of North Texas, 2014. https://digital.library.unt.edu/ark:/67531/metadc700078/.

Full text
Abstract:
The purpose of this project is to explore the GPU for general purpose computing. The GPU is a massively parallel computing device that has a high-throughput, exhibits high arithmetic intensity, has a large market presence, and with the increasing computation power being added to it each year through innovations, the GPU is a perfect candidate to complement the CPU in performing computations. The GPU follows the single instruction multiple data (SIMD) model for applying operations on its data. This model allows the GPU to be very useful for assisting the CPU in performing computations on data that is highly parallel in nature. The compute unified device architecture (CUDA) is a parallel computing and programming platform for NVIDIA GPUs. The main focus of this project is to show the power, speed, and performance of a CUDA-enabled GPU for digital video watermark insertion in the H.264 video compression domain. Digital video watermarking in general is a highly computationally intensive process that is strongly dependent on the video compression format in place. The H.264/MPEG-4 AVC video compression format has high compression efficiency at the expense of having high computational complexity and leaving little room for an imperceptible watermark to be inserted. Employing a human visual model to limit distortion and degradation of visual quality introduced by the watermark is a good choice for designing a video watermarking algorithm though this does introduce more computational complexity to the algorithm. Research is being conducted into how the CPU-GPU execution of the digital watermark application can boost the speed of the applications several times compared to running the application on a standalone CPU using NVIDIA visual profiler to optimize the application.
APA, Harvard, Vancouver, ISO, and other styles
15

Smith, Quentin D. "AN EVOLUTIONARY APPROACHTO A COMMUNICATIONS INFRASTRUCTURE FOR INTEGRATED VOICE, VIDEO AND HIGH SPEED DATA FROM RANGETO DESKTOP USING ATM." International Foundation for Telemetering, 1993. http://hdl.handle.net/10150/608864.

Full text
Abstract:
International Telemetering Conference Proceedings / October 25-28, 1993 / Riviera Hotel and Convention Center, Las Vegas, Nevada
As technology progresses we are faced with ever increasing volumes and rates of raw and processed telemetry data along with digitized high resolution video and the less demanding areas of video conferencing, voice communications and general LAN-based data communications. The distribution of all this data has traditionally been accomplished by solutions designed to each particular data type. With the advent of Asynchronous Transfer Modes or ATM, a single technology now exists for providing an integrated solution to distributing these diverse data types. This allows an integrated set of switches, transmission equipment and fiber optics to provide multi-session connection speeds of 622 Megabits per second. ATM allows for the integration of many of the most widely used and emerging low, medium and high speed communications standards. These include SONET, FDDI, Broadband ISDN, Cell Relay, DS-3, Token Ring and Ethernet LANs. However, ATM is also very well suited to handle unique data formats and speeds, as is often the case with telemetry data. Additionally, ATM is the only data communications technology in recent times to be embraced by both the computer and telecommunications industries. Thus, ATM is a single solution for connectivity within a test center, across a test range, or between ranges. ATM can be implemented in an evolutionary manner as the needs develop. This means the rate of capital investment can be gradual and older technologies can be replaced slowly as they become the communications bottlenecks. However, success of this evolution requires some planning now. This paper provides an overview of ATM, its application to test ranges and telemetry distribution. A road map is laid out which can guide the evolutionary changeover from today's technologies to a full ATM communications infrastructure. Special applications such as the support of high performance multimedia workstations are presented.
APA, Harvard, Vancouver, ISO, and other styles
16

Chen, Howard. "AZIP, audio compression system: Research on audio compression, comparison of psychoacoustic principles and genetic algorithms." CSUSB ScholarWorks, 2005. https://scholarworks.lib.csusb.edu/etd-project/2617.

Full text
Abstract:
The purpose of this project is to investigate the differences between psychoacoustic principles and genetic algorithms (GA0). These will be discussed separately. The review will also compare the compression ratio and the quality of the decompressed files decoded by these two methods.
APA, Harvard, Vancouver, ISO, and other styles
17

Vila, Duran Marius. "Information theory techniques for multimedia data classification and retrieval." Doctoral thesis, Universitat de Girona, 2015. http://hdl.handle.net/10803/302664.

Full text
Abstract:
We are in the information age where most data is stored in digital format. Thus, the management of digital documents and videos requires the development of efficient techniques for automatic analysis. Among them, capturing the similarity or dissimilarity between different document images or video frames are extremely important. In this thesis, we first analyze for several image resolutions the behavior of three different families of image-based similarity measures applied to invoice classification. In these three set of measures, the computation of the similarity between two images is based, respectively, on intensity differences, mutual information, and normalized compression distance. As the best results are obtained with mutual information-based measures, we proceed to investigate the application of three different Tsallis-based generalizations of mutual information for different entropic indexes. These three generalizations derive respectively from the Kullback-Leibler distance, the difference between entropy and conditional entropy, and the Jensen-Shannon divergence. In relation to digital video processing, we propose two different information-theoretic approaches based, respectively, on Tsallis mutual information and Jensen-Tsallis divergence to detect the abrupt shot boundaries of a video sequence and to select the most representative keyframe of each shot. Finally, Shannon entropy has been commonly used to quantify the image informativeness. The main drawback of this measure is that it does not take into account the spatial distribution of pixels. In this thesis, we analyze four information-theoretic measures that overcome this limitation. Three of them (entropy rate, excess entropy, and erasure entropy) consider the image as a stationary stochastic process, while the fourth (partitional information) is based on an information channel between image regions and histogram bins
Ens trobem a l’era de la informació on la majoria de les dades s’emmagatzemen en format digital. Per tant, la gestió de documents i vídeos digitals requereix el desenvolupament de tècniques eficients per a l’anàlisi automàtic. Entre elles, la captura de la similitud o dissimilitud entre diferents imatges de documents o fotogrames de vídeo és extremadament important. En aquesta tesi, analitzem, a diverses resolucions d’imatge, el comportament de tres famílies diferents de mesures basades en similitud d’imatges i aplicades a la classificació de factures. En aquests tres conjunt de mesures, el càlcul de la similitud entre dues imatges es basa, respectivament, en les diferències d’intensitat, en la informació mútua, i en la distància de compressió normalitzada. Degut a que els millors resultats s’obtenen amb les mesures basades en la informació mútua, es procedeix a investigar l’aplicació de tres generalitzacions de la informació mútua basades en Tsallis en diferents índexs entròpics. Aquestes tres generalitzacions es deriven respectivament de la distància de Kullback-Leibler, la diferència entre l’entropia i entropia condicional, i la divergència de Jensen-Shannon. En relació al processament de vídeo digital, proposem dos enfocaments diferents de teoria de la informació basats respectivament en la informació mútua de Tsallis i en la divergència de Jensen-Tsallis, per detectar els límits d’un pla cinematogràfic en una seqüència de vídeo i per seleccionar el fotograma clau més representatiu de cada pla. Finalment, l’entropia de Shannon s’ha utilitzat habitualment per quantificar la informativitat d’una imatge. El principal inconvenient d’aquesta mesura és que no té en compte la distribució espacial dels píxels. En aquesta tesi, s’analitzen quatre mesures de teoria de la informació que superen aquesta limitació. Tres d’elles (entropy rate, excess entropy i erasure entropy) consideren la imatge com un procés estocàstic estacionari, mentre que la quarta (partitional information) es basa en un canal d’informació entre les regions d’una imatge i els intervals de l’histograma
APA, Harvard, Vancouver, ISO, and other styles
18

Стригун, В. В. "Програмне забезпечення для обробки та відображення інформації про параметри руху, отриманої від засобів відеореєстрації." Thesis, Чернігів, 2021. http://ir.stu.cn.ua/123456789/24987.

Full text
Abstract:
Стригун, В. В. Програмне забезпечення для обробки та відображення інформації про параметри руху, отриманої від засобів відеореєстрації : випускна кваліфікаційна робота : 123 "Комп’ютерна інженерія" / В. В. Стригун ; керівник роботи Є. В. Риндич ; НУ "Чернігівська політехніка", кафедра інформаційних і комп’ютерних систем. – Чернігів, 2021. – 211 с.
Об’єкт дослідження – програмний засіб для обробки інформації отриманої від засобів відеореєстрації та визначення характеристик положення об’єкту в просторі. Мета роботи – підвищення ефективності процесу визначення траєкторних параметрів рухомих об'єктів. Методи дослідження – системний аналіз та синтез організаційно-технічних систем, теорія ефективності та оптимізації, моделювання складних технічних систем. Основні результати, отримані в роботі: проведено аналіз існуючих та перспективних програмних засобів, призначених для обробки інформації, що отримана від засобів відеореєстрації; розроблена загальна структура програмного засобу та алгоритми роботи окремих модулів; розроблено спеціалізоване програмне забезпечення для обробки та відображення інформації, отриманої від засобів відеореєстрації; проведено попередні та приймальні випробування спеціалізованого програмного забезпечення для обробки та відображення інформації, отриманої від засобів відеореєстрації; відпрацьовано технічну документацію на програмний засіб. Результати роботи можуть бути використані для прискорення обробки даних та відображення інформації, отриманої від засобів відеореєстрації під час проведення випробувань. Для реалізації запропонованого сервісу буде використовуватися мова програмування Java та інтегроване середовище розробки IntelliJ IDEA.
Object of research - a software tool for processing information obtained from video recording and determining the characteristics of the position of the object in space. The purpose of the work is to increase the efficiency of the process of determining the trajectory parameters of moving objects. Research methods - system analysis and synthesis of organizational and technical systems, theory of efficiency and optimization, modeling of complex technical systems. The main results obtained in the work: the analysis of the existing and perspective software means intended for processing of the information received from means of video registration is carried out; the general structure of software and algorithms of work of separate modules is developed; developed specialized software for processing and displaying information received from video recording equipment; preliminary and acceptance tests of the specialized software for processing and display of the information received from means of video registration are carried out; the technical documentation for the software has been worked out. The results of the work can be used to speed up data processing and display information obtained from video recording equipment during testing. The Java programming language and the IntelliJ IDEA integrated development environment will be used to implement the proposed service.
APA, Harvard, Vancouver, ISO, and other styles
19

Pinto, Allan da Silva 1984. "A countermeasure method for video-based face spoofing attacks : Detecção de tentativas de ataque com vídeos digitais em sistemas de biometria de face." [s.n.], 2013. http://repositorio.unicamp.br/jspui/handle/REPOSIP/275616.

Full text
Abstract:
Orientador: Anderson de Rezende Rocha
Dissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Computação
Made available in DSpace on 2018-08-23T22:22:57Z (GMT). No. of bitstreams: 1 Pinto_AllandaSilva_M.pdf: 47523880 bytes, checksum: 072eb0490c26631b80cdcc47d55a4817 (MD5) Previous issue date: 2013
Resumo: O resumo poderá ser visualizado no texto completo da tese digital
Abstract: The complete abstract is available with the full electronic document
Mestrado
Ciência da Computação
Mestre em Ciência da Computação
APA, Harvard, Vancouver, ISO, and other styles
20

Jiang, Xiaofeng. "Multipoint digital video communications." Thesis, University of Cambridge, 1991. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.239548.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Tsoi, Yau Chat. "Video cosmetics : digital removal of blemishes from video /." View Abstract or Full-Text, 2003. http://library.ust.hk/cgi/db/thesis.pl?COMP%202003%20TSOI.

Full text
Abstract:
Thesis (M. Phil.)--Hong Kong University of Science and Technology, 2003.
Includes bibliographical references (leaves 83-86). Also available in electronic version. Access restricted to campus users.
APA, Harvard, Vancouver, ISO, and other styles
22

Chen, Juan. "Content-based Digital Video Processing. Digital Videos Segmentation, Retrieval and Interpretation." Thesis, University of Bradford, 2009. http://hdl.handle.net/10454/4256.

Full text
Abstract:
Recent research approaches in semantics based video content analysis require shot boundary detection as the first step to divide video sequences into sections. Furthermore, with the advances in networking and computing capability, efficient retrieval of multimedia data has become an important issue. Content-based retrieval technologies have been widely implemented to protect intellectual property rights (IPR). In addition, automatic recognition of highlights from videos is a fundamental and challenging problem for content-based indexing and retrieval applications. In this thesis, a paradigm is proposed to segment, retrieve and interpret digital videos. Five algorithms are presented to solve the video segmentation task. Firstly, a simple shot cut detection algorithm is designed for real-time implementation. Secondly, a systematic method is proposed for shot detection using content-based rules and FSM (finite state machine). Thirdly, the shot detection is implemented using local and global indicators. Fourthly, a context awareness approach is proposed to detect shot boundaries. Fifthly, a fuzzy logic method is implemented for shot detection. Furthermore, a novel analysis approach is presented for the detection of video copies. It is robust to complicated distortions and capable of locating the copy of segments inside original videos. Then, iv objects and events are extracted from MPEG Sequences for Video Highlights Indexing and Retrieval. Finally, a human fighting detection algorithm is proposed for movie annotation.
APA, Harvard, Vancouver, ISO, and other styles
23

Biswas, Mainak. "Content adaptive video processing algorithms for digital TV /." Diss., Connect to a 24 p. preview or request complete full text in PDF format. Access restricted to UC campuses, 2005. http://wwwlib.umi.com/cr/ucsd/fullcit?p3189792.

Full text
APA, Harvard, Vancouver, ISO, and other styles
24

Chintala, Venkatram Reddy. "Digital image data representation." Ohio : Ohio University, 1986. http://www.ohiolink.edu/etd/view.cgi?ohiou1183128563.

Full text
APA, Harvard, Vancouver, ISO, and other styles
25

Mueller, Guenter. "DIGITAL DATA RECORDING: NEW WAYS IN DATA PROCESSING." International Foundation for Telemetering, 2000. http://hdl.handle.net/10150/606505.

Full text
Abstract:
International Telemetering Conference Proceedings / October 23-26, 2000 / Town & Country Hotel and Conference Center, San Diego, California
With the introduction of digital data recorders new ways of data processing have been developed. The three most important improvements are discussed in this paper: A) By processing PCM Data from a digital recorder by using the SCSI-Interface our ground station has developed software to detect the synchronization pattern of the PCM data and then perform software frame decommutation. Many advantages will be found with this method. B) New digital recorders already use the CCSDS Standard as the internal recording format. Once this technique is implemented in our ground station’s software and becomes part of our software engineering team’s general know-how, the switch to CCSDS telemetry in the future will require no quantum leap in effort. C) Digital recorders offer a very new application: Writing data to a digital tape in the recorder’s own format, allows the replay of data using the recorder’s interfaces; i.e. writing vibration data from the host system to tape, using the analog format of the digital recorder, allows the analysis of the data either in analog form, using the analog interface of the recorder, or in digital form.
APA, Harvard, Vancouver, ISO, and other styles
26

Wedge, Daniel John. "Video sequence synchronization." University of Western Australia. School of Computer Science and Software Engineering, 2008. http://theses.library.uwa.edu.au/adt-WU2008.0084.

Full text
Abstract:
[Truncated abstract] Video sequence synchronization is necessary for any computer vision application that integrates data from multiple simultaneously recorded video sequences. With the increased availability of video cameras as either dedicated devices, or as components within digital cameras or mobile phones, a large volume of video data is available as input for a growing range of computer vision applications that process multiple video sequences. To ensure that the output of these applications is correct, accurate video sequence synchronization is essential. Whilst hardware synchronization methods can embed timestamps into each sequence on-the-fly, they require specialized hardware and it is necessary to set up the camera network in advance. On the other hand, computer vision-based software synchronization algorithms can be used to post-process video sequences recorded by cameras that are not networked, such as common consumer hand-held video cameras or cameras embedded in mobile phones, or to synchronize historical videos for which hardware synchronization was not possible. The current state-of-the-art software algorithms vary in their input and output requirements and camera configuration assumptions. ... Next, I describe an approach that synchronizes two video sequences where an object exhibits ballistic motions. Given the epipolar geometry relating the two cameras and the imaged ballistic trajectory of an object, the algorithm uses a novel iterative approach that exploits object motion to rapidly determine pairs of temporally corresponding frames. This algorithm accurately synchronizes videos recorded at different frame rates and takes few iterations to converge to sub-frame accuracy. Whereas the method presented by the first algorithm integrates tracking data from all frames to synchronize the sequences as a whole, this algorithm recovers the synchronization by locating pairs of temporally corresponding frames in each sequence. Finally, I introduce an algorithm for synchronizing two video sequences recorded by stationary cameras with unknown epipolar geometry. This approach is unique in that it recovers both the frame rate ratio and the frame offset of the two sequences by finding matching space-time interest points that represent events in each sequence; the algorithm does not require object tracking. RANSAC-based approaches that take a set of putatively matching interest points and recover either a homography or a fundamental matrix relating a pair of still images are well known. This algorithm extends these techniques using space-time interest points in place of spatial features, and uses nested instances of RANSAC to also recover the frame rate ratio and frame offset of a pair of video sequences. In this thesis, it is demonstrated that each of the above algorithms can accurately recover the frame rate ratio and frame offset of a range of real video sequences. Each algorithm makes a contribution to the body of video sequence synchronization literature, and it is shown that the synchronization problem can be solved using a range of approaches.
APA, Harvard, Vancouver, ISO, and other styles
27

Grundmann, Matthias. "Real-time content aware resizing of video." Thesis, Atlanta, Ga. : Georgia Institute of Technology, 2008. http://hdl.handle.net/1853/26622.

Full text
Abstract:
Thesis (M. S.)--Computing, Georgia Institute of Technology, 2009.
Committee Chair: Essa, Irfan; Committee Member: Dellaert, Frank; Committee Member: Turk, Greg. Part of the SMARTech Electronic Thesis and Dissertation Collection.
APA, Harvard, Vancouver, ISO, and other styles
28

Chen, Jiawen (Jiawen Kevin). "Efficient data structures for piecewise-smooth video processing." Thesis, Massachusetts Institute of Technology, 2011. http://hdl.handle.net/1721.1/66003.

Full text
Abstract:
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2011.
Cataloged from PDF version of thesis.
Includes bibliographical references (p. 95-102).
A number of useful image and video processing techniques, ranging from low level operations such as denoising and detail enhancement to higher level methods such as object manipulation and special effects, rely on piecewise-smooth functions computed from the input data. In this thesis, we present two computationally efficient data structures for representing piecewise-smooth visual information and demonstrate how they can dramatically simplify and accelerate a variety of video processing algorithms. We start by introducing the bilateral grid, an image representation that explicitly accounts for intensity edges. By interpreting brightness values as Euclidean coordinates, the bilateral grid enables simple expressions for edge-aware filters. Smooth functions defined on the bilateral grid are piecewise-smooth in image space. Within this framework, we derive efficient reinterpretations of a number of edge-aware filters commonly used in computational photography as operations on the bilateral grid, including the bilateral filter, edgeaware scattered data interpolation, and local histogram equalization. We also show how these techniques can be easily parallelized onto modern graphics hardware for real-time processing of high definition video. The second data structure we introduce is the video mesh, designed as a flexible central data structure for general-purpose video editing. It represents objects in a video sequence as 2.5D "paper cutouts" and allows interactive editing of moving objects and modeling of depth, which enables 3D effects and post-exposure camera control. In our representation, we assume that motion and depth are piecewise-smooth, and encode them sparsely as a set of points tracked over time. The video mesh is a triangulation over this point set and per-pixel information is obtained by interpolation. To handle occlusions and detailed object boundaries, we rely on the user to rotoscope the scene at a sparse set of frames using spline curves. We introduce an algorithm to robustly and automatically cut the mesh into local layers with proper occlusion topology, and propagate the splines to the remaining frames. Object boundaries are refined with per-pixel alpha mattes. At its core, the video mesh is a collection of texture-mapped triangles, which we can edit and render interactively using graphics hardware. We demonstrate the effectiveness of our representation with special effects such as 3D viewpoint changes, object insertion, depthof- field manipulation, and 2D to 3D video conversion.
by Jiawen Chen.
Ph.D.
APA, Harvard, Vancouver, ISO, and other styles
29

Hu, Yongtao, and 胡永涛. "Multimodal speaker localization and identification for video processing." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2014. http://hdl.handle.net/10722/212633.

Full text
APA, Harvard, Vancouver, ISO, and other styles
30

Kang, Jung Won. "Effective temporal video segmentation and content-based audio-visual video clustering." Diss., Georgia Institute of Technology, 2003. http://hdl.handle.net/1853/13731.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Huang, Guo Heng. "On-line video object segmentation using superpixel approach." Thesis, University of Macau, 2017. http://umaclib3.umac.mo/record=b3691897.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Hamed, Mahmoud S. "Film and video restoration using nonlinear digital image processing techniques." Thesis, University of Strathclyde, 2004. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.400321.

Full text
APA, Harvard, Vancouver, ISO, and other styles
33

Dickinson, Keith William. "Traffic data capture and analysis using video image processing." Thesis, University of Sheffield, 1986. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.306374.

Full text
APA, Harvard, Vancouver, ISO, and other styles
34

LI, WEI. "HIERARCHICAL SUMMARIZATION OF VIDEO DATA." University of Cincinnati / OhioLINK, 2007. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1186941444.

Full text
APA, Harvard, Vancouver, ISO, and other styles
35

Turkmen, Muserref. "Digital Image Processing Of Remotely Sensed Oceanographic Data." Master's thesis, METU, 2008. http://etd.lib.metu.edu.tr/upload/12609948/index.pdf.

Full text
Abstract:
Developing remote sensing instrumentation allows obtaining information about an area rapidly and with low costs. This fact offers a challenge to remote sensing algorithms aimed at extracting information about an area from the available re¬
mote sensing data. A very typical and important problem being interpretation of satellite images. A very efficient approach to remote sensing is employing discrim¬
inant functions to distinguish different landscape classes from satellite images. Various methods on this direction are already studied. However, the efficiency of the studied methods are still not very high. In this thesis, we will improve efficiency of remote sensing algorithms. Besides we will investigate improving boundary detection methods on satellite images.
APA, Harvard, Vancouver, ISO, and other styles
36

Ilg, Markus. "Digital processing of map data in raster format /." Zürich : Geographisches Institut Eidgenössische Technische Hochschule, 1986. http://e-collection.ethbib.ethz.ch/show?type=diss&nr=7973.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Yip, Shu Kei. "Digital watermarking and data hiding /." View abstract or full-text, 2006. http://library.ust.hk/cgi/db/thesis.pl?ECED%202006%20YIP.

Full text
APA, Harvard, Vancouver, ISO, and other styles
38

Li, Richard Yuk Ming. "Digital watermarking and data hiding /." View abstract or full-text, 2007. http://library.ust.hk/cgi/db/thesis.pl?ECED%202007%20LIR.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

AZAMBUJA, MARCELLO DE LIMA. "A CLOUD COMPUTING ARCHITECTURE FOR LARGE SCALE VIDEO DATA PROCESSING." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2011. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=28923@1.

Full text
Abstract:
O advento da Internet coloca grandes desafios para o projeto de sistemas de submissão abertos, uma vez que elimina barreiras físicas e geográficas. A redução de custos, associada a passagem de mídias analógicas para digitais, dificultou a projeção de capacidades e recursos necessários para a construção de tais sistemas. Neste trabalho propomos uma arquitetura de software, baseada em computação na nuvem, que provê a escalabilidade necessária para a construção de sistemas de submissão abertos. Estes sistemas são caracterizados pela geração de grandes volumes de dados. Um caso real de uso é analisado utilizando o processamento de vídeos.
The advent of the Internet poses great challenges to the design of public submission systems as it eliminates traditional barriers, such as geographical location and cost. With open global access, it is very hard to estimate storage space and processing power required by this class of applications. In this thesis we explore cloud computing technology as an alternative solution. The main contribution of this work is a general architecture in which to built open access, data intensive, public submission systems. A real world scenario is analyzed using this architecture for video processing.
APA, Harvard, Vancouver, ISO, and other styles
40

Young, Robert W. "Video coding using lapped transforms." Thesis, University of Cambridge, 1993. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.309337.

Full text
APA, Harvard, Vancouver, ISO, and other styles
41

Liu, Sam J. "Low bit-rate image and video compression using adaptive segmentation and quantization." Diss., Georgia Institute of Technology, 1993. http://hdl.handle.net/1853/14850.

Full text
APA, Harvard, Vancouver, ISO, and other styles
42

Vemulapalli, Smita. "Audio-video based handwritten mathematical content recognition." Diss., Georgia Institute of Technology, 2012. http://hdl.handle.net/1853/45958.

Full text
Abstract:
Recognizing handwritten mathematical content is a challenging problem, and more so when such content appears in classroom videos. However, given the fact that in such videos the handwritten text and the accompanying audio refer to the same content, a combination of video and audio based recognizer has the potential to significantly improve the content recognition accuracy. This dissertation, using a combination of video and audio based recognizers, focuses on improving the recognition accuracy associated with handwritten mathematical content in such videos. Our approach makes use of a video recognizer as the primary recognizer and a multi-stage assembly, developed as part of this research, is used to facilitate effective combination with an audio recognizer. Specifically, we address the following challenges related to audio-video based handwritten mathematical content recognition: (1) Video Preprocessing - generates a timestamped sequence of segmented characters from the classroom video in the face of occlusions and shadows caused by the instructor, (2) Ambiguity Detection - determines the subset of input characters that may have been incorrectly recognized by the video based recognizer and forwards this subset for disambiguation, (3) A/V Synchronization - establishes correspondence between the handwritten character and the spoken content, (4) A/V Combination - combines the synchronized outputs from the video and audio based recognizers and generates the final recognized character, and (5) Grammar Assisted A/V Based Mathematical Content Recognition - utilizes a base mathematical speech grammar for both character and structure disambiguation. Experiments conducted using videos recorded in a classroom-like environment demonstrate the significant improvements in recognition accuracy that can be achieved using our techniques.
APA, Harvard, Vancouver, ISO, and other styles
43

So, Wai-ki, and 蘇慧琪. "Shadow identification in traffic video sequences." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2005. http://hub.hku.hk/bib/B32045967.

Full text
APA, Harvard, Vancouver, ISO, and other styles
44

Wang, Jue. "Foreground segmentation in images and video : methods, systems, and applications /." Thesis, Connect to this title online; UW restricted, 2007. http://hdl.handle.net/1773/6130.

Full text
APA, Harvard, Vancouver, ISO, and other styles
45

Gunturk, Bahadir K. "Multi-frame information fusion for image and video enhancement." Diss., Available online, Georgia Institute of Technology, 2004:, 2003. http://etd.gatech.edu/theses/available/etd-04072004-180015/unrestricted/gunturk%5Fbahadir%5Fk%5F200312%5Fphd.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
46

Neo, Yew Lam. "Digital processing algorithms for bistatic Synthetic Aperture Radar data." Thesis, University of British Columbia, 2007. http://hdl.handle.net/2429/31451.

Full text
Abstract:
The motivations for this thesis are the investigation of bistatic Synthetic Aperture Radar (SAR) image formation and the development of bistatic SAR algorithms to accommodate various bistatic SAR geometries. Traditional monostatic SAR algorithms based on frequency domain methods assume a single square root (a hyperbolic) range equation. In bistatic SAR data, the range history of a target has a Double Square Root (DSR) in the range equation as both the transmitter and receiver can assume different motion trajectories. Thus, monostatic algorithms are not able to focus bistatic SAR data. The key step to many frequency based algorithms is to find an analytical solution for the spectrum of a reference point target. No simple analytical solution exists for the bistatic case because of the DSR in the range equation. Several algorithms have been developed to overcome this difficulty. These algorithms are reviewed and analyzed in this thesis to compare their processing accuracies and the type of operations they require. A solution to the two-dimensional point target spectrum based on the reversion of a power series for the general bistatic case is presented in this thesis. The accuracy of this new point target spectrum is compared with existing analytical point target spectra. Using this spectrum result, a bistatic Range Doppler Algorithm (RDA) is developed to handle the azimuth-invariant, bistatic case. In addition, the algorithm is used to focus real bistatic data acquired with two X-band SAR systems from Forschungsgesellschaft für Angewandte Naturwissenschaften (FLAN), namely the Airborne Experimental Radar II (AER-II) and the Phased Array Multifunctional Imaging Radar (PAMIR). To handle azimuth-variant cases, the Non-Linear Chirp Scaling (NLCS) algorithm is used. The original NLCS algorithm is developed to focus only short-wavelength bistatic cases with one platform moving and imaging at broadside and the other stationary. It is found that the NLCS is able to cope with the general bistatic case since it is able to handle range and azimuth-variant signals. To exploit this processing capability, the algorithm is extended further to accommodate squinted and long-wavelength bistatic cases where both platforms have dissimilar velocities and flight paths slightly non-parallel.
Applied Science, Faculty of
Electrical and Computer Engineering, Department of
Graduate
APA, Harvard, Vancouver, ISO, and other styles
47

Nagahara, Masaaki. "Multirate digital signal processing via sampled-data H∞ optimization." 京都大学 (Kyoto University), 2003. http://hdl.handle.net/2433/120982.

Full text
APA, Harvard, Vancouver, ISO, and other styles
48

Kee, Eric. "Video Mosaicking Using Ancillary Data to Facilitate Size Estimation." Thesis, Virginia Tech, 2003. http://hdl.handle.net/10919/32693.

Full text
Abstract:
This thesis describes a mosaicking system designed to generate image mosaics that facilitate size estimation of 3-dimensional objects by improving data obtained with a multi-sensor video camera. The multi-sensor camera is equipped with a pulse laser-rangefinder and internally mounted inclinometers that measure instrument orientation about three axes. Using orientation data and video data, mosaics are constructed to reduce orientation data errors by augmenting orientation data with image information. Mosaicking is modeled as a 7-step refinement process: 1) an initial mosaic is constructed using orientation information obtained from the camera's inclinometers; 2) mosaics are refined by using coarse-to-fine processing to minimize an energy metric and, consequently, align overlapping video frames; 3) pair-wise mosaicking errors are detected, and removed, using an energy-based confidence metric; 4) mosaic accuracy is refined via color analysis; 5) mosaic accuracy is refined by estimating an affine transformation to align overlapping frames; 6) affine transformation approximations between overlapping video frames are used to reduce image noise through super-resolution; 7) original orientation data are corrected given the refined orientations of images within the mosaic. The mosaicking system has been tested using objects of known size and orientation accuracy has been improved by 86% for these cases.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
49

Srinivasan, Sabeshan. "Object Tracking in Distributed Video Networks Using Multi-Dimentional Signatures." Fogler Library, University of Maine, 2006. http://www.library.umaine.edu/theses/pdf/SrinivasanSX2006.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
50

Elnaggar, Ayman Ibrahim. "Scalable parallel VLSI architectures and algorithms for digital signal and video processing." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk2/tape17/PQDD_0032/NQ27134.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography