Gotowa bibliografia na temat „Video synchronization”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Spis treści
Zobacz listy aktualnych artykułów, książek, rozpraw, streszczeń i innych źródeł naukowych na temat „Video synchronization”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Artykuły w czasopismach na temat "Video synchronization"
EL-Sallam, Amar A., i Ajmal S. Mian. "Correlation based speech-video synchronization". Pattern Recognition Letters 32, nr 6 (kwiecień 2011): 780–86. http://dx.doi.org/10.1016/j.patrec.2011.01.001.
Pełny tekst źródłaLin, E. T., i E. J. Delp. "Temporal Synchronization in Video Watermarking". IEEE Transactions on Signal Processing 52, nr 10 (październik 2004): 3007–22. http://dx.doi.org/10.1109/tsp.2004.833866.
Pełny tekst źródłaFu, Jia Bing, i He Wei Yu. "Audio-Video Synchronization Method Based on Playback Time". Applied Mechanics and Materials 300-301 (luty 2013): 1677–80. http://dx.doi.org/10.4028/www.scientific.net/amm.300-301.1677.
Pełny tekst źródłaLi, Xiao Ni, He Xin Chen i Da Zhong Wang. "Research on Audio-Video Synchronization Coding Based on Mode Selection in H.264". Applied Mechanics and Materials 182-183 (czerwiec 2012): 701–5. http://dx.doi.org/10.4028/www.scientific.net/amm.182-183.701.
Pełny tekst źródłaLiu, Yiguang, Menglong Yang i Zhisheng You. "Video synchronization based on events alignment". Pattern Recognition Letters 33, nr 10 (lipiec 2012): 1338–48. http://dx.doi.org/10.1016/j.patrec.2012.02.009.
Pełny tekst źródłaMu Li i Vishal Monga. "Twofold Video Hashing With Automatic Synchronization". IEEE Transactions on Information Forensics and Security 10, nr 8 (sierpień 2015): 1727–38. http://dx.doi.org/10.1109/tifs.2015.2425362.
Pełny tekst źródłaZhou, Zhongyi, Anran Xu i Koji Yatani. "SyncUp". Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, nr 3 (9.09.2021): 1–25. http://dx.doi.org/10.1145/3478120.
Pełny tekst źródłaYang, Shu Zhen, Guang Lin Chu i Ming Wang. "A Study on Parallel Processing Video Splicing System with Multi-Processor". Applied Mechanics and Materials 198-199 (wrzesień 2012): 304–9. http://dx.doi.org/10.4028/www.scientific.net/amm.198-199.304.
Pełny tekst źródłaKwon, Ohsung. "Class Analysis Method Using Video Synchronization Algorithm". Journal of The Korean Association of Information Education 19, nr 4 (30.12.2015): 441–48. http://dx.doi.org/10.14352/jkaie.2015.19.4.441.
Pełny tekst źródłaChen, T., H. P. Graf i K. Wang. "Lip synchronization using speech-assisted video processing". IEEE Signal Processing Letters 2, nr 4 (kwiecień 1995): 57–59. http://dx.doi.org/10.1109/97.376913.
Pełny tekst źródłaRozprawy doktorskie na temat "Video synchronization"
Wedge, Daniel John. "Video sequence synchronization". University of Western Australia. School of Computer Science and Software Engineering, 2008. http://theses.library.uwa.edu.au/adt-WU2008.0084.
Pełny tekst źródłaYang, Hsueh-szu, i Benjamin Kupferschmidt. "Time Stamp Synchronization in Video Systems". International Foundation for Telemetering, 2010. http://hdl.handle.net/10150/605988.
Pełny tekst źródłaSynchronized video is crucial for data acquisition and telecommunication applications. For real-time applications, out-of-sync video may cause jitter, choppiness and latency. For data analysis, it is important to synchronize multiple video channels and data that are acquired from PCM, MIL-STD-1553 and other sources. Nowadays, video codecs can be easily obtained to play most types of video. However, a great deal of effort is still required to develop the synchronization methods that are used in a data acquisition system. This paper will describe several methods that TTC has adopted in our system to improve the synchronization of multiple data sources.
Gaskill, David M. "TECHNIQUES FOR SYNCHRONIZING THERMAL ARRAY CHART RECORDERS TO VIDEO". International Foundation for Telemetering, 1992. http://hdl.handle.net/10150/608901.
Pełny tekst źródłaVideo tape is becoming more and more popular for storing and analyzing missions. Video tape is inexpensive, it can hold a two hour test, and it can be edited and manipulated by easily available consumer electronics equipment. Standard technology allows each frame to be time stamped with SMPTE code, so that any point in the mission can be displayed on a CRT. To further correlate data from multiple acquisition systems, the SMPTE code can be derived from IRIG using commercially available code converters. Unfortunately, acquiring and storing analog data has not been so easy. Typically, analog signals from various sensors are coded, transmitted, decoded and sent to a chart recorder. Since chart recorders cannot normally store an entire mission internally, or time stamp each data value, it is very difficult for an analyst to accurately correlate analog data to an individual video frame. Normally the only method is to note the time stamp on the video frame and unroll the chart to the appropriate second or minute, depending on the code used, noted in the margin, and estimate the frame location as a percentage of the time code period. This is very inconvenient if the telemetrist is trying to establish an on-line data retreival system. To make matters worse, the methods of presentation are very different, chart paper as opposed to a CRT, and require the analyst to shift focus constantly. For these reasons, many telemetry stations do not currently have a workable plan to integrate analog and video subsystems even though it is now generally agreed that such integration is ultimately desirable.
Daami, Mourad. "Synchronization control of coded video streams, algorithms and implementation". Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk2/ftp04/mq26314.pdf.
Pełny tekst źródłaAbraham, Justin Kuruvilla. "Study of the TR Synchronization and Video Conversion Unit". Master's thesis, University of Cape Town, 2012. http://hdl.handle.net/11427/14137.
Pełny tekst źródłaYilmaz, Ayhan. "Robust Video Transmission Using Data Hiding". Master's thesis, METU, 2003. http://etd.lib.metu.edu.tr/upload/1093509/index.pdf.
Pełny tekst źródłahence, improves the reconstruction video quality without significant extra channel utilization. A complete error resilient video transmission codec is proposed, utilizing imperceptible embedded information for combined detecting, resynchronization and reconstruction of the errors and lost data. The data, which is imperceptibly embedded into the video itself at the encoder, is extracted from the video at the decoder side to be utilized in error concealment. A spatial domain error recovery technique, which hides edge orientation information of a block, and a resynchronization technique, which embeds bit length of a block into other blocks are combined, as well as some parity information about the hidden data, to conceal channel errors on intra-coded frames of a video sequence. The errors on inter-coded frames are basically recovered by hiding motion vector information along with a checksum into the next frames. The simulation results show that the proposed approach performs superior to conventional approaches for concealing the errors in binary symmetric channels, especially for higher bit rates and error rates.
Potetsianakis, Emmanouil. "Enhancing video applications through timed metadata". Electronic Thesis or Diss., Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLT029.
Pełny tekst źródłaVideo recording devices are often equipped with sensors (smartphones for example, with GPS receiver, gyroscope etc.), or used in settings where sensors are present (e.g. monitor cameras, in areas with temperature and/or humidity sensors). As a result, many systems process and distribute video together with timed metadata streams, often sourced as User-Generated Content. Video delivery has been thoroughly studied, however timed metadata streams have varying characteristics and forms, thus a consistent and effective way to handle them in conjunction with the video streams does not exist. In this Thesis we study ways to enhance video applications through timed metadata. We define as timed metadata all the non-audiovisual data recorded or produced, that are relevant to a specific time on the media timeline. ”Enhancing” video applications has a double meaning, and this work consists of two respective parts. First, using the timed metadata to extend the capabilities of multimedia applications, by introducing novel functionalities. Second, using the timed metadata to improve the content delivery for such applications. To extend multimedia applications, we have taken an exploratory approach, and we demonstrate two use cases with application examples. In the first case, timed metadata is used as input for generating content, and in the second, it is used to extend the navigational capabilities for the underlying multimedia content. By designing and implementing two different application scenarios we were able to identify the potential and limitations of video systems with timed metadata. We use the findings from the first part, to work from the perspective of enhancing video applications, by using the timed metadata to improve delivery of the content. More specifically, we study the use of timed metadata for multi-variable adaptation in multi-view video delivery - and we test our proposals on one of the platforms developed previously. Our final contribution is a buffering scheme for synchronous and lowlatency playback in live streaming systems
Carranza, López José Camilo. "On the synchronization of two metronomes and their related dynamics /". Ilha Solteira, 2017. http://hdl.handle.net/11449/151204.
Pełny tekst źródłaResumo: Nesta tese são investigadas, teórica e experimentalmente, a sincronização em fase e a sincronização em anti-fase de dois metrônomos oscilando sobre uma base móvel, a partir de um modelo aqui proposto. Uma descrição do funcionamento do mecanismo de escapamento dos metrônomos é feita, junto a um estudo da relação entre este e o oscilador de van der Pol. Também uma aproximação experimental do valor do amortecimento do metrônomo é fornecida. A frequência instantânea das respostas, numérica e experimental, do sistema é usada na analise. A diferença de outros trabalhos prévios, os dados experimentais têm sido adquiridos usando vídeos dos experimentos e extraídos com ajuda do software Tracker. Para investigar a relação entre as condições iniciais do sistema e seu estado final de sincronização, foram usados mapas bidimensionais chamados ‘basins of attraction’. A relação entre o modelo proposto e um modelo prévio também é mostrada. Encontrou-se que os parâmetros relevantes em relação a ambos os tipos de sincronização são a razão entre a massa do metrônomo e a massa da base, e o amortecimento do sistema. Tem-se encontrado, tanto experimental quanto teoricamente, que a frequência de oscilação dos metrônomos aumenta quando o sistema sincroniza-se em fase, e se mantém a mesma de um metrônomo isolado quando o sistema sincroniza-se em anti-fase. A partir de simulações numéricas encontrou-se que, em geral, incrementos no amortecimento do sistema levam ao sistema se sincronizar mais em fase d... (Resumo completo, clicar acesso eletrônico abaixo)
Doutor
Wehbe, Hassan. "Synchronisation automatique d'un contenu audiovisuel avec un texte qui le décrit". Thesis, Toulouse 3, 2016. http://www.theses.fr/2016TOU30104/document.
Pełny tekst źródłaWe address the problem of automatic synchronization of an audiovisual content with a procedural text that describes it. The strategy consists in extracting pieces of information about the structure from both contents, and in matching them depending on their types. We propose two video analysis tools that respectively extract: * Limits of events of interest using an approach inspired by dictionary quantization. * Segments that enclose a repeated action based on the YIN frequency analysis method. We then propose a synchronization system that merges results coming from these tools in order to establish links between textual instructions and the corresponding video segments. To do so, a "Confidence Matrix" is built and recursively processed in order to identify these links in respect with their reliability
Carranza, López José Camilo [UNESP]. "On the synchronization of two metronomes and their related dynamics". Universidade Estadual Paulista (UNESP), 2017. http://hdl.handle.net/11449/151204.
Pełny tekst źródłaApproved for entry into archive by Luiz Galeffi (luizgaleffi@gmail.com) on 2017-07-26T18:31:30Z (GMT) No. of bitstreams: 1 carranzalopez_jc_dr_ilha.pdf: 11035322 bytes, checksum: efe400c07b13cabff41e927078789c59 (MD5)
Made available in DSpace on 2017-07-26T18:31:30Z (GMT). No. of bitstreams: 1 carranzalopez_jc_dr_ilha.pdf: 11035322 bytes, checksum: efe400c07b13cabff41e927078789c59 (MD5) Previous issue date: 2017-06-05
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Nesta tese são investigadas, teórica e experimentalmente, a sincronização em fase e a sincronização em anti-fase de dois metrônomos oscilando sobre uma base móvel, a partir de um modelo aqui proposto. Uma descrição do funcionamento do mecanismo de escapamento dos metrônomos é feita, junto a um estudo da relação entre este e o oscilador de van der Pol. Também uma aproximação experimental do valor do amortecimento do metrônomo é fornecida. A frequência instantânea das respostas, numérica e experimental, do sistema é usada na analise. A diferença de outros trabalhos prévios, os dados experimentais têm sido adquiridos usando vídeos dos experimentos e extraídos com ajuda do software Tracker. Para investigar a relação entre as condições iniciais do sistema e seu estado final de sincronização, foram usados mapas bidimensionais chamados ‘basins of attraction’. A relação entre o modelo proposto e um modelo prévio também é mostrada. Encontrou-se que os parâmetros relevantes em relação a ambos os tipos de sincronização são a razão entre a massa do metrônomo e a massa da base, e o amortecimento do sistema. Tem-se encontrado, tanto experimental quanto teoricamente, que a frequência de oscilação dos metrônomos aumenta quando o sistema sincroniza-se em fase, e se mantém a mesma de um metrônomo isolado quando o sistema sincroniza-se em anti-fase. A partir de simulações numéricas encontrou-se que, em geral, incrementos no amortecimento do sistema levam ao sistema se sincronizar mais em fase do que em anti-fase. Adicionalmente se encontrou que, para dado valor de amortecimento, diminuir a massa da base leva a uma situação em que a sincronização em anti-fase é mais comum do que a sincronização em fase.
This thesis concerns a theoretical and experimental investigation into the synchronization of two coupled metronomes. A simplified model is proposed to study in-phase and anti-phase synchronization of two metronomes oscillating on a mobile base. A description of the escapement mechanism driving metronomes is given and its relationship with the van der Pol oscillator is discussed. Also an experimental value for the damping in the metronome is determined. The instantaneous frequency of the responses from both numerical and experimental data is used in the analysis. Unlike previous studies, measurements are made using videos and the time domain responses of the metronomes extracted by means of tracker software. Basins of attraction are used to investigate the relationship between initial conditions, parameters and both final synchronization states. The relationship between the model and a previous pendulum model is also shown. The key parameters concerning both kind of synchronization have been found to be the mass ratio between the metronome mass and the base mass, and the damping in the system. It has been shown, both theoretically and experimentally, that the frequency of oscillation of the metronomes increases when the system reaches in-phase synchronization, and is the same as an isolated metronome when the system synchronizes in anti-phase. From numerical simulations, it has been found that, in general, increasing damping leads the system to synchronize more in-phase than in anti-phase. It has also been found that, for a given damping value, decreasing the mass of the base results in the situation where anti-phase synchronization is more common than in-phase synchronization.
Książki na temat "Video synchronization"
Rona, Jeffrey C. Synchronization from reel to reel: A complete guide for the synchronization of audio, film & video. Milwaukee, WI: Hal Leonard Publishing Corporation, 1990.
Znajdź pełny tekst źródłaRona, Jeffrey C. Synchronization, from reel to reel: A complete guide for the synchronization of audio, film & video. Redaktorzy Schiff Ronny S i Wilkinson Scott R. 1953-. Milwaukee, WI: H. Leonard Pub., Corp., 1989.
Znajdź pełny tekst źródłaHawkins, Stan. Aesthetics and Hyperembodiment in Pop Videos. Redaktorzy John Richardson, Claudia Gorbman i Carol Vernallis. Oxford University Press, 2013. http://dx.doi.org/10.1093/oxfordhb/9780199733866.013.002.
Pełny tekst źródłaST 318:2015: Synchronization of 59.94- or 50-Hz Related Video and Audio Systems in Analog and Digital Areas — Reference Signals. 3 Barker Avenue., White Plains, NY 10601: The Society of Motion Picture and Television Engineers SMPTE, 2015. http://dx.doi.org/10.5594/smpte.st318.2015.
Pełny tekst źródłaBarrière, Jean-Baptiste, i Aleksi Barrière. When Music Unfolds into Image. Redaktor Yael Kaduri. Oxford University Press, 2016. http://dx.doi.org/10.1093/oxfordhb/9780199841547.013.39.
Pełny tekst źródłaCzęści książek na temat "Video synchronization"
Wang, Xue, i Qing Wang. "Video Synchronization with Trajectory Pulse". W Communications in Computer and Information Science, 12–19. Singapore: Springer Singapore, 2016. http://dx.doi.org/10.1007/978-981-10-3476-3_2.
Pełny tekst źródłaWedge, Daniel, Du Huynh i Peter Kovesi. "Motion Guided Video Sequence Synchronization". W Computer Vision – ACCV 2006, 832–41. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11612704_83.
Pełny tekst źródłaBazin, Jean-Charles, i Alexander Sorkine-Hornung. "ActionSnapping: Motion-Based Video Synchronization". W Computer Vision – ECCV 2016, 155–69. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-46454-1_10.
Pełny tekst źródłaFurht, Borko, Stephen W. Smoliar i HongJiang Zhang. "Multimedia Networking and Synchronization". W Video and Image Processing in Multimedia Systems, 33–57. Boston, MA: Springer US, 1995. http://dx.doi.org/10.1007/978-1-4615-2277-5_2.
Pełny tekst źródłaKryston, Kevin, Eric Novotny, Ralf Schmälzle i Ron Tamborini. "Social Demand in Video Games and the Synchronization Theory of Flow". W Video Games, 161–77. New York, NY : Routledge, 2018. | Series: Electronic media research series: Routledge, 2018. http://dx.doi.org/10.4324/9781351235266-10.
Pełny tekst źródłaCai, Ying, i Grammati E. Pantziou. "A Synchronization Mechanism for Multimedia Presentation". W Multimedia Communications and Video Coding, 157–62. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_20.
Pełny tekst źródłaZuniga, Gabriel, i Ephraim Feig. "Synchronization Issues on Software MPEG Playback Systems". W Multimedia Communications and Video Coding, 141–45. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_18.
Pełny tekst źródłaLiu, Changdong, Yong Xie, Myung J. Lee i Tarek N. Saadawi. "Adaptive Synchronization in Real-Time Multimedia Applications". W Multimedia Communications and Video Coding, 147–56. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_19.
Pełny tekst źródłaBulterman, Dick C. A., i Robert Liere. "Multimedia synchronization and UNIX". W Network and Operating System Support for Digital Audio and Video, 105–19. Berlin, Heidelberg: Springer Berlin Heidelberg, 1992. http://dx.doi.org/10.1007/3-540-55639-7_10.
Pełny tekst źródłaRothermel, Kurt, i Gabriel Dermler. "Synchronization in Joint-Viewing environments". W Network and Operating System Support for Digital Audio and Video, 106–18. Berlin, Heidelberg: Springer Berlin Heidelberg, 1993. http://dx.doi.org/10.1007/3-540-57183-3_10.
Pełny tekst źródłaStreszczenia konferencji na temat "Video synchronization"
Waingankar, P., i D. Valsan. "Audio-video synchronization". W the International Conference & Workshop. New York, New York, USA: ACM Press, 2011. http://dx.doi.org/10.1145/1980022.1980068.
Pełny tekst źródłaSung, Chih-Ta S. "MPEG audio-video synchronization". W Electronic Imaging '97, redaktorzy Sethuraman Panchanathan i Frans Sijstermans. SPIE, 1997. http://dx.doi.org/10.1117/12.263515.
Pełny tekst źródłaMediavilla, Ricardo. "Automatic test station for network synchronization performance characterization". W Voice, Video, and Data Communications, redaktorzy John M. Senior, Robert A. Cryan i Chunming Qiao. SPIE, 1997. http://dx.doi.org/10.1117/12.290373.
Pełny tekst źródłaShankar, Sukrit, Joan Lasenby i Anil Kokaram. "Warping trajectories for video synchronization". W the 4th ACM/IEEE international workshop. New York, New York, USA: ACM Press, 2013. http://dx.doi.org/10.1145/2510650.2510654.
Pełny tekst źródłaLin, Eugene T., i Edward J. Delp III. "Temporal synchronization in video watermarking". W Electronic Imaging 2002, redaktorzy Edward J. Delp III i Ping W. Wong. SPIE, 2002. http://dx.doi.org/10.1117/12.465310.
Pełny tekst źródłaStokking, Hans, Pablo Cesar, Fernando Boronat i Mario Montagud. "Media Synchronization Workshop". W TVX'15: ACM International Conference on Interactive Experiences for TV and Online Video. New York, NY, USA: ACM, 2015. http://dx.doi.org/10.1145/2745197.2745699.
Pełny tekst źródłaWieschollek, Patrick, Ido Freeman i Hendrik P. A. Lensch. "Learning Robust Video Synchronization without Annotations". W 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA). IEEE, 2017. http://dx.doi.org/10.1109/icmla.2017.0-173.
Pełny tekst źródłaWu, Yuanyuan, Xiaohai He i Truong Q. Nguyen. "Subframe video synchronization by matching trajectories". W ICASSP 2013 - 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2013. http://dx.doi.org/10.1109/icassp.2013.6638060.
Pełny tekst źródłaLi, Mu, i Vishal Monga. "Twofold video hashing with automatic synchronization". W 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 2014. http://dx.doi.org/10.1109/icip.2014.7026085.
Pełny tekst źródłaYang, Ming, Nikolaos Bourbakis, Zizhong Chen i Monica Trifas. "An Efficient Audio-Video Synchronization Methodology". W Multimedia and Expo, 2007 IEEE International Conference on. IEEE, 2007. http://dx.doi.org/10.1109/icme.2007.4284763.
Pełny tekst źródła