Inhaltsverzeichnis
Auswahl der wissenschaftlichen Literatur zum Thema „Video synchronization“
Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an
Machen Sie sich mit den Listen der aktuellen Artikel, Bücher, Dissertationen, Berichten und anderer wissenschaftlichen Quellen zum Thema "Video synchronization" bekannt.
Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.
Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.
Zeitschriftenartikel zum Thema "Video synchronization"
EL-Sallam, Amar A., und Ajmal S. Mian. „Correlation based speech-video synchronization“. Pattern Recognition Letters 32, Nr. 6 (April 2011): 780–86. http://dx.doi.org/10.1016/j.patrec.2011.01.001.
Der volle Inhalt der QuelleLin, E. T., und E. J. Delp. „Temporal Synchronization in Video Watermarking“. IEEE Transactions on Signal Processing 52, Nr. 10 (Oktober 2004): 3007–22. http://dx.doi.org/10.1109/tsp.2004.833866.
Der volle Inhalt der QuelleFu, Jia Bing, und He Wei Yu. „Audio-Video Synchronization Method Based on Playback Time“. Applied Mechanics and Materials 300-301 (Februar 2013): 1677–80. http://dx.doi.org/10.4028/www.scientific.net/amm.300-301.1677.
Der volle Inhalt der QuelleLi, Xiao Ni, He Xin Chen und Da Zhong Wang. „Research on Audio-Video Synchronization Coding Based on Mode Selection in H.264“. Applied Mechanics and Materials 182-183 (Juni 2012): 701–5. http://dx.doi.org/10.4028/www.scientific.net/amm.182-183.701.
Der volle Inhalt der QuelleLiu, Yiguang, Menglong Yang und Zhisheng You. „Video synchronization based on events alignment“. Pattern Recognition Letters 33, Nr. 10 (Juli 2012): 1338–48. http://dx.doi.org/10.1016/j.patrec.2012.02.009.
Der volle Inhalt der QuelleMu Li und Vishal Monga. „Twofold Video Hashing With Automatic Synchronization“. IEEE Transactions on Information Forensics and Security 10, Nr. 8 (August 2015): 1727–38. http://dx.doi.org/10.1109/tifs.2015.2425362.
Der volle Inhalt der QuelleZhou, Zhongyi, Anran Xu und Koji Yatani. „SyncUp“. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, Nr. 3 (09.09.2021): 1–25. http://dx.doi.org/10.1145/3478120.
Der volle Inhalt der QuelleYang, Shu Zhen, Guang Lin Chu und Ming Wang. „A Study on Parallel Processing Video Splicing System with Multi-Processor“. Applied Mechanics and Materials 198-199 (September 2012): 304–9. http://dx.doi.org/10.4028/www.scientific.net/amm.198-199.304.
Der volle Inhalt der QuelleKwon, Ohsung. „Class Analysis Method Using Video Synchronization Algorithm“. Journal of The Korean Association of Information Education 19, Nr. 4 (30.12.2015): 441–48. http://dx.doi.org/10.14352/jkaie.2015.19.4.441.
Der volle Inhalt der QuelleChen, T., H. P. Graf und K. Wang. „Lip synchronization using speech-assisted video processing“. IEEE Signal Processing Letters 2, Nr. 4 (April 1995): 57–59. http://dx.doi.org/10.1109/97.376913.
Der volle Inhalt der QuelleDissertationen zum Thema "Video synchronization"
Wedge, Daniel John. „Video sequence synchronization“. University of Western Australia. School of Computer Science and Software Engineering, 2008. http://theses.library.uwa.edu.au/adt-WU2008.0084.
Der volle Inhalt der QuelleYang, Hsueh-szu, und Benjamin Kupferschmidt. „Time Stamp Synchronization in Video Systems“. International Foundation for Telemetering, 2010. http://hdl.handle.net/10150/605988.
Der volle Inhalt der QuelleSynchronized video is crucial for data acquisition and telecommunication applications. For real-time applications, out-of-sync video may cause jitter, choppiness and latency. For data analysis, it is important to synchronize multiple video channels and data that are acquired from PCM, MIL-STD-1553 and other sources. Nowadays, video codecs can be easily obtained to play most types of video. However, a great deal of effort is still required to develop the synchronization methods that are used in a data acquisition system. This paper will describe several methods that TTC has adopted in our system to improve the synchronization of multiple data sources.
Gaskill, David M. „TECHNIQUES FOR SYNCHRONIZING THERMAL ARRAY CHART RECORDERS TO VIDEO“. International Foundation for Telemetering, 1992. http://hdl.handle.net/10150/608901.
Der volle Inhalt der QuelleVideo tape is becoming more and more popular for storing and analyzing missions. Video tape is inexpensive, it can hold a two hour test, and it can be edited and manipulated by easily available consumer electronics equipment. Standard technology allows each frame to be time stamped with SMPTE code, so that any point in the mission can be displayed on a CRT. To further correlate data from multiple acquisition systems, the SMPTE code can be derived from IRIG using commercially available code converters. Unfortunately, acquiring and storing analog data has not been so easy. Typically, analog signals from various sensors are coded, transmitted, decoded and sent to a chart recorder. Since chart recorders cannot normally store an entire mission internally, or time stamp each data value, it is very difficult for an analyst to accurately correlate analog data to an individual video frame. Normally the only method is to note the time stamp on the video frame and unroll the chart to the appropriate second or minute, depending on the code used, noted in the margin, and estimate the frame location as a percentage of the time code period. This is very inconvenient if the telemetrist is trying to establish an on-line data retreival system. To make matters worse, the methods of presentation are very different, chart paper as opposed to a CRT, and require the analyst to shift focus constantly. For these reasons, many telemetry stations do not currently have a workable plan to integrate analog and video subsystems even though it is now generally agreed that such integration is ultimately desirable.
Daami, Mourad. „Synchronization control of coded video streams, algorithms and implementation“. Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk2/ftp04/mq26314.pdf.
Der volle Inhalt der QuelleAbraham, Justin Kuruvilla. „Study of the TR Synchronization and Video Conversion Unit“. Master's thesis, University of Cape Town, 2012. http://hdl.handle.net/11427/14137.
Der volle Inhalt der QuelleYilmaz, Ayhan. „Robust Video Transmission Using Data Hiding“. Master's thesis, METU, 2003. http://etd.lib.metu.edu.tr/upload/1093509/index.pdf.
Der volle Inhalt der Quellehence, improves the reconstruction video quality without significant extra channel utilization. A complete error resilient video transmission codec is proposed, utilizing imperceptible embedded information for combined detecting, resynchronization and reconstruction of the errors and lost data. The data, which is imperceptibly embedded into the video itself at the encoder, is extracted from the video at the decoder side to be utilized in error concealment. A spatial domain error recovery technique, which hides edge orientation information of a block, and a resynchronization technique, which embeds bit length of a block into other blocks are combined, as well as some parity information about the hidden data, to conceal channel errors on intra-coded frames of a video sequence. The errors on inter-coded frames are basically recovered by hiding motion vector information along with a checksum into the next frames. The simulation results show that the proposed approach performs superior to conventional approaches for concealing the errors in binary symmetric channels, especially for higher bit rates and error rates.
Potetsianakis, Emmanouil. „Enhancing video applications through timed metadata“. Electronic Thesis or Diss., Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLT029.
Der volle Inhalt der QuelleVideo recording devices are often equipped with sensors (smartphones for example, with GPS receiver, gyroscope etc.), or used in settings where sensors are present (e.g. monitor cameras, in areas with temperature and/or humidity sensors). As a result, many systems process and distribute video together with timed metadata streams, often sourced as User-Generated Content. Video delivery has been thoroughly studied, however timed metadata streams have varying characteristics and forms, thus a consistent and effective way to handle them in conjunction with the video streams does not exist. In this Thesis we study ways to enhance video applications through timed metadata. We define as timed metadata all the non-audiovisual data recorded or produced, that are relevant to a specific time on the media timeline. ”Enhancing” video applications has a double meaning, and this work consists of two respective parts. First, using the timed metadata to extend the capabilities of multimedia applications, by introducing novel functionalities. Second, using the timed metadata to improve the content delivery for such applications. To extend multimedia applications, we have taken an exploratory approach, and we demonstrate two use cases with application examples. In the first case, timed metadata is used as input for generating content, and in the second, it is used to extend the navigational capabilities for the underlying multimedia content. By designing and implementing two different application scenarios we were able to identify the potential and limitations of video systems with timed metadata. We use the findings from the first part, to work from the perspective of enhancing video applications, by using the timed metadata to improve delivery of the content. More specifically, we study the use of timed metadata for multi-variable adaptation in multi-view video delivery - and we test our proposals on one of the platforms developed previously. Our final contribution is a buffering scheme for synchronous and lowlatency playback in live streaming systems
Carranza, López José Camilo. „On the synchronization of two metronomes and their related dynamics /“. Ilha Solteira, 2017. http://hdl.handle.net/11449/151204.
Der volle Inhalt der QuelleResumo: Nesta tese são investigadas, teórica e experimentalmente, a sincronização em fase e a sincronização em anti-fase de dois metrônomos oscilando sobre uma base móvel, a partir de um modelo aqui proposto. Uma descrição do funcionamento do mecanismo de escapamento dos metrônomos é feita, junto a um estudo da relação entre este e o oscilador de van der Pol. Também uma aproximação experimental do valor do amortecimento do metrônomo é fornecida. A frequência instantânea das respostas, numérica e experimental, do sistema é usada na analise. A diferença de outros trabalhos prévios, os dados experimentais têm sido adquiridos usando vídeos dos experimentos e extraídos com ajuda do software Tracker. Para investigar a relação entre as condições iniciais do sistema e seu estado final de sincronização, foram usados mapas bidimensionais chamados ‘basins of attraction’. A relação entre o modelo proposto e um modelo prévio também é mostrada. Encontrou-se que os parâmetros relevantes em relação a ambos os tipos de sincronização são a razão entre a massa do metrônomo e a massa da base, e o amortecimento do sistema. Tem-se encontrado, tanto experimental quanto teoricamente, que a frequência de oscilação dos metrônomos aumenta quando o sistema sincroniza-se em fase, e se mantém a mesma de um metrônomo isolado quando o sistema sincroniza-se em anti-fase. A partir de simulações numéricas encontrou-se que, em geral, incrementos no amortecimento do sistema levam ao sistema se sincronizar mais em fase d... (Resumo completo, clicar acesso eletrônico abaixo)
Doutor
Wehbe, Hassan. „Synchronisation automatique d'un contenu audiovisuel avec un texte qui le décrit“. Thesis, Toulouse 3, 2016. http://www.theses.fr/2016TOU30104/document.
Der volle Inhalt der QuelleWe address the problem of automatic synchronization of an audiovisual content with a procedural text that describes it. The strategy consists in extracting pieces of information about the structure from both contents, and in matching them depending on their types. We propose two video analysis tools that respectively extract: * Limits of events of interest using an approach inspired by dictionary quantization. * Segments that enclose a repeated action based on the YIN frequency analysis method. We then propose a synchronization system that merges results coming from these tools in order to establish links between textual instructions and the corresponding video segments. To do so, a "Confidence Matrix" is built and recursively processed in order to identify these links in respect with their reliability
Carranza, López José Camilo [UNESP]. „On the synchronization of two metronomes and their related dynamics“. Universidade Estadual Paulista (UNESP), 2017. http://hdl.handle.net/11449/151204.
Der volle Inhalt der QuelleApproved for entry into archive by Luiz Galeffi (luizgaleffi@gmail.com) on 2017-07-26T18:31:30Z (GMT) No. of bitstreams: 1 carranzalopez_jc_dr_ilha.pdf: 11035322 bytes, checksum: efe400c07b13cabff41e927078789c59 (MD5)
Made available in DSpace on 2017-07-26T18:31:30Z (GMT). No. of bitstreams: 1 carranzalopez_jc_dr_ilha.pdf: 11035322 bytes, checksum: efe400c07b13cabff41e927078789c59 (MD5) Previous issue date: 2017-06-05
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Nesta tese são investigadas, teórica e experimentalmente, a sincronização em fase e a sincronização em anti-fase de dois metrônomos oscilando sobre uma base móvel, a partir de um modelo aqui proposto. Uma descrição do funcionamento do mecanismo de escapamento dos metrônomos é feita, junto a um estudo da relação entre este e o oscilador de van der Pol. Também uma aproximação experimental do valor do amortecimento do metrônomo é fornecida. A frequência instantânea das respostas, numérica e experimental, do sistema é usada na analise. A diferença de outros trabalhos prévios, os dados experimentais têm sido adquiridos usando vídeos dos experimentos e extraídos com ajuda do software Tracker. Para investigar a relação entre as condições iniciais do sistema e seu estado final de sincronização, foram usados mapas bidimensionais chamados ‘basins of attraction’. A relação entre o modelo proposto e um modelo prévio também é mostrada. Encontrou-se que os parâmetros relevantes em relação a ambos os tipos de sincronização são a razão entre a massa do metrônomo e a massa da base, e o amortecimento do sistema. Tem-se encontrado, tanto experimental quanto teoricamente, que a frequência de oscilação dos metrônomos aumenta quando o sistema sincroniza-se em fase, e se mantém a mesma de um metrônomo isolado quando o sistema sincroniza-se em anti-fase. A partir de simulações numéricas encontrou-se que, em geral, incrementos no amortecimento do sistema levam ao sistema se sincronizar mais em fase do que em anti-fase. Adicionalmente se encontrou que, para dado valor de amortecimento, diminuir a massa da base leva a uma situação em que a sincronização em anti-fase é mais comum do que a sincronização em fase.
This thesis concerns a theoretical and experimental investigation into the synchronization of two coupled metronomes. A simplified model is proposed to study in-phase and anti-phase synchronization of two metronomes oscillating on a mobile base. A description of the escapement mechanism driving metronomes is given and its relationship with the van der Pol oscillator is discussed. Also an experimental value for the damping in the metronome is determined. The instantaneous frequency of the responses from both numerical and experimental data is used in the analysis. Unlike previous studies, measurements are made using videos and the time domain responses of the metronomes extracted by means of tracker software. Basins of attraction are used to investigate the relationship between initial conditions, parameters and both final synchronization states. The relationship between the model and a previous pendulum model is also shown. The key parameters concerning both kind of synchronization have been found to be the mass ratio between the metronome mass and the base mass, and the damping in the system. It has been shown, both theoretically and experimentally, that the frequency of oscillation of the metronomes increases when the system reaches in-phase synchronization, and is the same as an isolated metronome when the system synchronizes in anti-phase. From numerical simulations, it has been found that, in general, increasing damping leads the system to synchronize more in-phase than in anti-phase. It has also been found that, for a given damping value, decreasing the mass of the base results in the situation where anti-phase synchronization is more common than in-phase synchronization.
Bücher zum Thema "Video synchronization"
Rona, Jeffrey C. Synchronization from reel to reel: A complete guide for the synchronization of audio, film & video. Milwaukee, WI: Hal Leonard Publishing Corporation, 1990.
Den vollen Inhalt der Quelle findenS, Schiff Ronny, und Wilkinson Scott R. 1953-, Hrsg. Synchronization, from reel to reel: A complete guide for the synchronization of audio, film & video. Milwaukee, WI: H. Leonard Pub., Corp., 1989.
Den vollen Inhalt der Quelle findenHawkins, Stan. Aesthetics and Hyperembodiment in Pop Videos. Herausgegeben von John Richardson, Claudia Gorbman und Carol Vernallis. Oxford University Press, 2013. http://dx.doi.org/10.1093/oxfordhb/9780199733866.013.002.
Der volle Inhalt der QuelleST 318:2015: Synchronization of 59.94- or 50-Hz Related Video and Audio Systems in Analog and Digital Areas — Reference Signals. 3 Barker Avenue., White Plains, NY 10601: The Society of Motion Picture and Television Engineers SMPTE, 2015. http://dx.doi.org/10.5594/smpte.st318.2015.
Der volle Inhalt der QuelleBarrière, Jean-Baptiste, und Aleksi Barrière. When Music Unfolds into Image. Herausgegeben von Yael Kaduri. Oxford University Press, 2016. http://dx.doi.org/10.1093/oxfordhb/9780199841547.013.39.
Der volle Inhalt der QuelleBuchteile zum Thema "Video synchronization"
Wang, Xue, und Qing Wang. „Video Synchronization with Trajectory Pulse“. In Communications in Computer and Information Science, 12–19. Singapore: Springer Singapore, 2016. http://dx.doi.org/10.1007/978-981-10-3476-3_2.
Der volle Inhalt der QuelleWedge, Daniel, Du Huynh und Peter Kovesi. „Motion Guided Video Sequence Synchronization“. In Computer Vision – ACCV 2006, 832–41. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11612704_83.
Der volle Inhalt der QuelleBazin, Jean-Charles, und Alexander Sorkine-Hornung. „ActionSnapping: Motion-Based Video Synchronization“. In Computer Vision – ECCV 2016, 155–69. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-46454-1_10.
Der volle Inhalt der QuelleFurht, Borko, Stephen W. Smoliar und HongJiang Zhang. „Multimedia Networking and Synchronization“. In Video and Image Processing in Multimedia Systems, 33–57. Boston, MA: Springer US, 1995. http://dx.doi.org/10.1007/978-1-4615-2277-5_2.
Der volle Inhalt der QuelleKryston, Kevin, Eric Novotny, Ralf Schmälzle und Ron Tamborini. „Social Demand in Video Games and the Synchronization Theory of Flow“. In Video Games, 161–77. New York, NY : Routledge, 2018. | Series: Electronic media research series: Routledge, 2018. http://dx.doi.org/10.4324/9781351235266-10.
Der volle Inhalt der QuelleCai, Ying, und Grammati E. Pantziou. „A Synchronization Mechanism for Multimedia Presentation“. In Multimedia Communications and Video Coding, 157–62. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_20.
Der volle Inhalt der QuelleZuniga, Gabriel, und Ephraim Feig. „Synchronization Issues on Software MPEG Playback Systems“. In Multimedia Communications and Video Coding, 141–45. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_18.
Der volle Inhalt der QuelleLiu, Changdong, Yong Xie, Myung J. Lee und Tarek N. Saadawi. „Adaptive Synchronization in Real-Time Multimedia Applications“. In Multimedia Communications and Video Coding, 147–56. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_19.
Der volle Inhalt der QuelleBulterman, Dick C. A., und Robert Liere. „Multimedia synchronization and UNIX“. In Network and Operating System Support for Digital Audio and Video, 105–19. Berlin, Heidelberg: Springer Berlin Heidelberg, 1992. http://dx.doi.org/10.1007/3-540-55639-7_10.
Der volle Inhalt der QuelleRothermel, Kurt, und Gabriel Dermler. „Synchronization in Joint-Viewing environments“. In Network and Operating System Support for Digital Audio and Video, 106–18. Berlin, Heidelberg: Springer Berlin Heidelberg, 1993. http://dx.doi.org/10.1007/3-540-57183-3_10.
Der volle Inhalt der QuelleKonferenzberichte zum Thema "Video synchronization"
Waingankar, P., und D. Valsan. „Audio-video synchronization“. In the International Conference & Workshop. New York, New York, USA: ACM Press, 2011. http://dx.doi.org/10.1145/1980022.1980068.
Der volle Inhalt der QuelleSung, Chih-Ta S. „MPEG audio-video synchronization“. In Electronic Imaging '97, herausgegeben von Sethuraman Panchanathan und Frans Sijstermans. SPIE, 1997. http://dx.doi.org/10.1117/12.263515.
Der volle Inhalt der QuelleMediavilla, Ricardo. „Automatic test station for network synchronization performance characterization“. In Voice, Video, and Data Communications, herausgegeben von John M. Senior, Robert A. Cryan und Chunming Qiao. SPIE, 1997. http://dx.doi.org/10.1117/12.290373.
Der volle Inhalt der QuelleShankar, Sukrit, Joan Lasenby und Anil Kokaram. „Warping trajectories for video synchronization“. In the 4th ACM/IEEE international workshop. New York, New York, USA: ACM Press, 2013. http://dx.doi.org/10.1145/2510650.2510654.
Der volle Inhalt der QuelleLin, Eugene T., und Edward J. Delp III. „Temporal synchronization in video watermarking“. In Electronic Imaging 2002, herausgegeben von Edward J. Delp III und Ping W. Wong. SPIE, 2002. http://dx.doi.org/10.1117/12.465310.
Der volle Inhalt der QuelleStokking, Hans, Pablo Cesar, Fernando Boronat und Mario Montagud. „Media Synchronization Workshop“. In TVX'15: ACM International Conference on Interactive Experiences for TV and Online Video. New York, NY, USA: ACM, 2015. http://dx.doi.org/10.1145/2745197.2745699.
Der volle Inhalt der QuelleWieschollek, Patrick, Ido Freeman und Hendrik P. A. Lensch. „Learning Robust Video Synchronization without Annotations“. In 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA). IEEE, 2017. http://dx.doi.org/10.1109/icmla.2017.0-173.
Der volle Inhalt der QuelleWu, Yuanyuan, Xiaohai He und Truong Q. Nguyen. „Subframe video synchronization by matching trajectories“. In ICASSP 2013 - 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2013. http://dx.doi.org/10.1109/icassp.2013.6638060.
Der volle Inhalt der QuelleLi, Mu, und Vishal Monga. „Twofold video hashing with automatic synchronization“. In 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 2014. http://dx.doi.org/10.1109/icip.2014.7026085.
Der volle Inhalt der QuelleYang, Ming, Nikolaos Bourbakis, Zizhong Chen und Monica Trifas. „An Efficient Audio-Video Synchronization Methodology“. In Multimedia and Expo, 2007 IEEE International Conference on. IEEE, 2007. http://dx.doi.org/10.1109/icme.2007.4284763.
Der volle Inhalt der Quelle