Literatura científica selecionada sobre o tema "Video synchronization"
Crie uma referência precisa em APA, MLA, Chicago, Harvard, e outros estilos
Consulte a lista de atuais artigos, livros, teses, anais de congressos e outras fontes científicas relevantes para o tema "Video synchronization".
Ao lado de cada fonte na lista de referências, há um botão "Adicionar à bibliografia". Clique e geraremos automaticamente a citação bibliográfica do trabalho escolhido no estilo de citação de que você precisa: APA, MLA, Harvard, Chicago, Vancouver, etc.
Você também pode baixar o texto completo da publicação científica em formato .pdf e ler o resumo do trabalho online se estiver presente nos metadados.
Artigos de revistas sobre o assunto "Video synchronization"
EL-Sallam, Amar A., e Ajmal S. Mian. "Correlation based speech-video synchronization". Pattern Recognition Letters 32, n.º 6 (abril de 2011): 780–86. http://dx.doi.org/10.1016/j.patrec.2011.01.001.
Texto completo da fonteLin, E. T., e E. J. Delp. "Temporal Synchronization in Video Watermarking". IEEE Transactions on Signal Processing 52, n.º 10 (outubro de 2004): 3007–22. http://dx.doi.org/10.1109/tsp.2004.833866.
Texto completo da fonteFu, Jia Bing, e He Wei Yu. "Audio-Video Synchronization Method Based on Playback Time". Applied Mechanics and Materials 300-301 (fevereiro de 2013): 1677–80. http://dx.doi.org/10.4028/www.scientific.net/amm.300-301.1677.
Texto completo da fonteLi, Xiao Ni, He Xin Chen e Da Zhong Wang. "Research on Audio-Video Synchronization Coding Based on Mode Selection in H.264". Applied Mechanics and Materials 182-183 (junho de 2012): 701–5. http://dx.doi.org/10.4028/www.scientific.net/amm.182-183.701.
Texto completo da fonteLiu, Yiguang, Menglong Yang e Zhisheng You. "Video synchronization based on events alignment". Pattern Recognition Letters 33, n.º 10 (julho de 2012): 1338–48. http://dx.doi.org/10.1016/j.patrec.2012.02.009.
Texto completo da fonteMu Li e Vishal Monga. "Twofold Video Hashing With Automatic Synchronization". IEEE Transactions on Information Forensics and Security 10, n.º 8 (agosto de 2015): 1727–38. http://dx.doi.org/10.1109/tifs.2015.2425362.
Texto completo da fonteZhou, Zhongyi, Anran Xu e Koji Yatani. "SyncUp". Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 5, n.º 3 (9 de setembro de 2021): 1–25. http://dx.doi.org/10.1145/3478120.
Texto completo da fonteYang, Shu Zhen, Guang Lin Chu e Ming Wang. "A Study on Parallel Processing Video Splicing System with Multi-Processor". Applied Mechanics and Materials 198-199 (setembro de 2012): 304–9. http://dx.doi.org/10.4028/www.scientific.net/amm.198-199.304.
Texto completo da fonteKwon, Ohsung. "Class Analysis Method Using Video Synchronization Algorithm". Journal of The Korean Association of Information Education 19, n.º 4 (30 de dezembro de 2015): 441–48. http://dx.doi.org/10.14352/jkaie.2015.19.4.441.
Texto completo da fonteChen, T., H. P. Graf e K. Wang. "Lip synchronization using speech-assisted video processing". IEEE Signal Processing Letters 2, n.º 4 (abril de 1995): 57–59. http://dx.doi.org/10.1109/97.376913.
Texto completo da fonteTeses / dissertações sobre o assunto "Video synchronization"
Wedge, Daniel John. "Video sequence synchronization". University of Western Australia. School of Computer Science and Software Engineering, 2008. http://theses.library.uwa.edu.au/adt-WU2008.0084.
Texto completo da fonteYang, Hsueh-szu, e Benjamin Kupferschmidt. "Time Stamp Synchronization in Video Systems". International Foundation for Telemetering, 2010. http://hdl.handle.net/10150/605988.
Texto completo da fonteSynchronized video is crucial for data acquisition and telecommunication applications. For real-time applications, out-of-sync video may cause jitter, choppiness and latency. For data analysis, it is important to synchronize multiple video channels and data that are acquired from PCM, MIL-STD-1553 and other sources. Nowadays, video codecs can be easily obtained to play most types of video. However, a great deal of effort is still required to develop the synchronization methods that are used in a data acquisition system. This paper will describe several methods that TTC has adopted in our system to improve the synchronization of multiple data sources.
Gaskill, David M. "TECHNIQUES FOR SYNCHRONIZING THERMAL ARRAY CHART RECORDERS TO VIDEO". International Foundation for Telemetering, 1992. http://hdl.handle.net/10150/608901.
Texto completo da fonteVideo tape is becoming more and more popular for storing and analyzing missions. Video tape is inexpensive, it can hold a two hour test, and it can be edited and manipulated by easily available consumer electronics equipment. Standard technology allows each frame to be time stamped with SMPTE code, so that any point in the mission can be displayed on a CRT. To further correlate data from multiple acquisition systems, the SMPTE code can be derived from IRIG using commercially available code converters. Unfortunately, acquiring and storing analog data has not been so easy. Typically, analog signals from various sensors are coded, transmitted, decoded and sent to a chart recorder. Since chart recorders cannot normally store an entire mission internally, or time stamp each data value, it is very difficult for an analyst to accurately correlate analog data to an individual video frame. Normally the only method is to note the time stamp on the video frame and unroll the chart to the appropriate second or minute, depending on the code used, noted in the margin, and estimate the frame location as a percentage of the time code period. This is very inconvenient if the telemetrist is trying to establish an on-line data retreival system. To make matters worse, the methods of presentation are very different, chart paper as opposed to a CRT, and require the analyst to shift focus constantly. For these reasons, many telemetry stations do not currently have a workable plan to integrate analog and video subsystems even though it is now generally agreed that such integration is ultimately desirable.
Daami, Mourad. "Synchronization control of coded video streams, algorithms and implementation". Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk2/ftp04/mq26314.pdf.
Texto completo da fonteAbraham, Justin Kuruvilla. "Study of the TR Synchronization and Video Conversion Unit". Master's thesis, University of Cape Town, 2012. http://hdl.handle.net/11427/14137.
Texto completo da fonteYilmaz, Ayhan. "Robust Video Transmission Using Data Hiding". Master's thesis, METU, 2003. http://etd.lib.metu.edu.tr/upload/1093509/index.pdf.
Texto completo da fontehence, improves the reconstruction video quality without significant extra channel utilization. A complete error resilient video transmission codec is proposed, utilizing imperceptible embedded information for combined detecting, resynchronization and reconstruction of the errors and lost data. The data, which is imperceptibly embedded into the video itself at the encoder, is extracted from the video at the decoder side to be utilized in error concealment. A spatial domain error recovery technique, which hides edge orientation information of a block, and a resynchronization technique, which embeds bit length of a block into other blocks are combined, as well as some parity information about the hidden data, to conceal channel errors on intra-coded frames of a video sequence. The errors on inter-coded frames are basically recovered by hiding motion vector information along with a checksum into the next frames. The simulation results show that the proposed approach performs superior to conventional approaches for concealing the errors in binary symmetric channels, especially for higher bit rates and error rates.
Potetsianakis, Emmanouil. "Enhancing video applications through timed metadata". Electronic Thesis or Diss., Université Paris-Saclay (ComUE), 2019. http://www.theses.fr/2019SACLT029.
Texto completo da fonteVideo recording devices are often equipped with sensors (smartphones for example, with GPS receiver, gyroscope etc.), or used in settings where sensors are present (e.g. monitor cameras, in areas with temperature and/or humidity sensors). As a result, many systems process and distribute video together with timed metadata streams, often sourced as User-Generated Content. Video delivery has been thoroughly studied, however timed metadata streams have varying characteristics and forms, thus a consistent and effective way to handle them in conjunction with the video streams does not exist. In this Thesis we study ways to enhance video applications through timed metadata. We define as timed metadata all the non-audiovisual data recorded or produced, that are relevant to a specific time on the media timeline. ”Enhancing” video applications has a double meaning, and this work consists of two respective parts. First, using the timed metadata to extend the capabilities of multimedia applications, by introducing novel functionalities. Second, using the timed metadata to improve the content delivery for such applications. To extend multimedia applications, we have taken an exploratory approach, and we demonstrate two use cases with application examples. In the first case, timed metadata is used as input for generating content, and in the second, it is used to extend the navigational capabilities for the underlying multimedia content. By designing and implementing two different application scenarios we were able to identify the potential and limitations of video systems with timed metadata. We use the findings from the first part, to work from the perspective of enhancing video applications, by using the timed metadata to improve delivery of the content. More specifically, we study the use of timed metadata for multi-variable adaptation in multi-view video delivery - and we test our proposals on one of the platforms developed previously. Our final contribution is a buffering scheme for synchronous and lowlatency playback in live streaming systems
Carranza, López José Camilo. "On the synchronization of two metronomes and their related dynamics /". Ilha Solteira, 2017. http://hdl.handle.net/11449/151204.
Texto completo da fonteResumo: Nesta tese são investigadas, teórica e experimentalmente, a sincronização em fase e a sincronização em anti-fase de dois metrônomos oscilando sobre uma base móvel, a partir de um modelo aqui proposto. Uma descrição do funcionamento do mecanismo de escapamento dos metrônomos é feita, junto a um estudo da relação entre este e o oscilador de van der Pol. Também uma aproximação experimental do valor do amortecimento do metrônomo é fornecida. A frequência instantânea das respostas, numérica e experimental, do sistema é usada na analise. A diferença de outros trabalhos prévios, os dados experimentais têm sido adquiridos usando vídeos dos experimentos e extraídos com ajuda do software Tracker. Para investigar a relação entre as condições iniciais do sistema e seu estado final de sincronização, foram usados mapas bidimensionais chamados ‘basins of attraction’. A relação entre o modelo proposto e um modelo prévio também é mostrada. Encontrou-se que os parâmetros relevantes em relação a ambos os tipos de sincronização são a razão entre a massa do metrônomo e a massa da base, e o amortecimento do sistema. Tem-se encontrado, tanto experimental quanto teoricamente, que a frequência de oscilação dos metrônomos aumenta quando o sistema sincroniza-se em fase, e se mantém a mesma de um metrônomo isolado quando o sistema sincroniza-se em anti-fase. A partir de simulações numéricas encontrou-se que, em geral, incrementos no amortecimento do sistema levam ao sistema se sincronizar mais em fase d... (Resumo completo, clicar acesso eletrônico abaixo)
Doutor
Wehbe, Hassan. "Synchronisation automatique d'un contenu audiovisuel avec un texte qui le décrit". Thesis, Toulouse 3, 2016. http://www.theses.fr/2016TOU30104/document.
Texto completo da fonteWe address the problem of automatic synchronization of an audiovisual content with a procedural text that describes it. The strategy consists in extracting pieces of information about the structure from both contents, and in matching them depending on their types. We propose two video analysis tools that respectively extract: * Limits of events of interest using an approach inspired by dictionary quantization. * Segments that enclose a repeated action based on the YIN frequency analysis method. We then propose a synchronization system that merges results coming from these tools in order to establish links between textual instructions and the corresponding video segments. To do so, a "Confidence Matrix" is built and recursively processed in order to identify these links in respect with their reliability
Carranza, López José Camilo [UNESP]. "On the synchronization of two metronomes and their related dynamics". Universidade Estadual Paulista (UNESP), 2017. http://hdl.handle.net/11449/151204.
Texto completo da fonteApproved for entry into archive by Luiz Galeffi (luizgaleffi@gmail.com) on 2017-07-26T18:31:30Z (GMT) No. of bitstreams: 1 carranzalopez_jc_dr_ilha.pdf: 11035322 bytes, checksum: efe400c07b13cabff41e927078789c59 (MD5)
Made available in DSpace on 2017-07-26T18:31:30Z (GMT). No. of bitstreams: 1 carranzalopez_jc_dr_ilha.pdf: 11035322 bytes, checksum: efe400c07b13cabff41e927078789c59 (MD5) Previous issue date: 2017-06-05
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Nesta tese são investigadas, teórica e experimentalmente, a sincronização em fase e a sincronização em anti-fase de dois metrônomos oscilando sobre uma base móvel, a partir de um modelo aqui proposto. Uma descrição do funcionamento do mecanismo de escapamento dos metrônomos é feita, junto a um estudo da relação entre este e o oscilador de van der Pol. Também uma aproximação experimental do valor do amortecimento do metrônomo é fornecida. A frequência instantânea das respostas, numérica e experimental, do sistema é usada na analise. A diferença de outros trabalhos prévios, os dados experimentais têm sido adquiridos usando vídeos dos experimentos e extraídos com ajuda do software Tracker. Para investigar a relação entre as condições iniciais do sistema e seu estado final de sincronização, foram usados mapas bidimensionais chamados ‘basins of attraction’. A relação entre o modelo proposto e um modelo prévio também é mostrada. Encontrou-se que os parâmetros relevantes em relação a ambos os tipos de sincronização são a razão entre a massa do metrônomo e a massa da base, e o amortecimento do sistema. Tem-se encontrado, tanto experimental quanto teoricamente, que a frequência de oscilação dos metrônomos aumenta quando o sistema sincroniza-se em fase, e se mantém a mesma de um metrônomo isolado quando o sistema sincroniza-se em anti-fase. A partir de simulações numéricas encontrou-se que, em geral, incrementos no amortecimento do sistema levam ao sistema se sincronizar mais em fase do que em anti-fase. Adicionalmente se encontrou que, para dado valor de amortecimento, diminuir a massa da base leva a uma situação em que a sincronização em anti-fase é mais comum do que a sincronização em fase.
This thesis concerns a theoretical and experimental investigation into the synchronization of two coupled metronomes. A simplified model is proposed to study in-phase and anti-phase synchronization of two metronomes oscillating on a mobile base. A description of the escapement mechanism driving metronomes is given and its relationship with the van der Pol oscillator is discussed. Also an experimental value for the damping in the metronome is determined. The instantaneous frequency of the responses from both numerical and experimental data is used in the analysis. Unlike previous studies, measurements are made using videos and the time domain responses of the metronomes extracted by means of tracker software. Basins of attraction are used to investigate the relationship between initial conditions, parameters and both final synchronization states. The relationship between the model and a previous pendulum model is also shown. The key parameters concerning both kind of synchronization have been found to be the mass ratio between the metronome mass and the base mass, and the damping in the system. It has been shown, both theoretically and experimentally, that the frequency of oscillation of the metronomes increases when the system reaches in-phase synchronization, and is the same as an isolated metronome when the system synchronizes in anti-phase. From numerical simulations, it has been found that, in general, increasing damping leads the system to synchronize more in-phase than in anti-phase. It has also been found that, for a given damping value, decreasing the mass of the base results in the situation where anti-phase synchronization is more common than in-phase synchronization.
Livros sobre o assunto "Video synchronization"
Rona, Jeffrey C. Synchronization from reel to reel: A complete guide for the synchronization of audio, film & video. Milwaukee, WI: Hal Leonard Publishing Corporation, 1990.
Encontre o texto completo da fonteS, Schiff Ronny, e Wilkinson Scott R. 1953-, eds. Synchronization, from reel to reel: A complete guide for the synchronization of audio, film & video. Milwaukee, WI: H. Leonard Pub., Corp., 1989.
Encontre o texto completo da fonteHawkins, Stan. Aesthetics and Hyperembodiment in Pop Videos. Editado por John Richardson, Claudia Gorbman e Carol Vernallis. Oxford University Press, 2013. http://dx.doi.org/10.1093/oxfordhb/9780199733866.013.002.
Texto completo da fonteST 318:2015: Synchronization of 59.94- or 50-Hz Related Video and Audio Systems in Analog and Digital Areas — Reference Signals. 3 Barker Avenue., White Plains, NY 10601: The Society of Motion Picture and Television Engineers SMPTE, 2015. http://dx.doi.org/10.5594/smpte.st318.2015.
Texto completo da fonteBarrière, Jean-Baptiste, e Aleksi Barrière. When Music Unfolds into Image. Editado por Yael Kaduri. Oxford University Press, 2016. http://dx.doi.org/10.1093/oxfordhb/9780199841547.013.39.
Texto completo da fonteCapítulos de livros sobre o assunto "Video synchronization"
Wang, Xue, e Qing Wang. "Video Synchronization with Trajectory Pulse". In Communications in Computer and Information Science, 12–19. Singapore: Springer Singapore, 2016. http://dx.doi.org/10.1007/978-981-10-3476-3_2.
Texto completo da fonteWedge, Daniel, Du Huynh e Peter Kovesi. "Motion Guided Video Sequence Synchronization". In Computer Vision – ACCV 2006, 832–41. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11612704_83.
Texto completo da fonteBazin, Jean-Charles, e Alexander Sorkine-Hornung. "ActionSnapping: Motion-Based Video Synchronization". In Computer Vision – ECCV 2016, 155–69. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-46454-1_10.
Texto completo da fonteFurht, Borko, Stephen W. Smoliar e HongJiang Zhang. "Multimedia Networking and Synchronization". In Video and Image Processing in Multimedia Systems, 33–57. Boston, MA: Springer US, 1995. http://dx.doi.org/10.1007/978-1-4615-2277-5_2.
Texto completo da fonteKryston, Kevin, Eric Novotny, Ralf Schmälzle e Ron Tamborini. "Social Demand in Video Games and the Synchronization Theory of Flow". In Video Games, 161–77. New York, NY : Routledge, 2018. | Series: Electronic media research series: Routledge, 2018. http://dx.doi.org/10.4324/9781351235266-10.
Texto completo da fonteCai, Ying, e Grammati E. Pantziou. "A Synchronization Mechanism for Multimedia Presentation". In Multimedia Communications and Video Coding, 157–62. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_20.
Texto completo da fonteZuniga, Gabriel, e Ephraim Feig. "Synchronization Issues on Software MPEG Playback Systems". In Multimedia Communications and Video Coding, 141–45. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_18.
Texto completo da fonteLiu, Changdong, Yong Xie, Myung J. Lee e Tarek N. Saadawi. "Adaptive Synchronization in Real-Time Multimedia Applications". In Multimedia Communications and Video Coding, 147–56. Boston, MA: Springer US, 1996. http://dx.doi.org/10.1007/978-1-4613-0403-6_19.
Texto completo da fonteBulterman, Dick C. A., e Robert Liere. "Multimedia synchronization and UNIX". In Network and Operating System Support for Digital Audio and Video, 105–19. Berlin, Heidelberg: Springer Berlin Heidelberg, 1992. http://dx.doi.org/10.1007/3-540-55639-7_10.
Texto completo da fonteRothermel, Kurt, e Gabriel Dermler. "Synchronization in Joint-Viewing environments". In Network and Operating System Support for Digital Audio and Video, 106–18. Berlin, Heidelberg: Springer Berlin Heidelberg, 1993. http://dx.doi.org/10.1007/3-540-57183-3_10.
Texto completo da fonteTrabalhos de conferências sobre o assunto "Video synchronization"
Waingankar, P., e D. Valsan. "Audio-video synchronization". In the International Conference & Workshop. New York, New York, USA: ACM Press, 2011. http://dx.doi.org/10.1145/1980022.1980068.
Texto completo da fonteSung, Chih-Ta S. "MPEG audio-video synchronization". In Electronic Imaging '97, editado por Sethuraman Panchanathan e Frans Sijstermans. SPIE, 1997. http://dx.doi.org/10.1117/12.263515.
Texto completo da fonteMediavilla, Ricardo. "Automatic test station for network synchronization performance characterization". In Voice, Video, and Data Communications, editado por John M. Senior, Robert A. Cryan e Chunming Qiao. SPIE, 1997. http://dx.doi.org/10.1117/12.290373.
Texto completo da fonteShankar, Sukrit, Joan Lasenby e Anil Kokaram. "Warping trajectories for video synchronization". In the 4th ACM/IEEE international workshop. New York, New York, USA: ACM Press, 2013. http://dx.doi.org/10.1145/2510650.2510654.
Texto completo da fonteLin, Eugene T., e Edward J. Delp III. "Temporal synchronization in video watermarking". In Electronic Imaging 2002, editado por Edward J. Delp III e Ping W. Wong. SPIE, 2002. http://dx.doi.org/10.1117/12.465310.
Texto completo da fonteStokking, Hans, Pablo Cesar, Fernando Boronat e Mario Montagud. "Media Synchronization Workshop". In TVX'15: ACM International Conference on Interactive Experiences for TV and Online Video. New York, NY, USA: ACM, 2015. http://dx.doi.org/10.1145/2745197.2745699.
Texto completo da fonteWieschollek, Patrick, Ido Freeman e Hendrik P. A. Lensch. "Learning Robust Video Synchronization without Annotations". In 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA). IEEE, 2017. http://dx.doi.org/10.1109/icmla.2017.0-173.
Texto completo da fonteWu, Yuanyuan, Xiaohai He e Truong Q. Nguyen. "Subframe video synchronization by matching trajectories". In ICASSP 2013 - 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2013. http://dx.doi.org/10.1109/icassp.2013.6638060.
Texto completo da fonteLi, Mu, e Vishal Monga. "Twofold video hashing with automatic synchronization". In 2014 IEEE International Conference on Image Processing (ICIP). IEEE, 2014. http://dx.doi.org/10.1109/icip.2014.7026085.
Texto completo da fonteYang, Ming, Nikolaos Bourbakis, Zizhong Chen e Monica Trifas. "An Efficient Audio-Video Synchronization Methodology". In Multimedia and Expo, 2007 IEEE International Conference on. IEEE, 2007. http://dx.doi.org/10.1109/icme.2007.4284763.
Texto completo da fonte