Academic literature on the topic 'Audio-visual attention'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Audio-visual attention.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "Audio-visual attention"
Chen, Yanxiang, Tam V. Nguyen, Mohan Kankanhalli, Jun Yuan, Shuicheng Yan, and Meng Wang. "Audio Matters in Visual Attention." IEEE Transactions on Circuits and Systems for Video Technology 24, no. 11 (November 2014): 1992–2003. http://dx.doi.org/10.1109/tcsvt.2014.2329380.
Full textLee, Yong-Hyeok, Dong-Won Jang, Jae-Bin Kim, Rae-Hong Park, and Hyung-Min Park. "Audio–Visual Speech Recognition Based on Dual Cross-Modality Attentions with the Transformer Model." Applied Sciences 10, no. 20 (October 17, 2020): 7263. http://dx.doi.org/10.3390/app10207263.
Full textIwaki, Sunao, Mitsuo Tonoike, Masahiko Yamaguchi, and Takashi Hamada. "Modulation of extrastriate visual processing by audio-visual intermodal selective attention." NeuroImage 11, no. 5 (May 2000): S21. http://dx.doi.org/10.1016/s1053-8119(00)90956-x.
Full textNAGASAKI, Yoshiki, Masaki HAYASHI, Naoshi KANEKO, and Yoshimitsu AOKI. "Temporal Cross-Modal Attention for Audio-Visual Event Localization." Journal of the Japan Society for Precision Engineering 88, no. 3 (March 5, 2022): 263–68. http://dx.doi.org/10.2493/jjspe.88.263.
Full textXuan, Hanyu, Zhenyu Zhang, Shuo Chen, Jian Yang, and Yan Yan. "Cross-Modal Attention Network for Temporal Inconsistent Audio-Visual Event Localization." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 01 (April 3, 2020): 279–86. http://dx.doi.org/10.1609/aaai.v34i01.5361.
Full textIwaki, Sunao. "Audio-visual intermodal orientation of attention modulates task-specific extrastriate visual processing." Neuroscience Research 68 (January 2010): e269. http://dx.doi.org/10.1016/j.neures.2010.07.1195.
Full textKeitel, Christian, and Matthias M. Müller. "Audio-visual synchrony and feature-selective attention co-amplify early visual processing." Experimental Brain Research 234, no. 5 (August 1, 2015): 1221–31. http://dx.doi.org/10.1007/s00221-015-4392-8.
Full textZhu, Hao, Man-Di Luo, Rui Wang, Ai-Hua Zheng, and Ran He. "Deep Audio-visual Learning: A Survey." International Journal of Automation and Computing 18, no. 3 (April 15, 2021): 351–76. http://dx.doi.org/10.1007/s11633-021-1293-0.
Full textRan, Yue, Hongying Tang, Baoqing Li, and Guohui Wang. "Self-Supervised Video Representation and Temporally Adaptive Attention for Audio-Visual Event Localization." Applied Sciences 12, no. 24 (December 9, 2022): 12622. http://dx.doi.org/10.3390/app122412622.
Full textZhao, Sicheng, Yunsheng Ma, Yang Gu, Jufeng Yang, Tengfei Xing, Pengfei Xu, Runbo Hu, Hua Chai, and Kurt Keutzer. "An End-to-End Visual-Audio Attention Network for Emotion Recognition in User-Generated Videos." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 01 (April 3, 2020): 303–11. http://dx.doi.org/10.1609/aaai.v34i01.5364.
Full textDissertations / Theses on the topic "Audio-visual attention"
Sharma, Dinkar. "Effects of attention on audio-visual speech." Thesis, University of Reading, 1989. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.329379.
Full textRen, Reede. "Audio-visual football video analysis, from structure detection to attention analysis." Thesis, Connect to e-thesis. Move to record for print version, 2008. http://theses.gla.ac.uk/77/.
Full textPh.D. thesis submitted to the Faculty of Information and Mathematical Sciences, Department of Computing Science, University of Glasgow, 2008. Includes bibliographical references. Print version also available.
Song, Guanghan. "Effect of sound in videos on gaze : contribution to audio-visual saliency modelling." Thesis, Grenoble, 2013. http://www.theses.fr/2013GRENT013/document.
Full textHumans receive large quantity of information from the environment with sight and hearing. To help us to react rapidly and properly, there exist mechanisms in the brain to bias attention towards particular regions, namely the salient regions. This attentional bias is not only influenced by vision, but also influenced by audio-visual interaction. According to existing literature, the visual attention can be studied towards eye movements, however the sound effect on eye movement in videos is little known. The aim of this thesis is to investigate the influence of sound in videos on eye movement and to propose an audio-visual saliency model to predict salient regions in videos more accurately. For this purpose, we designed a first audio-visual experiment of eye tracking. We created a database of short video excerpts selected from various films. These excerpts were viewed by participants either with their original soundtrack (AV condition), or without soundtrack (V condition). We analyzed the difference of eye positions between participants with AV and V conditions. The results show that there does exist an effect of sound on eye movement and the effect is greater for the on-screen speech class. Then, we designed a second audio-visual experiment with thirteen classes of sound. Through comparing the difference of eye positions between participants with AV and V conditions, we conclude that the effect of sound is different depending on the type of sound, and the classes with human voice (i.e. speech, singer, human noise and singers classes) have the greatest effect. More precisely, sound source significantly attracted eye position only when the sound was human voice. Moreover, participants with AV condition had a shorter average duration of fixation than with V condition. Finally, we proposed a preliminary audio-visual saliency model based on the findings of the above experiments. In this model, two fusion strategies of audio and visual information were described: one for speech sound class, and one for musical instrument sound class. The audio-visual fusion strategies defined in the model improves its predictability with AV condition
D'AMELIO, ALESSANDRO. "A STOCHASTIC FORAGING MODEL OF ATTENTIVE EYE GUIDANCE ON DYNAMIC STIMULI." Doctoral thesis, Università degli Studi di Milano, 2021. http://hdl.handle.net/2434/816678.
Full textMARINI, FRANCESCO. "Attentional control guides the strategic filtering of potential distraction as revealed by behavior and Fmri." Doctoral thesis, Università degli Studi di Milano-Bicocca, 2014. http://hdl.handle.net/10281/50236.
Full textKristal-Ern, Alfred. "Can sound be used to effectively direct players' attention in a visual gameplay oriented task?" Thesis, Luleå tekniska universitet, Medier ljudteknik och upplevelseproduktion och teater, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:ltu:diva-63621.
Full textLind, Erik. "The role of an audio-visual attentional stimulus in influencing affective responses during graded cycling exercise." [Ames, Iowa : Iowa State University], 2008.
Find full textChen, Ya-Ching, and 陳雅靖. "The Effect of Audio Rhythm on Visual Attention." Thesis, 2004. http://ndltd.ncl.edu.tw/handle/73567288538165563347.
Full text國立交通大學
應用藝術研究所
92
Vision and audition are the two major modalities we use to receive outside messages. There is evidence indicating that vision and audition do not function independently. A better understanding of the interaction between these two senses is of great value to a designer. The focus of our study is the effect of audio rhythm on visual attention. We hypothesize that if the visual and the auditory stimuli are synchronized, viewer can be cued by the auditory rhythm and would pay more attention to the synchronized visual event. The aim of this study is to test this hypothesis. We used rapid serial visual presentation (RSVP) as a means to probe subjects’ visual spatial attention on a given spot of the screen. Two RSVP streams of different rhythms were presented to the viewer on each trial. One of the RSVP stream is synchronized with an auditory stimulus while another is not. If the viewer’s attention can be guided by the auditory rhythm, one would predict that the performance in the synchronized RSVP be better than that in another stream. The results show that: 1. The auditory rhythm, while being task-irrelevant by itself, does cue subjects’ attention to the synchronized visual event. 2. The power of cueing visual events is critically dependent upon the acoustic properties of the auditory stimulus. 3. Some rhythms are more potent than others in binding visual and auditory events. 4. As most viewers were not aware that one of the RSVP streams was synchronized to the auditory event, we believe the enhancement effect by synchronization occurs at an early, preconscious level.
Lin, Yan-Bo, and 林彥伯. "Cross-Modality Co-Attention for Audio-Visual Event Localization." Thesis, 2019. http://ndltd.ncl.edu.tw/handle/mamwpe.
Full text國立臺灣大學
電信工程學研究所
107
Audio-visual event localization requires one to identify the event labelacross video frames by jointly observing visual and audio information. To address this task, we propose a deep neural network named Audio-Visual sequence-to-sequence dual network (AVSDN). By jointly taking both audio and visual features at each time segment as inputs, our proposed model learns global and local event information in a sequence to sequence manner. Besides, we also propose a deep learning framework of cross-modality co-attention for audio-visual event localization. The co-attention framework can be applied on existing methods and AVSDN. Our co-attention modelis able to exploit intra and inter-frame visual information, with audio features jointly observed to perform co-attention over the above three modalities.With visual, temporal, and audio information observed across consecutive video frames, our model achieves promising capability in extracting informative spatial/temporal features for improved event localization. Moreover,our model is able to produce instance-level attention, which would identify image regions at the instance level which are associated with the sound/event of interest. Experiments on a benchmark dataset confirm the effectiveness of our proposed framework,with ablation studies performed to verify the design of our propose network model.
ZHOU, YONG-FENG, and 周永豐. "Combining Brainwave Instrument to Develop Visual and Audio Attention Test System: Apply to Children with Attention Deficit Hyperactivity Disorder." Thesis, 2017. http://ndltd.ncl.edu.tw/handle/jsbfv5.
Full text朝陽科技大學
資訊工程系
105
In this study, we focus on the development of visual and auditory attention test system (VAAT), which is used to evaluate the visual and auditory focus. The purpose of this study is to verify the concurrent validity of VAAT in this study, and to explore the effects of ADHD (Attention deficit hyperactivity disorder) on VAAT and Conners Continuous Performance Test 3rd Edition (CPT 3) and the Conners Kiddie Continuous Performance Test 2rd Edition, K-CPT 2). We also study the correlation between the scores of the parameters and the state of the brain waves. The subjects were by the 16 ADHD children and 2 other symptomatic and healthy children consisting of preschool group, the age from 4 to 7 years old; 11 ADHD children and 6 other symptomatic children consisting of kid group (age from 8 to 13 years old). Each child in the preschool group received K-CPT2 and VAAT. Each child in the kid group received CPT 3 and VAAT. To use the Neurosky brain cube mobile version (MindWave Mobile) measures the brain waves. The results of this study show that VAAT and K-CPT 2 have significantly positively correlated with more than half of the parameters. VAAT and CPT 3 have significantly positively correlated with most of parameters. Therefore, VAAT has good concurrent validity, it can be applied to the attention assessment. In compared with the children with ADHD of the visual and auditory ability, the preschool group of children with ADHD have poor auditory ability to identify target and non-target ability, and slower auditory response than the visual ability; the kid group of children with ADHD in visual and audio ability comparison, easy to omission auditory target and slower auditory response. The children with ADHD do right and woring brain wave difference, preschool group of children with ADHD when the Beta wave is reduced, showing inattention, and easy to do wrong (mistakenly press) non-target, but kid group of children with ADHD were not found in the test, because the number of cases is too small. In the analysis of the correlation between the parameters and brain wave performance, it was found that preschool group of children with ADHD in the identification ability and Alpha wave were negative correlattion, and the identification ability was lower, the Alpha wave will reduce and showed a tense phenomenon. The change of the reaction block and Delta wave were a negative correlation, representing the worse performance of change of the reaction block, the Delta wave will reduce, showed inattention and cognitive decline phenomenon. The kid group of children with ADHD in error rate, the Theta wave and Delta wave were negative correlation, indicating that the higher error rate score, the Theta and Delta wave will reduce, showed inattention and cognitive decline phenomenon. Preschool group with different types of ADHD children found in the K-CPT 2 performance, inattentive type was detected inattentive problem; impulsive type was detected vigilant problem, but wasn’t detected impulsive problem; combined type was detected inattentive problem, but wasn’t detected impulsive problem. Kid group with different types of ADHD children found in the CPT 3 performance, inattentive type was detected inattentive and sustained attention problems; impulsive type was detected sustained attention problem, but wasn’t detected impulsive problem; combined type was detected inattentive, sustained attention and vigilant problems, but wasn’t detected impulsive problem. This study developed VAAT not only to evaluate audio and visual attention, but also to explore the significant correlation between test performance and brain wave statistics. Therefore, if we can expand the test of normal children and ADHD children in the future, continued to verify and improve its reliability and validity. We believe by immediately and unique features of VAAT, it should be universal in children's attention tests and provide children's medical clinical application reference.
Books on the topic "Audio-visual attention"
The power of multisensory preaching and teaching: Increase attention, comprehension, and retention. Grand Rapids, Mich: Zondervan, 2008.
Find full textSchacher, Jan C. Algorithmic Spatialization. Edited by Roger T. Dean and Alex McLean. Oxford University Press, 2018. http://dx.doi.org/10.1093/oxfordhb/9780190226992.013.12.
Full textColmeiro, José. Peripheral Visions / Global Sounds. Liverpool University Press, 2018. http://dx.doi.org/10.5949/liverpool/9781786940308.001.0001.
Full textCruz, Gabriela. Grand Illusion. Oxford University Press, 2020. http://dx.doi.org/10.1093/oso/9780190915056.001.0001.
Full textBook chapters on the topic "Audio-visual attention"
Fang, Yinghong, Junpeng Zhang, and Cewu Lu. "Attention-Based Audio-Visual Fusion for Video Summarization." In Neural Information Processing, 328–40. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-36711-4_28.
Full textSchauerte, Boris. "Bottom-Up Audio-Visual Attention for Scene Exploration." In Cognitive Systems Monographs, 35–113. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-33796-8_3.
Full textPleshkova, Snejana, and Alexander Bekiarski. "Audio Visual Attention Models in the Mobile Robots Navigation." In New Approaches in Intelligent Image Analysis, 253–94. Cham: Springer International Publishing, 2016. http://dx.doi.org/10.1007/978-3-319-32192-9_8.
Full textLin, Yan-Bo, and Yu-Chiang Frank Wang. "Audiovisual Transformer with Instance Attention for Audio-Visual Event Localization." In Computer Vision – ACCV 2020, 274–90. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-69544-6_17.
Full textMercea, Otniel-Bogdan, Thomas Hummel, A. Sophia Koepke, and Zeynep Akata. "Temporal and Cross-modal Attention for Audio-Visual Zero-Shot Learning." In Lecture Notes in Computer Science, 488–505. Cham: Springer Nature Switzerland, 2022. http://dx.doi.org/10.1007/978-3-031-20044-1_28.
Full textSun, Zhongbo, Yannan Wang, and Li Cao. "An Attention Based Speaker-Independent Audio-Visual Deep Learning Model for Speech Enhancement." In MultiMedia Modeling, 722–28. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-37734-2_60.
Full textTzinis, Efthymios, Scott Wisdom, Tal Remez, and John R. Hershey. "AudioScopeV2: Audio-Visual Attention Architectures for Calibrated Open-Domain On-Screen Sound Separation." In Lecture Notes in Computer Science, 368–85. Cham: Springer Nature Switzerland, 2022. http://dx.doi.org/10.1007/978-3-031-19836-6_21.
Full textChen, Chin-Ling, Yung-Wen Tang, Yong-Feng Zhou, and Yue-Xun Chen. "Development of Audio and Visual Attention Assessment System in Combination with Brain Wave Instrument: Apply to Children with Attention Deficit Hyperactivity Disorder." In Advances in Intelligent Systems and Computing, 153–61. Singapore: Springer Singapore, 2017. http://dx.doi.org/10.1007/978-981-10-6487-6_19.
Full textYeromin, Mykola Borysovych, and Igor Charskykh. "Universal and Specific Codes of Cultural Context in Audio-Visual Media." In Cross-Cultural Perspectives on Technology-Enhanced Language Learning, 68–82. IGI Global, 2018. http://dx.doi.org/10.4018/978-1-5225-5463-9.ch004.
Full textAbankwah, Ruth Mpatawuwa. "Managing Audio-Visual Resources in Selected Developed and Developing Countries." In Handbook of Research on Heritage Management and Preservation, 126–49. IGI Global, 2018. http://dx.doi.org/10.4018/978-1-5225-3137-1.ch007.
Full textConference papers on the topic "Audio-visual attention"
Cheng, Shuaiyang, Xing Gao, Liang Song, and Jianbing Xiahou. "Audio-Visual Salieny Network with Audio Attention Module." In ICAIIS 2021: 2021 2nd International Conference on Artificial Intelligence and Information Systems. New York, NY, USA: ACM, 2021. http://dx.doi.org/10.1145/3469213.3470254.
Full textZhang, Wen, and Jie Shao. "Multi-Attention Audio-Visual Fusion Network for Audio Spatialization." In ICMR '21: International Conference on Multimedia Retrieval. New York, NY, USA: ACM, 2021. http://dx.doi.org/10.1145/3460426.3463624.
Full textLee, Jong-Seok, Francesca De Simone, and Touradj Ebrahimi. "Video coding based on audio-visual attention." In 2009 IEEE International Conference on Multimedia and Expo (ICME). IEEE, 2009. http://dx.doi.org/10.1109/icme.2009.5202435.
Full textLee, Jiyoung, Sunok Kim, Seungryong Kim, and Kwanghoon Sohn. "Audio-Visual Attention Networks for Emotion Recognition." In MM '18: ACM Multimedia Conference. New York, NY, USA: ACM, 2018. http://dx.doi.org/10.1145/3264869.3264873.
Full textChianese, Angelo, Vincenzo Moscato, Antonio Penta, and Antonio Picariello. "Scene Detection using Visual and Audio Attention." In 1st International ICST Conference on Ambient Media and Systems. ICST, 2008. http://dx.doi.org/10.4108/icst.ambisys2008.2828.
Full textSUGANO, Y., and S. IWAMIYA. "THE EFFECTS OF AUDIO—VISUAL SYNCHRONIZATION ON THE ATTENTION TO THE AUDIO—VISUAL MATERIALS." In MMM 2000. WORLD SCIENTIFIC, 2000. http://dx.doi.org/10.1142/9789812791993_0001.
Full textJust, N., M. Laabs, E. Unver, B. Gunel, S. Worrall, and A. M. Kondoz. "AVISION Audio and visual attention models applied to 2D and 3D audio-visual content." In 2011 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB). IEEE, 2011. http://dx.doi.org/10.1109/bmsb.2011.5954949.
Full textLi, Chenda, and Yanmin Qian. "Deep Audio-Visual Speech Separation with Attention Mechanism." In ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020. http://dx.doi.org/10.1109/icassp40776.2020.9054180.
Full textWu, Yu, Linchao Zhu, Yan Yan, and Yi Yang. "Dual Attention Matching for Audio-Visual Event Localization." In 2019 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE, 2019. http://dx.doi.org/10.1109/iccv.2019.00639.
Full textGuo, Ningning, Huaping Liu, and Linhua Jiang. "Attention-based Visual-Audio Fusion for Video Caption Generation." In 2019 IEEE 4th International Conference on Advanced Robotics and Mechatronics (ICARM). IEEE, 2019. http://dx.doi.org/10.1109/icarm.2019.8834066.
Full textReports on the topic "Audio-visual attention"
Tarasenko, Rostyslav O., Svitlana M. Amelina, Yuliya M. Kazhan, and Olga V. Bondarenko. The use of AR elements in the study of foreign languages at the university. CEUR Workshop Proceedings, November 2020. http://dx.doi.org/10.31812/123456789/4421.
Full textTarasenko, Rostyslav O., Svitlana M. Amelina, Yuliya M. Kazhan, and Olga V. Bondarenko. The use of AR elements in the study of foreign languages at the university. CEUR Workshop Proceedings, November 2020. http://dx.doi.org/10.31812/123456789/4421.
Full text