Gotowa bibliografia na temat „Auditory spatial perception”
Utwórz poprawne odniesienie w stylach APA, MLA, Chicago, Harvard i wielu innych
Spis treści
Zobacz listy aktualnych artykułów, książek, rozpraw, streszczeń i innych źródeł naukowych na temat „Auditory spatial perception”.
Przycisk „Dodaj do bibliografii” jest dostępny obok każdej pracy w bibliografii. Użyj go – a my automatycznie utworzymy odniesienie bibliograficzne do wybranej pracy w stylu cytowania, którego potrzebujesz: APA, MLA, Harvard, Chicago, Vancouver itp.
Możesz również pobrać pełny tekst publikacji naukowej w formacie „.pdf” i przeczytać adnotację do pracy online, jeśli odpowiednie parametry są dostępne w metadanych.
Artykuły w czasopismach na temat "Auditory spatial perception"
Recanzone, Gregg H. "Auditory Influences on Visual Temporal Rate Perception". Journal of Neurophysiology 89, nr 2 (1.02.2003): 1078–93. http://dx.doi.org/10.1152/jn.00706.2002.
Pełny tekst źródłaHaas, Ellen C. "Auditory Perception". Proceedings of the Human Factors Society Annual Meeting 36, nr 3 (październik 1992): 247. http://dx.doi.org/10.1518/107118192786751817.
Pełny tekst źródłaBest, Virginia, Jorg M. Buchholz i Tobias Weller. "Measuring auditory spatial perception in realistic environments". Journal of the Acoustical Society of America 141, nr 5 (maj 2017): 3692. http://dx.doi.org/10.1121/1.4988040.
Pełny tekst źródłaLau, Bonnie K., Tanya St. John, Annette Estes i Stephen Dager. "Auditory processing in neurodiverse children". Journal of the Acoustical Society of America 155, nr 3_Supplement (1.03.2024): A75. http://dx.doi.org/10.1121/10.0026855.
Pełny tekst źródłaPeng, Z. Ellen. "School-age children show poor use of spatial cues in reverberation for speech-in-speech perception". Journal of the Acoustical Society of America 151, nr 4 (kwiecień 2022): A169. http://dx.doi.org/10.1121/10.0011001.
Pełny tekst źródłaKoohi, Nehzat, Gilbert Thomas-Black, Paola Giunti i Doris-Eva Bamiou. "Auditory Phenotypic Variability in Friedreich’s Ataxia Patients". Cerebellum 20, nr 4 (18.02.2021): 497–508. http://dx.doi.org/10.1007/s12311-021-01236-9.
Pełny tekst źródłaCui, Qi N., Babak Razavi, William E. O'Neill i Gary D. Paige. "Perception of Auditory, Visual, and Egocentric Spatial Alignment Adapts Differently to Changes in Eye Position". Journal of Neurophysiology 103, nr 2 (luty 2010): 1020–35. http://dx.doi.org/10.1152/jn.00500.2009.
Pełny tekst źródłaStrybel, Thomas Z. "Auditory Spatial Information and Head-Coupled Display Systems". Proceedings of the Human Factors Society Annual Meeting 32, nr 2 (październik 1988): 75. http://dx.doi.org/10.1177/154193128803200215.
Pełny tekst źródłaUpadhya, Sushmitha, Rohit Bhattacharyya, Ritwik Jargar i K. Nisha Venkateswaran. "Closed-field Auditory Spatial Perception and Its Relationship to Musical Aptitude". Journal of Indian Speech Language & Hearing Association 37, nr 2 (2023): 61–65. http://dx.doi.org/10.4103/jisha.jisha_20_23.
Pełny tekst źródłaTerrence, Peter I., J. Christopher Brill i Richard D. Gilson. "Body Orientation and the Perception of Spatial Auditory and Tactile Cues". Proceedings of the Human Factors and Ergonomics Society Annual Meeting 49, nr 17 (wrzesień 2005): 1663–67. http://dx.doi.org/10.1177/154193120504901735.
Pełny tekst źródłaRozprawy doktorskie na temat "Auditory spatial perception"
Keating, Peter. "Plasticity and integration of auditory spatial cues". Thesis, University of Oxford, 2011. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.561113.
Pełny tekst źródłaGeeseman, Joseph W. "The influence of auditory cues on visual spatial perception". OpenSIUC, 2010. https://opensiuc.lib.siu.edu/theses/286.
Pełny tekst źródłaGriffiths, Shaaron S., i shaaron griffiths@deakin edu au. "Spatial and temporal disparaties in aurally aided visual search". Deakin University. School of Psychology, 2001. http://tux.lib.deakin.edu.au./adt-VDU/public/adt-VDU20061207.134032.
Pełny tekst źródłaElias, Bartholomew. "Cross-modal facilitation of spatial frequency discriminations through auditory frequency cue presentations". Thesis, Georgia Institute of Technology, 1991. http://hdl.handle.net/1853/28611.
Pełny tekst źródłaBest, Virginia Ann. "Spatial Hearing with Simultaneous Sound Sources: A Psychophysical Investigation". University of Sydney. Medicine, 2004. http://hdl.handle.net/2123/576.
Pełny tekst źródłaJin, Craig T. "Spectral analysis and resolving spatial ambiguities in human sound localization". Connect to full text, 2001. http://hdl.handle.net/2123/1342.
Pełny tekst źródłaTitle from title screen (viewed 13 January 2009). Submitted in fulfilment of the requirements for the degree of Doctor of Philosophy to the School of Electrical and Information Engineering, Faculty of Engineering. Includes bibliographical references. Also available in print form.
Nuckols, Richard. "Localization of Auditory Spatial Targets in Sighted and Blind Subjects". VCU Scholars Compass, 2013. http://scholarscompass.vcu.edu/etd/3286.
Pełny tekst źródłaEuston, David Raymond. "From spectrum to space the integration of frequency-specific intensity cues to produce auditory spatial receptive fields in the barn owl inferior colliculus /". [Eugene, Or. : University of Oregon Library System], 2000. http://libweb.uoregon.edu/UOTheses/2000/eustond00.pdf.
Pełny tekst źródłaEuston, David Raymond 1964. "From spectrum to space: the integration of frequency-specific intensity cues to produce auditory spatial receptive fields in the barn owl inferior colliculus". Thesis, University of Oregon, 2000. http://hdl.handle.net/1794/143.
Pełny tekst źródłaNeurons in the barn owl's inferior colliculus (IC) derive their spatial receptive fields (RF) from two auditory cues: interaural time difference (ITD) and interaural level difference (ILD). ITD serves to restrict a RF in azimuth but the precise role of ILD was, up to this point, unclear. Filtering by the ears and head insures that each spatial location is associated with a unique combination of frequency-specific ILD values (i.e., an ILD spectrum). We isolated the effect of ILD spectra using virtual sound sources in which ITD was held fixed for all spatial locations while ILD spectra were allowed to vary normally. A cell's response to these stimuli reflects the contribution of ILD to spatial tuning, referred to as an “ILD-alone RF”. In a sample of 34 cells, individual ILD-alone RFs were distributed and amorphous, but consistently showed that the ILD spectrum is facilatory at the cell's best location and inhibitory above and/or below. Prior results have suggested that an IC cell's spatial specificity is generated by summing inputs which are narrowly tuned to frequency and selective for both ILD and ITD. Based on this premise, we present a developmental model which, when trained solely on a cell's true spatial RF, reproduces both the cell's true RF and its ILD-alone RF. According to the model, the connectivity between a space-tuned IC cell and its frequency-specific inputs develops subject to two constraints: the cell must be excited by ILD spectra from the cell's best location and inhibited by spectra from locations above and below but along the vertical strip defined by the best ITD. To assess how frequency-specific inputs are integrated to form restricted spatial RFs, we measured the responses of 47 space-tuned IC cells to pure tones at varying ILDs and frequencies. ILD tuning varied with frequency. Further, pure-tone responses, summed according to the head-related filters, accounted for 56 percent of the variance in broadband ILD-alone RFs. Modelling suggests that, with broadband sounds, cells behave as though they are linearly summing their inputs, but when testing with pure tones, non-linearities arise. This dissertation includes unpublished co-authored materials.
Cogné, Mélanie. "Influence de modulations sensorielles sur la navigation et la mémoire spatiale en réalité virtuelle : Processus cognitifs impliqués". Thesis, Bordeaux, 2017. http://www.theses.fr/2017BORD0704.
Pełny tekst źródłaNavigating in a familiar or unfamiliar environment is a frequent challenge for human beings. Many patients with brain injury suffer from topographical difficulties, which influences their autonomy in daily life. Virtual Reality Tools enable the evaluation of largescale spatial navigation and spatial memory, resembling a real environment. Virtual reality also permits to add stimuli to the software. These stimuli can be contextual, that is to say linked to the task that participants have to accomplish in the Virtual Environment, or non-contextual, i.e. with no link with the require task. This thesis investigates whether visual or auditory stimuli influence spatial navigation and memory in Virtual Environments of patients with brain injury or with a neurodegenerative disease. The first part of the thesis showed contextual auditory stimuli type a sonar effect and the names of products of the shopping list improved spatial navigation of brain-injured patients during a shopping task in the virtual supermarket VAP-S. The second part of this thesis highlighted that non-contextual auditory stimuli with a high perceptual or cognitive salience decreased spatial navigation performance of brain-injured patients during a shopping task in the VAP-S. The third part of this thesis showed that visual cues like directional arrows and salient landmarks improved spatial navigation and some aspects of spatial memory of patients with Alzheimer’s disease or Mild Cognitive Impairments during a navigation task in a virtual district. The last part of this thesis demonstrated that auditory cues, i.e. beeping sounds indicating the directions, increased spatial navigation in a virtual district of patients who have had a stroke with contra-lesional visual and auditory neglect. These results suggest that some visual and auditory stimuli could be helpful for spatial navigation and memory tasks in patients with brain injury of neuro-degenerative disease. It further offers new research avenues for neuro-rehabilitation, such as the use of augmented reality in real-life settings to support the navigational capabilities of these patients
Książki na temat "Auditory spatial perception"
Blauert, Jens. Spatial hearing: The psychophysics of human sound localization. Cambridge, Mass: MIT Press, 1997.
Znajdź pełny tekst źródłaH, Gilkey Robert, i Anderson Timothy R, red. Binaural and spatial hearing in real and virtual environments. Mahwah, N.J: Lawrence Erlbaum Associates, 1997.
Znajdź pełny tekst źródłaPrinciples And Applications Of Spatial Hearing. World Scientific Publishing Company, 2011.
Znajdź pełny tekst źródłaSuzuki, Yoiti, Douglas Brungart i Kazuhiro Iida. Principles and Applications of Spatial Hearing. World Scientific Publishing Co Pte Ltd, 2011.
Znajdź pełny tekst źródłaAnderson, Timothy R., i Robert Gilkey. Binaural and Spatial Hearing in Real and Virtual Environments. Taylor & Francis Group, 2014.
Znajdź pełny tekst źródłaAnderson, Timothy R., i Robert Gilkey. Binaural and Spatial Hearing in Real and Virtual Environments. Taylor & Francis Group, 2014.
Znajdź pełny tekst źródłaAnderson, Timothy R., i Robert Gilkey. Binaural and Spatial Hearing in Real and Virtual Environments. Taylor & Francis Group, 2014.
Znajdź pełny tekst źródłaAnderson, Timothy R., i Robert Gilkey. Binaural and Spatial Hearing in Real and Virtual Environments. Taylor & Francis Group, 2014.
Znajdź pełny tekst źródłaAnderson, Timothy R., i Robert H. Gilkey. Binaural and Spatial Hearing in Real and Virtual Environments. Taylor & Francis Group, 2015.
Znajdź pełny tekst źródłaParnas, Josef, i Annick Urfer-Parnas. The ontology and epistemology of symptoms: The case of auditory verbal hallucinations in schizophrenia. Redaktorzy Kenneth S. Kendler i Josef Parnas. Oxford University Press, 2017. http://dx.doi.org/10.1093/med/9780198796022.003.0026.
Pełny tekst źródłaCzęści książek na temat "Auditory spatial perception"
Hehrmann, P., J. K. Maier, N. S. Harper, D. McAlpine i Maneesh Sahani. "Adaptive Coding for Auditory Spatial Cues". W The Neurophysiological Bases of Auditory Perception, 357–66. New York, NY: Springer New York, 2010. http://dx.doi.org/10.1007/978-1-4419-5686-6_34.
Pełny tekst źródłaElhilali, Mounya, Ling Ma, Christophe Micheyl, Andrew Oxenham i Shihab Shamma. "Rate Versus Temporal Code? A Spatio-Temporal Coherence Model of the Cortical Basis of Streaming". W The Neurophysiological Bases of Auditory Perception, 497–506. New York, NY: Springer New York, 2010. http://dx.doi.org/10.1007/978-1-4419-5686-6_46.
Pełny tekst źródłaCedolin, Leonardo, i Bertrand Delgutte. "Spatio-Temporal Representation of the Pitch of Complex Tones in the Auditory Nerve". W Hearing – From Sensory Processing to Perception, 61–70. Berlin, Heidelberg: Springer Berlin Heidelberg, 2007. http://dx.doi.org/10.1007/978-3-540-73009-5_8.
Pełny tekst źródłaBarrie, John M., i Walter J. Freeman. "Perceptual Topography: Spatio-Temporal Analysis of Prepyriform, Visual, Auditory, and Somesthetic EEGs in Perception by Trained Rabbits". W The Neurobiology of Computation, 173–78. Boston, MA: Springer US, 1995. http://dx.doi.org/10.1007/978-1-4615-2235-5_28.
Pełny tekst źródłaWightman, Frederic L., i Rick Jenison. "Auditory Spatial Layout". W Perception of Space and Motion, 365–400. Elsevier, 1995. http://dx.doi.org/10.1016/b978-012240530-3/50012-2.
Pełny tekst źródłaZAHORIK, P., E. BRANDEWIE i V. P. SIVONEN. "AUDITORY PERCEPTION IN REVERBERANT SOUND FIELDS AND EFFECTS OF PRIOR LISTENING EXPOSURE". W Principles and Applications of Spatial Hearing, 24–34. WORLD SCIENTIFIC, 2011. http://dx.doi.org/10.1142/9789814299312_0003.
Pełny tekst źródłaMilner, A. David, H. Chris Dijkerman i David P. Carey. "Visuospatial processing in a pure case of visual-form agnosia". W The Hippocampal and Parietal Foundations of Spatial Cognition, 443–66. Oxford University PressOxford, 1998. http://dx.doi.org/10.1093/oso/9780198524533.003.0023.
Pełny tekst źródłavon Kriegstein, Katharina, i Christa Müller-Axt. "The Role of the Thalamus for Human Auditory and Visual Speech Perception". W The Cerebral Cortex and Thalamus, redaktorzy Andrew J. King i Judith A. Hirsch, 239–47. Oxford University PressNew York, 2023. http://dx.doi.org/10.1093/med/9780197676158.003.0023.
Pełny tekst źródłaLadavas, Elisabetta, i Alessandro Farne. "Multisensory Representation of Peripe onal Space". W Human Body Perception From The Inside Out, 89–104. Oxford University PressNew York, NY, 2005. http://dx.doi.org/10.1093/oso/9780195178371.003.0005.
Pełny tekst źródłaPesic, Peter. "Helmholtz and the Sirens". W Music and the Making of Modern Science. The MIT Press, 2014. http://dx.doi.org/10.7551/mitpress/9780262027274.003.0015.
Pełny tekst źródłaStreszczenia konferencji na temat "Auditory spatial perception"
Chao, Yujing, Zhijun Zhao, Chang Liu i Lingyun Xie. "Auditory Space Perception Influenced by Visual Spatial Information". W 2018 IEEE 3rd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC). IEEE, 2018. http://dx.doi.org/10.1109/iaeac.2018.8577596.
Pełny tekst źródłaBukvic, Ivica Ico, Gregory Earle, Disha Sardana i Woohun Joo. "Studies in Spatial Aural Perception: Establishing Foundations for Immersive Sonification". W ICAD 2019: The 25th International Conference on Auditory Display. Newcastle upon Tyne, United Kingdom: Department of Computer and Information Sciences, Northumbria University, 2019. http://dx.doi.org/10.21785/icad2019.017.
Pełny tekst źródłaBăcilă, Bogdan Ioan, i Hyunkook Lee. "Subjective Elicitation Of Listener-Perspective-Dependent Spatial Attributes in a Rerverberant Room, using the Repertory Grid Technique". W ICAD 2019: The 25th International Conference on Auditory Display. Newcastle upon Tyne, United Kingdom: Department of Computer and Information Sciences, Northumbria University, 2019. http://dx.doi.org/10.21785/icad2019.073.
Pełny tekst źródłaTomoriova, Beata, i Norbert Kopco. "Auditory Spatial Cuing for Speech Perception in a Dynamic Multi-talker Environment". W 2008 6th International Symposium on Applied Machine Intelligence and Informatics (SAMI '08). IEEE, 2008. http://dx.doi.org/10.1109/sami.2008.4469177.
Pełny tekst źródłaSardana, Disha, Woohun Joo, Ivica Ico Bukvic i Gregory Earle. "Perception of spatial data properties in an immersive multi-layered auditory environment". W AM'20: Audio Mostly 2020. New York, NY, USA: ACM, 2020. http://dx.doi.org/10.1145/3411109.3411134.
Pełny tekst źródłaBeadling, Andrew, i Paul Vickers. "Listener Perception of Spatialised Audio for Embodied Interaction in Sonification". W ICAD 2023: The 28th International Conference on Auditory Display. icad.org: International Community for Auditory Display, 2023. http://dx.doi.org/10.21785/icad2023.3814.
Pełny tekst źródłaVoong, Tray Minh, i Michael Oehler. "Auditory Spatial Perception Using Bone Conduction Headphones along with Fitted Head Related Transfer Functions". W 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR). IEEE, 2019. http://dx.doi.org/10.1109/vr.2019.8798218.
Pełny tekst źródłaHirahara, Tatsuya, Daisuke Yoshisaki i Daisuke Morikawa. "Impact of dynamic binaural signal associated with listener's voluntary movement in auditory spatial perception". W ICA 2013 Montreal. ASA, 2013. http://dx.doi.org/10.1121/1.4799864.
Pełny tekst źródłaLarsson, Pontus, Johanna Bergfelt Ramos de Souza i Joel Begnert. "An Auditory Display for Remote Road Vehicle Operation That Increases Awareness and Presence". W ICAD 2023: The 28th International Conference on Auditory Display. icad.org: International Community for Auditory Display, 2023. http://dx.doi.org/10.21785/icad2023.8296.
Pełny tekst źródłaGodfroy-Cooper, Martine, Elizabeth Wenzel, Joel Miller i Edward Bachelder. "Cheeseman Award Paper: Isomorphic Spatial Visual Auditory Displays for Operations in DVE for Obstacle Avoidance". W Vertical Flight Society 75th Annual Forum & Technology Display. The Vertical Flight Society, 2019. http://dx.doi.org/10.4050/f-0075-2019-14563.
Pełny tekst źródłaRaporty organizacyjne na temat "Auditory spatial perception"
Letowski, Tomasz R., i Szymon T. Letowski. Auditory Spatial Perception: Auditory Localization. Fort Belvoir, VA: Defense Technical Information Center, maj 2012. http://dx.doi.org/10.21236/ada562292.
Pełny tekst źródłaMahat, Marian, Vivienne Awad, Christopher Bradbeer, Chengxin Guo, Wesley Imms i Julia Morris. Furniture for Engagement. University of Melbourne, luty 2023. http://dx.doi.org/10.46580/124374.
Pełny tekst źródła