Dissertations / Theses on the topic 'Visual grounding of text'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Visual grounding of text.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Engilberge, Martin. "Deep Inside Visual-Semantic Embeddings." Electronic Thesis or Diss., Sorbonne université, 2020. http://www.theses.fr/2020SORUS150.
Full textNowadays Artificial Intelligence (AI) is omnipresent in our society. The recentdevelopment of learning methods based on deep neural networks alsocalled "Deep Learning" has led to a significant improvement in visual representation models.and textual.In this thesis, we aim to further advance image representation and understanding.Revolving around Visual Semantic Embedding (VSE) approaches, we explore different directions: We present relevant background covering images and textual representation and existing multimodal approaches. We propose novel architectures further improving retrieval capability of VSE and we extend VSE models to novel applications and leverage embedding models to visually ground semantic concept. Finally, we delve into the learning process andin particular the loss function by learning differentiable approximation of ranking based metric
Emmott, Stephen J. "The visual processing of text." Thesis, University of Stirling, 1993. http://hdl.handle.net/1893/1837.
Full textMi, Jinpeng Verfasser], and Jianwei [Akademischer Betreuer] [Zhang. "Natural Language Visual Grounding via Multimodal Learning / Jinpeng Mi ; Betreuer: Jianwei Zhang." Hamburg : Staats- und Universitätsbibliothek Hamburg, 2020. http://d-nb.info/1205070885/34.
Full textMi, Jinpeng [Verfasser], and Jianwei [Akademischer Betreuer] Zhang. "Natural Language Visual Grounding via Multimodal Learning / Jinpeng Mi ; Betreuer: Jianwei Zhang." Hamburg : Staats- und Universitätsbibliothek Hamburg, 2020. http://d-nb.info/1205070885/34.
Full textPrince, Md Enamul Hoque. "Visual text analytics for online conversations." Thesis, University of British Columbia, 2017. http://hdl.handle.net/2429/61772.
Full textScience, Faculty of
Computer Science, Department of
Graduate
Chauhan, Aneesh. "Grounding human vocabulary in robot perception through interaction." Doctoral thesis, Universidade de Aveiro, 2014. http://hdl.handle.net/10773/12841.
Full textThis thesis addresses the problem of word learning in computational agents. The motivation behind this work lies in the need to support language-based communication between service robots and their human users, as well as grounded reasoning using symbols relevant for the assigned tasks. The research focuses on the problem of grounding human vocabulary in robotic agent’s sensori-motor perception. Words have to be grounded in bodily experiences, which emphasizes the role of appropriate embodiments. On the other hand, language is a cultural product created and acquired through social interactions. This emphasizes the role of society as a source of linguistic input. Taking these aspects into account, an experimental scenario is set up where a human instructor teaches a robotic agent the names of the objects present in a visually shared environment. The agent grounds the names of these objects in visual perception. Word learning is an open-ended problem. Therefore, the learning architecture of the agent will have to be able to acquire words and categories in an openended manner. In this work, four learning architectures were designed that can be used by robotic agents for long-term and open-ended word and category acquisition. The learning methods used in these architectures are designed for incrementally scaling-up to larger sets of words and categories. A novel experimental evaluation methodology, that takes into account the openended nature of word learning, is proposed and applied. This methodology is based on the realization that a robot’s vocabulary will be limited by its discriminatory capacity which, in turn, depends on its sensors and perceptual capabilities. An extensive set of systematic experiments, in multiple experimental settings, was carried out to thoroughly evaluate the described learning approaches. The results indicate that all approaches were able to incrementally acquire new words and categories. Although some of the approaches could not scale-up to larger vocabularies, one approach was shown to learn up to 293 categories, with potential for learning many more.
Esta tese aborda o problema da aprendizagem de palavras em agentes computacionais. A motivação por trás deste trabalho reside na necessidade de suportar a comunicação baseada em linguagem entre os robôs de serviço e os seus utilizadores humanos, bem como suportar o raciocínio baseado em símbolos que sejam relevantes no contexto das tarefas atribuídas e cujo significado seja definido com base na experiência perceptiva. Mais especificamente, o foco da investigação é o problema de estabelecer o significado das palavras na percepção do robô através da interacção homemrobô. A definição do significado das palavras com base em experiências perceptuais e perceptuo-motoras enfatiza o papel da configuração física e perceptuomotora do robô. Entretanto, a língua é um produto cultural criado e adquirido através de interacções sociais. Isso destaca o papel da sociedade como fonte linguística. Tendo em conta estes aspectos, um cenário experimental foi definido no qual um instrutor humano ensina a um agente robótico os nomes dos objectos presentes num ambiente visualmente partilhado. O agente associa os nomes desses objectos à sua percepção visual desses objectos. A aprendizagem de palavras é um problema sem objectivo pré-estabelecido. Nós adquirimos novas palavras ao longo das nossas vidas. Assim, a arquitectura de aprendizagem do agente deve poder adquirir palavras e categorias de uma forma semelhante. Neste trabalho foram concebidas quatro arquitecturas de aprendizagem que podem ser usadas por agentes robóticos para aprendizagem e aquisição de novas palavras e categorias, incrementalmente. Os métodos de aprendizagem utilizados nestas arquitecturas foram projectados para funcionar de forma incremental, acumulando um conjunto cada vez maior de palavras e categorias. É proposta e aplicada uma nova metodologia da avaliação experimental que leva em conta a natureza aberta e incremental da aprendizagem de palavras. Esta metodologia leva em consideração a constatação de que o vocabulário de um robô será limitado pela sua capacidade de discriminação, a qual, por sua vez, depende dos seus sensores e capacidades perceptuais. Foi realizado um extenso conjunto de experiências sistemáticas em múltiplas situações experimentais, para avaliar cuidadosamente estas abordagens de aprendizagem. Os resultados indicam que todas as abordagens foram capazes de adquirir novas palavras e categorias incrementalmente. Embora em algumas das abordagens não tenha sido possível atingir vocabulários maiores, verificou-se que uma das abordagens conseguiu aprender até 293 categorias, com potencial para aprender muitas mais.
Sabir, Ahmed. "Enhancing scene text recognition with visual context information." Doctoral thesis, Universitat Politècnica de Catalunya, 2020. http://hdl.handle.net/10803/670286.
Full textAquesta tesi aborda el problema de millorar els sistemes de reconeixement de text, que permeten detectar i reconèixer text en imatges no restringides (per exemple, un cartell al carrer, un anunci, una destinació d’autobús, etc.). L’objectiu és millorar el rendiment dels sistemes de visió existents explotant la informació semàntica derivada de la pròpia imatge. La idea principal és que conèixer el contingut de la imatge o el context visual en el que un text apareix, pot ajudar a decidir quines són les paraules correctes. Per exemple, el fet que una imatge mostri una cafeteria fa que sigui més probable que una paraula en un rètol es llegeixi com a Dunkin que no pas com unkind. Abordem aquest problema recorrent a avenços en el processament del llenguatge natural i l’aprenentatge automàtic, en particular, aprenent re-rankers i xarxes neuronals, per presentar solucions de postprocés que milloren els sistemes de l’estat de l’art de reconeixement de text, sense necessitat de costosos procediments de reentrenament o afinació que requereixin grans quantitats de dades. Descobrir el grau de relació semàntica entre les paraules candidates i el seu context d’imatge és una tasca relacionada amb l’avaluació de la semblança semàntica entre paraules o fragments de text. Tanmateix, determinar l’existència d’una relació semàntica és una tasca més general que avaluar la semblança (per exemple, cotxe, carretera i semàfor estan relacionats però no són similars) i per tant els mètodes existents requereixen certes adaptacions. Per satisfer els requisits d’aquestes perspectives més àmplies de relació semàntica, desenvolupem dos enfocaments per aprendre la relació semàntica de la paraula reconeguda i el seu context: paraula-a-paraula (amb els objectes a la imatge) o paraula-a-frase (subtítol de la imatge). En l’enfocament de paraula-a-paraula s’usen re-rankers basats en word-embeddings. El re-ranker pren les paraules proposades pel sistema base i les torna a reordenar en funció del context visual proporcionat pel classificador d’objectes. Per al segon cas, s’ha dissenyat un enfocament neuronal d’extrem a extrem per explotar la descripció de la imatge (subtítol) tant a nivell de frase com a nivell de paraula i re-ordenar les paraules candidates basant-se tant en el context visual com en les co-ocurrències amb el subtítol. Com a contribució addicional, per satisfer els requisits dels enfocs basats en dades com ara les xarxes neuronals, presentem un conjunt de dades de contextos visuals per a aquesta tasca, en el què el conjunt de dades COCO-text disponible públicament [Veit et al. 2016] s’ha ampliat amb informació sobre l’escena (inclosos els objectes i els llocs que apareixen a la imatge) per permetre als investigadors incloure les relacions semàntiques entre textos i escena als seus sistemes de reconeixement de text, i oferir una base d’avaluació comuna per a aquests enfocaments.
Willems, Heather Marie. "Writing the written: text as a visual image." The Ohio State University, 2005. http://rave.ohiolink.edu/etdc/view?acc_num=osu1382952227.
Full textKan, Jichao. "Visual-Text Translation with Deep Graph Neural Networks." Thesis, University of Sydney, 2020. https://hdl.handle.net/2123/23759.
Full textShmueli, Yael. "Integrating speech and visual text in multimodal interfaces." Thesis, University College London (University of London), 2005. http://discovery.ucl.ac.uk/1446688/.
Full textRohrbach, Anna [Verfasser], and Bernt [Akademischer Betreuer] Schiele. "Generation and grounding of natural language descriptions for visual data / Anna Rohrbach ; Betreuer: Bernt Schiele." Saarbrücken : Saarländische Universitäts- und Landesbibliothek, 2017. http://d-nb.info/113349207X/34.
Full textZablocki, Éloi. "Multimodal machine learning : complementarity of textual and visual contexts." Electronic Thesis or Diss., Sorbonne université, 2019. http://www.theses.fr/2019SORUS409.
Full textResearch looking at the interaction between language and vision, despite a growing interest, is relatively underexplored. Beyond trivial differences between texts and images, these two modalities have non overlapping semantics. On the one hand, language can express high-level semantics about the world, but it is biased in the sense that a large portion of its content is implicit (common-sense or implicit knowledge). On the other hand, images are aggregates of lower-level information, but they can depict a more direct view of real-world statistics and can be used to ground the meaning of objects. In this thesis, we exploit connections and leverage complementarity between language and vision. First, natural language understanding capacities can be augmented with the help of the visual modality, as language is known to be grounded in the visual world. In particular, representing language semantics is a long-standing problem for the natural language processing community, and to further improve traditional approaches towards that goal, leveraging visual information is crucial. We show that semantic linguistic representations can be enriched by visual information, and we especially focus on visual contexts and spatial organization of scenes. We present two models to learn grounded word or sentence semantic representations respectively, with the help of images. Conversely, integrating language with vision brings the possibility of expanding the horizons and tasks of the vision community. Assuming that language contains visual information about objects, and that this can be captured within linguistic semantic representation, we focus on the zero-shot object recognition task, which consists in recognizing objects that have never been seen thanks to linguistic knowledge acquired about the objects beforehand. In particular, we argue that linguistic representations not only contain visual information about the visual appearance of objects but also about their typical visual surroundings and visual occurrence frequencies. We thus present a model for zero-shot recognition that leverages the visual context of an object, and its visual occurrence likelihood, in addition to the region of interest as done in traditional approaches. Finally, we present prospective research directions to further exploit connections between language and images and to better understand the semantic gap between the two modalities
Larsmo, Sebastian. "Mellan text och bild." Thesis, Konstfack, Grafisk design & illustration, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:konstfack:diva-5914.
Full textFiorini, Sandro Rama. "S-Chart : um arcabouço para interpretação visual de gráficos." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2009. http://hdl.handle.net/10183/17039.
Full textSemantic image interpretation is one of the most promising frontiers in the Computer Vision area, specifically when applied to Image Interpretation. To reach semantic interpretation, visual knowledge explicitly represented is applied by symbolic reasoning algorithms combined with image processing algorithms in order to extract the content of the images and associate it with semantically rich models. This work describes the S-Chart approach, a semantic image interpretation approach designed for interpretation of line charts. It is structured as a set of knowledge models and algorithms that can be instantiated to accomplish chart interpretation in other domains. The models are represented in three semantic levels and apply the concept of symbol grounding in order to map the primitives between the levels. The interpretation algorithms carry out the interaction between the symbolic reasoning in the high level, and the signal processing algorithms in the low level data. In order to demonstrate the applicability of the S-Chart framework, we developed the InteliStrata system, an application in Geology for the semantic interpretation of well log profiles. Using the application, we have interpreted the graphs of two gamma-ray profiles captured in exploration wells, to indicate the position of Stratigraphic Sequences and the maximum flooding surfaces. The results were compared with the interpretation of an experienced geologist using the same data input. The system was able to point the same identified sequences and offered alternative interpretation that were compatible with the geologist interpretation over the data. The S-Chart framework demonstrates its effectiveness on interpretation of pictorial information in knowledge intensive domains. The stronger points of the approach are its domain independent models for visual knowledge representation and, specially, the application of a symbol grounding model to provide a correlation between representation primitives.
Pilz, Anja [Verfasser]. "Entity Linking to Wikipedia : Grounding entity mentions in natural language text using thematic context distance and collective search / Anja Pilz." Bonn : Universitäts- und Landesbibliothek Bonn, 2016. http://d-nb.info/1096330032/34.
Full textBradel, Lauren C. "Multi-Model Semantic Interaction for Scalable Text Analytics." Diss., Virginia Tech, 2015. http://hdl.handle.net/10919/52785.
Full textPh. D.
Heibi, Ivan. "A Visual Framework for Graph and Text Analytics in Email Investigation." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/13818/.
Full textKozlovski, Nikolai. "TEXT-IMAGE RESTORATION AND TEXT ALIGNMENT FOR MULTI-ENGINE OPTICAL CHARACTER RECOGNITION SYSTEMS." Master's thesis, University of Central Florida, 2006. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/3607.
Full textM.S.E.E.
Department of Electrical and Computer Engineering
Engineering and Computer Science
Electrical Engineering
Clavelli, Antonio. "A computational model of eye guidance, searching for text in real scene images." Doctoral thesis, Universitat Autònoma de Barcelona, 2014. http://hdl.handle.net/10803/283428.
Full textLa búsqueda de objetos de texto en imágenes de escena reales es un problema abierto y un área de investigación muy activa la visión por computador. Se han propuesto un gran número de métodos basados en la extensión de los métodos desde el análisis de documentos o inspirado en métodos de detección de objetos. Sin embargo, el problema de la búsqueda de objetos en imágenes de escena reales sigue siendo un problema extremadamente difícil debido a la gran variabilidad en la apariencia de los objetos. Esta tesis se basa en los más recientes hallazgos en la literatura de la atención visual, introduciendo un nuevo modelo computacional de visión guiada que apunta a describir la búsqueda de texto en imágenes de escenas reales. En primer lugar se presentan los resultados mas pertinentes de la literatura científica en relación con la atención visual, los movimientos oculares y la búsqueda visual. Los mas relevantes modelos de atención son discutidos e integrados con recientes obser- vaciones sobre la función de los denominados ’top-down constraints’ y la emergente necesidad de un modelo estratificado de atención en el que la saliencia no es el único factor guía de atención. La atención visual se explica por la interacción de varios factores moduladores, tales como objetos, valor, planes y saliencia. Se introduce nuestra formulación probabilística de los mecanismos de atención en escenas reales para la tarea de búsqueda de objetos. El modelo se basa en el argumento de que el despliegue de atención depende de dos procesos distintos pero interactuantes: un proceso de atención que asigna valor a las fuentes de información y un proceso motor que une flexiblemente información con la acción. En ese marco, la elección de dónde buscar la próxima tarea es dependiente y orientada a las clases de objetos incrustados en imágenes de escenas reales. La dependencia de la tarea se tiene en cuenta al explotar el valor y la recompensa de contemplar ciertas partes o proto-objetos de la imagen que proporcionan una rala representación de los objetos en la escena. En la sección experimental se prueba el modelo en condiciones de laboratorio, comparando las simulaciones del modelo con datos de experimentos de eye tracking. La comparación es cualitativa en términos de trayectorias de exploración y cuanti- tativa, en términos de similitud estadística de la amplitud de movimientos oculares. Los experimentos se han realizado con datos de eye tracking tanto de un conjunto de datos públicos de rostros humanos y texto, tanto con un nuevo conjunto de datos de eye tracking y de imágenes urbanas con texto. La última parte de esta tesis se dedica a estudiar en qué medida el modelo prop- uesto puede responder del despliegue de atención en un entorno complejo. Se ha utilizado un dispositivo móvil de eye tracking y una metodología desarrollada es- pecíficamente para comparar los datos simulados con los datos grabados de eye track- ing. Tal configuración permite poner a prueba el modelo en la tarea de búsqueda de texto muy parecida a una búsqueda real, en la condición de información visual incompleta.
Searching for text objects in real scene images is an open problem and a very ac- tive computer vision research area. A large number of methods have been proposed tackling the text search as extension of the ones from the document analysis field or inspired by general purpose object detection methods. However the general problem of object search in real scene images remains an extremely challenging problem due to the huge variability in object appearance. This thesis builds on top of the most recent findings in the visual attention literature presenting a novel computational model of eye guidance aiming to better describe text object search in real scene images. First are presented the relevant state-of-the-art results from the visual attention literature regarding eye movements and visual search. Relevant models of attention are discussed and integrated with recent observations on the role of top-down con- straints and the emerging need for a layered model of attention in which saliency is not the only factor guiding attention. Visual attention is then explained by the interaction of several modulating factors, such as objects, value, plans and saliency. Then we introduce our probabilistic formulation of attention deployment in real scene. The model is based on the rationale that oculomotor control depends on two interacting but distinct processes: an attentional process that assigns value to the sources of information and motor process that flexibly links information with action. In such framework, the choice of where to look next is task-dependent and oriented to classes of objects embedded within pictures of complex scenes. The dependence on task is taken into account by exploiting the value and the reward of gazing at certain image patches or proto-objects that provide a sparse representation of the scene objects. In the experimental section the model is tested in laboratory condition, compar- ing model simulations with data from eye tracking experiments. The comparison is qualitative in terms of observable scan paths and quantitative in terms of statisti- cal similarity of gaze shift amplitude. Experiments are performed using eye tracking data from both a publicly available dataset of face and text and from newly performed eye-tracking experiments on a dataset of street view pictures containing text. The last part of this thesis is dedicated to study the extent to which the proposed model can account for human eye movements in a low constrained setting. We used a mobile eye tracking device and an ad-hoc developed methodology to compare model simulated eye data with the human eye data from mobile eye tracking recordings. Such setting allow to test the model in an incomplete visual information condition, reproducing a close to real-life search task.
Shevlin, Joseph P. "Text orienting cues: Enhancing dynamic visual acuity for targets of unknown origin." Diss., Virginia Tech, 1999. http://hdl.handle.net/10919/37664.
Full textPh. D.
Brown, Ron. "Visual Cueing: Investigating the Effects of Text Annotation on Student Retention Rates." Thesis, University of North Texas, 2003. https://digital.library.unt.edu/ark:/67531/metadc4232/.
Full textBell, S. "The promise of the short text : writing risk into visual arts practice." Thesis, Coventry University, 2013. http://curve.coventry.ac.uk/open/items/06d7767a-9941-4ab8-892e-e73767b48aa7/1.
Full textAjazi, Sara <1995>. "Experience and engagement: the case of Venice. Combining text and visual research." Master's Degree Thesis, Università Ca' Foscari Venezia, 2021. http://hdl.handle.net/10579/18739.
Full textGlover, Arren John. "Developing grounded representations for robots through the principles of sensorimotor coordination." Thesis, Queensland University of Technology, 2014. https://eprints.qut.edu.au/71763/1/Arren_Glover_Thesis.pdf.
Full textLaurence, Sean Xavier. "Analyzing non-collocated synchronous shared visual workspace-mediated interaction and effects on conversational grounding : a study on collaborative intelligence analysis." Thesis, Middlesex University, 2016. http://eprints.mdx.ac.uk/21257/.
Full textScannavino, Katia Romero Felizardo. "Evidence-based software engineering: systematic literature review process based on visual text mining." Universidade de São Paulo, 2012. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-18072012-102032/.
Full textContexto: Revisão Sistemática (RS) é uma metodologia utilizada para reunir evidências sobre uma quest~ao de pesquisa específica. Uma das atividades associadas à RS é a seleção de estudos primários. Quando o pesquisador se depara com grandes volumes de estudos, torna-se difícil selecionar artigos relevantes para uma análise mais aprofundada. Outra atividade associada à RS é a apresentação dos resultados dos estudos primários que atendem aos propósitos da RS. Os resultados são geralmente resumidos em tabelas e uma alternativa para reduzir o tempo consumido para entender os dados é o uso de representações gráficas. Mapeamento sistemático (MS) é uma forma mais aberta de RS, usado para construir um esquema de classificação e categorização sobre uma área de interesse. As atividades de categorização e classificação no MS não são tarefas triviais, pois exigem um esforço manual e conhecimento do domínio por parte dos revisores para a geração de resultados adequados. Embora relevantes, ambos os processos de RS e MS são demorados e muita das atividades são realizadas manualmente. Objetivo: O objetivo desta pesquisa é a utilização de Mineração Visual de Texto (VTM) para apoiar as diferentes atividades dos processos de RS e MS como, por exemplo, suporte à seleção de estudos primários, apresentação de resultados de RSs e a categorização e classificação em MSs. Métodos: Foram propostas extensões para os processos de RS e MS com base em VTM. Uma série de estudos de caso foram realizados para demonstrar a utilidade de técnicas VTM no contexto de seleção, revisão, apresentação de resultados e categorização. Resultados: Os resultados mostraram que a aplicação de VTM é promissora em termos de apoio positivo para a atividade de seleção de estudos primários e que o uso de representações visuais para apresentar resultados de RSs leva a uma redução do tempo necessário para sua análise, sem perda de compreensão de dados. A aplicação da VTM é relevante também no contexto da MS. Conclus~oes: Técnicas VTM podem ser empregadas com sucesso para ajudar nos processos de RS e MS
Foley, Marius. "High fidelity image tracing the emergence of a new constructed image /." Access electronically, 2004. http://ro.uow.edu.au/theses/285.
Full textPalmer, Kelly. "Let's see the facts: The effects of text mapping on comprehension and visual representation of expository text with fourth-grade students." Thesis, Wichita State University, 2013. http://hdl.handle.net/10057/6832.
Full textThesis (M.Ed.)--Wichita State University, College of Education, Dept. of Curriculum and Instruction
Halliwell, K. "Photography and narrative : An investigation of serial imagery." Thesis, University of Kent, 1986. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.373238.
Full textSpeight, Amanda Gaye. "Op writing : text ornamenting vision." Queensland University of Technology, 2008. http://eprints.qut.edu.au/16611/.
Full textRoy, Marguerite Claire. "Comprehension and learning through multimedia : integrative processing of text and illustrations." Thesis, McGill University, 2000. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=37826.
Full textParticipants were twenty-four undergraduate engineering students who had little prior knowledge of the target domain (the human visual system), but were experienced in learning about functional systems using texts and illustrations. Six students were randomly assigned to each of four presentation conditions: (a) text only, (b) illustrations only, (c) text with controlled access to illustrations, or (d) text with free access to illustrations.
The materials were presented individually in a computer environment which recorded and timed all information accessed. Participants provided on-line interpretations as they read, post-input verbal and visual free recalls of the materials, and responses to integrative comprehension questions. Planned comparisons were used to contrast: (a) the two text with illustrations groups, (b) the combined text with illustrations groups to the text only group, (c) the text with illustrations groups to the illustrations only group, and (d) the processing of information which was privileged (structure) to other information.
The results indicated that the text and illustrations each provided mutually constraining information that functioned together to support comprehension. Illustrations aided the construction and elaboration of mental models by providing an external context that supported more active conceptual processing and integration of information. Text aided both literal and high level comprehension by communicating the meaning of illustrations and signaling what information was important. These results support perspectives on situated learning which emphasize the role that discourse plays in comprehending knowledge in environments involving more than one external information source. Future research is recommended to extend such findings to other populations of learners and materials.
Speight, Amanda Gaye. "Op writing : text ornamenting vision." Thesis, Queensland University of Technology, 2008. https://eprints.qut.edu.au/16611/1/Amanda_Speight_Thesis.pdf.
Full textHeimerl, Florian [Verfasser], and Thomas [Akademischer Betreuer] Ertl. "Exploratory visual text analytics in the scientific literature domain / Florian Heimerl ; Betreuer: Thomas Ertl." Stuttgart : Universitätsbibliothek der Universität Stuttgart, 2017. http://d-nb.info/1139709712/34.
Full textOlson, Christina Louise. "Message from the grave a text-in-context case study of Bikur Cholim Sephardic Cemetery /." Lynchburg, Va. : Liberty University, 2008. http://digitalcommons.liberty.edu.
Full textJacobs, Ilene. "Performing the self : autobiography, narrative, image and text in self-representations." Thesis, Stellenbosch : University of Stellenbosch, 2007. http://hdl.handle.net/10019.1/1552.
Full textThesis received without illustrations at the time of submission to this repository.
This research follows the assumption that the notion of performativity can be applied to the visual construction of identity within art-making discourse in order to explore the contingent and mutable nature of identity in representation. My interest in performativity, defined as the active, repetitive and ritualistic processes responsible for the construction of subjectivities, lies within the process of production. I indicate how this notion, within the context of self-representation, can provide the possibility for performing identity as a process. I investigate the extent to which gender, the gaze, memory and narrative contribute to the performative construction of self-representations and reveal, through the exploration of my practical research, that these concepts are themselves performative. Although agency to construct the self can be regarded as problematic, considering the role of language and discourse in determining subjectivities, this research suggests that it is possible to perform interventions from within language. I suggest that the notion of inscription provides a means through which identity constructions can be performed differently; and that my art-making process of repetitive inscription, erasure and re-inscription of image and text and the layering of paint not only reflect the notion of performativity, but also enable me to expose the multiple and fragmented nature of identities.
Weaver, Beth Nixon. "Interactive text-image conceptual models for literary interpretation and composition in the digital age." Doctoral diss., University of Central Florida, 2010. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/4602.
Full textID: 028916553; System requirements: World Wide Web browser and PDF reader.; Mode of access: World Wide Web.; Thesis (Ph.D.)--University of Central Florida, 2010.; Includes bibliographical references (p. 359-370).
Ph.D.
Doctorate
Department of English
Arts and Humanities
Brown, K. E. "'Text and image in early twentieth-century Irish literary and visual culture : an Irish synaesthetic'." Thesis, Queen's University Belfast, 2006. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.438110.
Full textHuang, Wang. "Visual Sensation and Performative Cultural Politics: Chinese Literary Text Messages and the Colors of Texts." The Ohio State University, 2010. http://rave.ohiolink.edu/etdc/view?acc_num=osu1275499580.
Full textJi, Xiaonan. "An Integrated Framework of Text and Visual Analytics to Facilitate Information Retrieval towards Biomedical Literature." The Ohio State University, 2018. http://rave.ohiolink.edu/etdc/view?acc_num=osu1524199589980214.
Full textGerrard, Emily Elizabeth. "Picturebooks as visual literacy the influence of illustrations on second-graders' comprehension of narrative text /." College Park, Md.: University of Maryland, 2008. http://hdl.handle.net/1903/8999.
Full textThesis research directed by: Dept. of Curriculum and Instruction. Title from t.p. of PDF. Includes bibliographical references. Published by UMI Dissertation Services, Ann Arbor, Mich. Also available in paper.
Garrido, Gabarrón Enric. "Comunicació Institucional i Producció de Sentit. La significació del text en l'estructura gràfica dels cartells de l'Ajuntament de Barcelona." Doctoral thesis, Universitat Autònoma de Barcelona, 2021. http://hdl.handle.net/10803/673313.
Full textLas calles de las ciudades son entornos visualmente fértiles para que las instituciones se relacionen con sus ciudadanos a través de múltiples medios de comunicación. La irrupción de las Tecnologías de la Información y la Comunicación han derivado esta comunicación hacia plataformas digitales y redes sociales, pero el entorno visual urbano es todavía un espacio de interacción comunicativo utilizado por empresas e instituciones públicas para comunicarse con sus usuarios. La tipología de mensajes que encontramos en el espacio público es múltiple y diverso. Las ciudades se han convertido en un entorno habitado por miles de personas que, día a día, transitan por sus calles y plazas, quedando así expuestos a todo tipo de impactos comunicativos. Las instituciones públicas utilizan esta exposición como una oportunidad para enviar mensajes a la ciudadanía, a través de campañas institucionales, para generar mensajes de temáticas muy diversas relacionadas con políticas de convivencia, derechos sociales, rehabilitación de viviendas, discriminación, cultura o festividades, entre otros. De todos los formatos en que se integran en estas campañas, el cartel es posiblemente el elemento urbano que más ha perdurado en la convivencia visual de las ciudades. En paredes o marquesinas, el cartel tiene la capacidad de relacionarse visualmente con los ciudadanos de forma más que efectiva, generando ambientes comunicativos agradables y recurrentes. Es en el contexto institucional que la presente investigación posiciona la mirada sobre el cartel, analizando desde varias disciplinas el comportamiento de los elementos gráficos que conforman la superficie gráfica. El diseño de una herramienta de análisis que actúa sobre una muestra de carteles del Ayuntamiento de Barcelona, ofrece una serie de parámetros que permiten analizar la significación concreta del texto sobre la estructura gráfica del cartel institucional, visibilizando aquellas áreas relacionadas con el Diseño Gráfico más relevantes para extraer información útil para una investigación en diseño. Así, se ofrece un escenario de relación de diferentes ámbitos que permiten articular la transferencia de conocimiento en el ámbito del Diseño Gráfico y las Metodologías de Análisis Visual, necesarios para estudiar las peculiaridades significantes y funcionalidades visuales del cartel institucional barcelonés.
City streets are visually fertile environments for institutions to relate to their citizens through multiple media. The iruption of Information and Communication Technologies has derived this communication to digital platforms and social networks, but the urban visual environment is still a communication interaction space used by companies and public institutions to communicate with their users. The type of messages that we find in public space are multiple and diverse. Cities have become an environment inhabited by thousands of people who, during the day, transit through their streets and squares, thus being exposed to all kinds of communication impacts. Public institutions use this exposure as an opportunity to send messages to the citizens, through institutional campaigns, generating messages from diverse topics related to coexistence policies, social rights, housing rehabilitation, discrimination, culture, or festivities, among others. Of all the formats in which they are integrated into these campaigns, the poster is possibly the most enduring urban element in the visual coexistence of cities. On walls or marquesses, the poster has the ability to relate visually to citizens more than effectively, generating pleasant and recurring communication environments. It is in the institutional context that this research positions the gaze on the poster, analysing from various disciplines the behavior of the graphic elements that make up the graphical surface. The design of an analysis tool that acts on a sample of Barcelona City Council posters, offers a number of parameters that allow us to analyze the concrete meaning of the text on the graphical structure of the institutional poster, showing the areas related to Graphic Design more relevants to extract useful information for a design investigation. Thus, we offer a relationship scenario of different fields that enable the transfer of knowledge in the field of Graphic Design and Visual Methodologies, necessary to study the significant peculiarities and visual features of the Barcelona institutional poster.
Universitat Autònoma de Barcelona. Programa de Doctorat en Filosofia
Molley, Sean. "Ta Da! - The Text Adventure Design Assistant a Visual Tool for the Development of Adventure Games." TopSCHOLAR®, 1997. http://digitalcommons.wku.edu/theses/347.
Full textChan, Mei Ling. "Web-Based Usability Evaluation of Text-Resizing Methods and Users' Visual Fatigue on Online Reading Tasks." Thesis, California State University, Long Beach, 2017. http://pqdtopen.proquest.com/#viewpdf?dispub=10638681.
Full textA web-based usability evaluation was conducted to examine the effects of two text-resizing methods on normal and low vision Internet users’ online reading experience. Normal or corrected to normal vision (N/CTN; N = 50) and low vision (N = 5) participants completed two blocks of four mock Scholastic Aptitude Test (SAT) reading tests on the experiment website using fixed width (FW) and text reflow (TR) text-resizing methods. Subjective visual fatigue scores, modified System Usability Scale (SUS) score, time on task, and accuracy were collected as dependent measures for user experience evaluation. Results showed that when FW method was presented in the second block, N/CtN participants’ visual fatigue continued to increase, along with a significant decrease in time spent on reading tasks. Such effects were not observed in TR method. Results of low vision participants were different from those obtained in lab settings, suggesting that testing environment may have played a role in low vision participants’ performance. Implications of the current research can be used to guide future web accessibility research in regards to online experiment design and recruitment of participants with visual impairments.
Bartell, Nicole E. "But what do they think it means? : the text/image relationship in informational communication /." Online version of thesis, 2004. http://hdl.handle.net/1850/11799.
Full textKingston, Matthew Patrick. "(Re)inventing the Novel: Examining the Use of Text and Image in the Twenty-First Century Novel." Fogler Library, University of Maine, 2008. http://www.library.umaine.edu/theses/pdf/KingstonMP2008.pdf.
Full textBennett, Ronald W. "A comparison of combined and separate text-picture combinations on recall and application of information." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1997. http://www.collectionscanada.ca/obj/s4/f2/dsk2/ftp04/mq25822.pdf.
Full textRyrå, Landgren Isabella. "Samspel i det berättartekniska : text, bild och effekter i musikvideor." Thesis, Högskolan Väst, Avd för medier och design, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:hv:diva-8965.
Full textMusikvideor har under de senaste 50 åren varit en form av underhållning för vårt samhälle. Somliga formas för att spegla känslor medan andra visar upp artisten. Det finns de som baserar sig på låttexten för att skapa en kortare film eller gestalta låttextens innehåll. Med hjälp av tekniker som visuella effekter kan dessa drömlika och omöjliga världar och historier komma till liv. Det är videor med sådana effekter jag valt att analysera i denna uppsats med syftet att ta reda påhur stor roll de visuella effekterna spelar i berättandet. För att komma fram till detta har jag gjort en semiotisk studie fokuserad på analys och tolkningar av fem valda videor skapade under eller efter 2000-talet. CGI, slow-motion och metaforer är tekniker jag kollat på och det har visat sig att de alla bidrar till hur berättandet utspelas och uppfattas. Sambandet mellan bild och text i de valda videorna har pendlat mellan tolkning till bokstavligt översatt till varandra.
Douglas, James Maxwell Art College of Fine Arts UNSW. "Pictures and page numbers - image, text and formal structure in the visual book and the artist's book." Awarded by:University of New South Wales. School of Art, 2005. http://handle.unsw.edu.au/1959.4/23945.
Full textWilliams, Stephen E. "An Experimental Comparison of Middle School Students Motivation and Preference Toward Text and Graphic-Based Programming." DigitalCommons@USU, 2009. https://digitalcommons.usu.edu/etd/487.
Full textKrantz, Richard. "PIPETT 1." Thesis, Konstfack, Institutionen för Konst (K), 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:konstfack:diva-5784.
Full text