Dissertations / Theses on the topic 'Vidéo texte'

To see the other types of publications on this topic, follow the link: Vidéo texte.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Vidéo texte.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Ayache, Stéphane. "Indexation de documents vidéos par concepts par fusion de caractéristiques audio, vidéo et texte." Grenoble INPG, 2007. http://www.theses.fr/2007INPG0071.

Full text
Abstract:
Cette thèse s'inscrit dans le cadre de la Recherche d'Information et vise à une indexation sémantique des documents multimédias. Les recherches dans ce domaine sont confrontées au « fossé sémantique» qui sépare les descriptions brutes de différentes modalités des descriptions conceptuelles compréhensibles par les utilisateurs. Nous proposons un modèle d'indexation basé sur des réseaux d'opérateurs dans lesquels les flots de données, appelés numcepts, unifient des informations provenant des différentes modalités et extraites à différents niveaux d'abstractions. Nous présentons une instance de ce modèle où nous décrivons une typologie des opérateurs et des numcepts mis en œuvre. Nous avons mené des expérimentations sur les corpus TREC VIDEO afin d'évaluer l'impact de l'agencement et de l' implémentation des opérateurs sur la qualité de l'indexation des documents vidéos. Nous montrons qu'un réseau doit être décliné relativement à un concept afin d'optimiser la qualité de l'indexation
Work deals with information retrieval and aims to reach semantic indexing of multimediaIments. The state of the art approach tackle this problem by bridging of the semantic gap between level features, from each modality, and high-Ievel features (concepts), which are useful for humans. We propose an indexing model based on networks of operators into which data flows, called numcepts, unify informations from the various modalities and extracted at several level of abstraction. We present an instance of this model where we describe a topology of the operators and the numcepts we have deveIoped. We have conducted experiments on TREC VIDEO corpora in order to evaluate various organizations of the networks and the choice of the operators. We have studied those effects on performance of concept detection. We show that a network have to be designed with respect to the concepts in order to optimize the indexing performance
APA, Harvard, Vancouver, ISO, and other styles
2

Wehbe, Hassan. "Synchronisation automatique d'un contenu audiovisuel avec un texte qui le décrit." Thesis, Toulouse 3, 2016. http://www.theses.fr/2016TOU30104/document.

Full text
Abstract:
Nous abordons le problème de la synchronisation automatique d'un contenu audiovisuel avec une procédure textuelle qui le décrit. La stratégie consiste à extraire des informations sur la structure des deux contenus puis à les mettre en correspondance. Nous proposons deux outils d'analyse vidéo qui extraient respectivement : * les limites des évènements d'intérêt à l'aide d'une méthode de quantification de type dictionnaire * les segments dans lesquels une action se répète en exploitant une méthode d'analyse fréquentielle : le YIN. Ensuite, nous proposons un système de synchronisation qui fusionne les informations fournies par ces outils pour établir des associations entre les instructions textuelles et les segments vidéo correspondants. Une "Matrice de confiance" est construite et exploitée de manière récursive pour établir ces associations en regard de leur fiabilité
We address the problem of automatic synchronization of an audiovisual content with a procedural text that describes it. The strategy consists in extracting pieces of information about the structure from both contents, and in matching them depending on their types. We propose two video analysis tools that respectively extract: * Limits of events of interest using an approach inspired by dictionary quantization. * Segments that enclose a repeated action based on the YIN frequency analysis method. We then propose a synchronization system that merges results coming from these tools in order to establish links between textual instructions and the corresponding video segments. To do so, a "Confidence Matrix" is built and recursively processed in order to identify these links in respect with their reliability
APA, Harvard, Vancouver, ISO, and other styles
3

Yousfi, Sonia. "Embedded Arabic text detection and recognition in videos." Thesis, Lyon, 2016. http://www.theses.fr/2016LYSEI069/document.

Full text
Abstract:
Cette thèse s'intéresse à la détection et la reconnaissance du texte arabe incrusté dans les vidéos. Dans ce contexte, nous proposons différents prototypes de détection et d'OCR vidéo (Optical Character Recognition) qui sont robustes à la complexité du texte arabe (différentes échelles, tailles, polices, etc.) ainsi qu'aux différents défis liés à l'environnement vidéo et aux conditions d'acquisitions (variabilité du fond, luminosité, contraste, faible résolution, etc.). Nous introduisons différents détecteurs de texte arabe qui se basent sur l'apprentissage artificiel sans aucun prétraitement. Les détecteurs se basent sur des Réseaux de Neurones à Convolution (ConvNet) ainsi que sur des schémas de boosting pour apprendre la sélection des caractéristiques textuelles manuellement conçus. Quant à notre méthodologie d'OCR, elle se passe de la segmentation en traitant chaque image de texte en tant que séquence de caractéristiques grâce à un processus de scanning. Contrairement aux méthodes existantes qui se basent sur des caractéristiques manuellement conçues, nous proposons des représentations pertinentes apprises automatiquement à partir des données. Nous utilisons différents modèles d'apprentissage profond, regroupant des Auto-Encodeurs, des ConvNets et un modèle d'apprentissage non-supervisé, qui génèrent automatiquement ces caractéristiques. Chaque modèle résulte en un système d'OCR bien spécifique. Le processus de reconnaissance se base sur une approche connexionniste récurrente pour l'apprentissage de l'étiquetage des séquences de caractéristiques sans aucune segmentation préalable. Nos modèles d'OCR proposés sont comparés à d'autres modèles qui se basent sur des caractéristiques manuellement conçues. Nous proposons, en outre, d'intégrer des modèles de langage (LM) arabes afin d'améliorer les résultats de reconnaissance. Nous introduisons différents LMs à base des Réseaux de Neurones Récurrents capables d'apprendre des longues interdépendances linguistiques. Nous proposons un schéma de décodage conjoint qui intègre les inférences du LM en parallèle avec celles de l'OCR tout en introduisant un ensemble d’hyper-paramètres afin d'améliorer la reconnaissance et réduire le temps de réponse. Afin de surpasser le manque de corpus textuels arabes issus de contenus multimédia, nous mettons au point de nouveaux corpus manuellement annotés à partir des flux TV arabes. Le corpus conçu pour l'OCR, nommé ALIF et composée de 6,532 images de texte annotées, a été publié a des fins de recherche. Nos systèmes ont été développés et évalués sur ces corpus. L’étude des résultats a permis de valider nos approches et de montrer leurs efficacité et généricité avec plus de 97% en taux de détection, 88.63% en taux de reconnaissance mots sur le corpus ALIF dépassant ainsi un des systèmes d'OCR commerciaux les mieux connus par 36 points
This thesis focuses on Arabic embedded text detection and recognition in videos. Different approaches robust to Arabic text variability (fonts, scales, sizes, etc.) as well as to environmental and acquisition condition challenges (contrasts, degradation, complex background, etc.) are proposed. We introduce different machine learning-based solutions for robust text detection without relying on any pre-processing. The first method is based on Convolutional Neural Networks (ConvNet) while the others use a specific boosting cascade to select relevant hand-crafted text features. For the text recognition, our methodology is segmentation-free. Text images are transformed into sequences of features using a multi-scale scanning scheme. Standing out from the dominant methodology of hand-crafted features, we propose to learn relevant text representations from data using different deep learning methods, namely Deep Auto-Encoders, ConvNets and unsupervised learning models. Each one leads to a specific OCR (Optical Character Recognition) solution. Sequence labeling is performed without any prior segmentation using a recurrent connectionist learning model. Proposed solutions are compared to other methods based on non-connectionist and hand-crafted features. In addition, we propose to enhance the recognition results using Recurrent Neural Network-based language models that are able to capture long-range linguistic dependencies. Both OCR and language model probabilities are incorporated in a joint decoding scheme where additional hyper-parameters are introduced to boost recognition results and reduce the response time. Given the lack of public multimedia Arabic datasets, we propose novel annotated datasets issued from Arabic videos. The OCR dataset, called ALIF, is publicly available for research purposes. As the best of our knowledge, it is first public dataset dedicated for Arabic video OCR. Our proposed solutions were extensively evaluated. Obtained results highlight the genericity and the efficiency of our approaches, reaching a word recognition rate of 88.63% on the ALIF dataset and outperforming well-known commercial OCR engine by more than 36%
APA, Harvard, Vancouver, ISO, and other styles
4

Bull, Hannah. "Learning sign language from subtitles." Electronic Thesis or Diss., université Paris-Saclay, 2023. http://www.theses.fr/2023UPASG013.

Full text
Abstract:
Les langues des signes sont un moyen de communication essentiel pour les communautés sourdes. Elles sont des langues visuo-gestuelles, qui utilisent comme modalités les mains, les expressions faciales, le regard et les mouvements du corps. Elles ont des structures grammaticales complexes et des lexiques riches qui sont considérablement différents de ceux que l'on trouve dans les langues parlées. Les spécificités des langues des signes en termes de canaux de communication, de structure et de grammaire exigent des méthodologies distinctes. Les performances des systèmes de traduction automatique entre des langues écrites ou parlées sont actuellement suffisantes pour de nombreux cas d'utilisation quotidienne, tels que la traduction de vidéos, de sites web, d'e-mails et de documents. En revanche, les systèmes de traduction automatique pour les langues des signes n'existent pas en dehors de cas d'utilisation très spécifiques avec un vocabulaire limité. La traduction automatique de langues des signes est un défi pour deux raisons principales. Premièrement, les langues des signes sont des langues à faibles ressources avec peu de données d'entraînement disponibles. Deuxièmement, les langues des signes sont des langues visuelles et spatiales sans forme écrite, naturellement représentées sous forme de vidéo plutôt que d'audio ou de texte. Pour relever le premier défi, nous fournissons de grands corpus de données pour l'entraînement et l'évaluation des systèmes de traduction automatique en langue des signes, avec des contenus vidéo en langue des signes interprétée et originale, ainsi que des sous-titres écrits. Alors que les données interprétées nous permettent de collecter un grand nombre d'heures de vidéos, les vidéos originalement en langue des signes sont plus représentatives de l'utilisation de la langue des signes au sein des communautés sourdes. Les sous-titres écrits peuvent être utilisés comme supervision faible pour diverses tâches de compréhension de la langue des signes. Pour relever le deuxième défi, cette thèse propose des méthodes permettant de mieux comprendre les vidéos en langue des signes. Alors que la segmentation des phrases est généralement triviale pour les langues écrites, la segmentation des vidéos en langue des signes en phrases repose sur la détection d'indices sémantiques et prosodiques subtils dans les vidéos. Nous utilisons des indices prosodiques pour apprendre à segmenter automatiquement une vidéo en langue des signes en unités de type phrase, déterminées par les limites des sous-titres. En développant cette méthode de segmentation, nous apprenons ensuite à aligner les sous-titres du texte sur les segments de la vidéo en langue des signes en utilisant des indices sémantiques et prosodiques, afin de créer des paires au niveau de la phrase entre la vidéo en langue des signes et le texte. Cette tâche est particulièrement importante pour les données interprétées, où les sous-titres sont généralement alignés sur l'audio et non sur la langue des signes. En utilisant ces paires vidéo-texte alignées automatiquement, nous développons et améliorons plusieurs méthodes différentes pour annoter de façon dense les signes lexicaux en interrogeant des mots dans le texte des sous-titres et en recherchant des indices visuels dans la vidéo en langue des signes pour les signes correspondants
Sign languages are an essential means of communication for deaf communities. Sign languages are visuo-gestual languages using the modalities of hand gestures, facial expressions, gaze and body movements. They possess rich grammar structures and lexicons that differ considerably from those found among spoken languages. The uniqueness of transmission medium, structure and grammar of sign languages requires distinct methodologies. The performance of automatic translations systems between high-resource written languages or spoken languages is currently sufficient for many daily use cases, such as translating videos, websites, emails and documents. On the other hand, automatic translation systems for sign languages do not exist outside of very specific use cases with limited vocabulary. Automatic sign language translation is challenging for two main reasons. Firstly, sign languages are low-resource languages with little available training data. Secondly, sign languages are visual-spatial languages with no written form, naturally represented as video rather than audio or text. To tackle the first challenge, we contribute large datasets for training and evaluating automatic sign language translation systems with both interpreted and original sign language video content, as well as written text subtitles. Whilst interpreted data allows us to collect large numbers of hours of videos, original sign language video is more representative of sign language usage within deaf communities. Written subtitles can be used as weak supervision for various sign language understanding tasks. To address the second challenge, we develop methods to better understand visual cues from sign language video. Whilst sentence segmentation is mostly trivial for written languages, segmenting sign language video into sentence-like units relies on detecting subtle semantic and prosodic cues from sign language video. We use prosodic cues to learn to automatically segment sign language video into sentence-like units, determined by subtitle boundaries. Expanding upon this segmentation method, we then learn to align text subtitles to sign language video segments using both semantic and prosodic cues, in order to create sentence-level pairs between sign language video and text. This task is particularly important for interpreted TV data, where subtitles are generally aligned to the audio and not to the signing. Using these automatically aligned video-text pairs, we develop and improve multiple different methods to densely annotate lexical signs by querying words in the subtitle text and searching for visual cues in the sign language video for the corresponding signs
APA, Harvard, Vancouver, ISO, and other styles
5

Couture, Matte Robin. "Digital games and negotiated interaction : integrating Club Penguin Island into two ESL grade 6 classes." Master's thesis, Université Laval, 2019. http://hdl.handle.net/20.500.11794/35458.

Full text
Abstract:
Cette étude avait pour objectif d’explorer l’interaction entre de jeunes apprenants (11-12 ans) lors de tâches communicatives accomplies face à face et supportées par Club Penguin Island, un jeu de rôle en ligne massivement multijoueur (MMORPG). Les questions de recherche étaient triples: évaluer la présence d’épisodes de centration sur la forme (FFEs) lors des tâches communicatives accomplies avec Club Penguin Island et identifier leurs caractéristiques; évaluer l’impact de différents types de tâches sur la présence de FFEs; et examiner les attitudes des participants. Ce projet de recherche a été réalisé auprès de 20 élèves de 6ième année en anglais intensif dans la province de Québec. Les participants ont exécuté une tâche de type «information gap», et deux tâches de type «reasoninggap» dont une incluant une composante écrite. Les tâches ont été réalisées en dyades et les enregistrements des interactions ont été transcrits et analysés pour identifier la présence de FFEs et leurs caractéristiques. Une analyse statistique fut utilisée pour évaluer l’impact du type de tâche sur la présence de FFEs, et un questionnaire a été administré pour examiner les attitudes des participants à la suite des tâches. Les résultats révèlent que des FFEs ont été produits lors des tâches accomplies avec le MMORPG, que les participants ont pu négocier les interactions sans l’aide de l’instructeur et que la majorité des FFEs visaient des mots trouvés dans les tâches et le jeu. L’analyse statistique démontre l’influence du type de tâche, puisque davantage de FFEs ont été produits lors de la tâche de type «information gap» que l’une des tâches de type «reasoning gap». Le questionnaire sur les attitudes révèle qu’elles ont été positives. Les implications pédagogiques soulignent l’impact des MMORPGs pour l’acquisition des langues secondes et les conclusions ajoutent à la littérature limitée sur les interactions entre de jeunes apprenants.
The objective of the present study was to explore negotiated interaction involving young children (age 11-12) who carried out communicative tasks supported by Club Penguin Island, a massively multiplayer online role-playing game (MMORPG). Unlike previous studies involving MMORPGs, the present study assessed the use of Club Penguin Island in the context of face-to-face interaction. More specifically, the research questions were three-fold: assess the presence focus-on-form episodes (FFEs) during tasks carried out with Club Penguin Island and identify their characteristics; evaluate the impact of task type on the presence of FFEs; and survey the attitudes of participants. The research project was carried out with 20 Grade 6 intensive English as a second language (ESL) students in the province of Quebec. The participants carried out one information-gap task and two reasoning-gap tasks including one with a writing component. The tasks were carriedout in dyads, and recordings were transcribed and analyzed to identify the presence of FFEs and their characteristics. A statistical analysis was used to assess the impact of task type on the presence of FFEs, and a questionnaire was administered to assess the attitudes of participants following the completion of all tasks. Findings revealed that carrying out tasks with the MMORPG triggered FFEs, that participants were able to successfully negotiate interaction without the help of the instructor, and that most FFEs were focused on the meaning of vocabulary found in the tasks and game. The statistical analysis showed the influence of task type since more FFEs were produced during the information-gap task than one of the reasoning-gap tasks. The attitude questionnaire revealed positive attitudes, which was in line with previous researchon digital games for language learning. Pedagogical implications point to the impact of MMORPGs for language learning and add to the scarce literature on negotiated interaction with young learners.
APA, Harvard, Vancouver, ISO, and other styles
6

Sidevåg, Emmilie. "Användarmanual text vs video." Thesis, Linnéuniversitetet, Institutionen för datavetenskap, fysik och matematik, DFM, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-17617.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Salway, Andrew. "Video annotation : the role of specialist text." Thesis, University of Surrey, 1999. http://epubs.surrey.ac.uk/843350/.

Full text
Abstract:
Digital video is among the most information-intensive modes of communication. The retrieval of video from digital libraries, along with sound and text, is a major challenge for the computing community in general and for the artificial intelligence community specifically. The advent of digital video has set some old questions in a new light. Questions relating to aesthetics and to the role of surrogates - image for reality and text for image, invariably touch upon the link between vision and language. Dealing with this link computationally is important for the artificial intelligence enterprise. Interesting images to consider both aesthetically and for research in video retrieval include those which are constrained and patterned, and which convey rich meanings; for example, dance. These are specialist images for us and require a special language for description and interpretation. Furthermore, they require specialist knowledge to be understood since there is usually more than meets the untrained eye: this knowledge may also be articulated in the language of the specialism. In order to be retrieved effectively and efficiently, video has to be annotated-, particularly so for specialist moving images. Annotation involves attaching keywords from the specialism along with, for us, commentaries produced by experts, including those written and spoken specifically for annotation and those obtained from a corpus of extant texts. A system that processes such collateral text for video annotation should perhaps be grounded in an understanding of the link between vision and language. This thesis attempts to synthesise ideas from artificial intelligence, multimedia systems, linguistics, cognitive psychology and aesthetics. The link between vision and language is explored by focusing on moving images of dance and the special language used to describe and interpret them. We have developed an object-oriented system, KAB, which helps to annotate a digital video library with a collateral corpus of texts and terminology. User evaluation has been encouraging. The system is now available on the WWW.
APA, Harvard, Vancouver, ISO, and other styles
8

Smith, Gregory. "VIDEO SCENE DETECTION USING CLOSED CAPTION TEXT." VCU Scholars Compass, 2009. http://scholarscompass.vcu.edu/etd/1932.

Full text
Abstract:
Issues in Automatic Video Biography Editing are similar to those in Video Scene Detection and Topic Detection and Tracking (TDT). The techniques of Video Scene Detection and TDT can be applied to interviews to reduce the time necessary to edit a video biography. The system has attacked the problems of extraction of video text, story segmentation, and correlation. This thesis project was divided into three parts: extraction, scene detection, and correlation. The project successfully detected scene breaks in series television episodes and displayed scenes that had similar content.
APA, Harvard, Vancouver, ISO, and other styles
9

Zhang, Jing. "Extraction of Text Objects in Image and Video Documents." Scholar Commons, 2012. http://scholarcommons.usf.edu/etd/4266.

Full text
Abstract:
The popularity of digital image and video is increasing rapidly. To help users navigate libraries of image and video, Content Based Information Retrieval (CBIR) system that can automatically index image and video documents are needed. However, due to the semantic gap between low-level machine descriptors and high-level semantic descriptors, the existing CBIR systems are still far from perfect. Text embedded in multi-media data, as a well-defined model of concepts for humans' communication, contains much semantic information related to the content. This text information can provide a much truer form of content-based access to the image and video documents if it can be extracted and harnessed efficiently. This dissertation solves the problem involved in detecting text object in image and video and tracking text event in video. For text detection problem, we propose a new unsupervised text detection algorithm. A new text model is constructed to describe text object using pictorial structure. Each character is a part in the model and every two neighboring characters are connected by a spring-like link. Two characters and the link connecting them are defined as a text unit. We localize candidate parts by extracting closed boundaries and initialize the links by connecting two neighboring candidate parts based on the spatial relationship of characters. For every candidate part, we compute character energy using three new character features, averaged angle difference of corresponding pairs, fraction of non-noise pairs, and vector of stroke width. They are extracted based on our observation that the edge of a character can be divided into two sets with high similarities in length, curvature, and orientation. For every candidate link, we compute link energy based on our observation that the characters of a text typically align along certain direction with similar color, size, and stroke width. For every candidate text unit, we combine character and link energies to compute text unit energy which indicates the probability that the candidate text model is a real text object. The final text detection results are generated using a text unit energy based thresholding. For text tracking problem, we construct a text event model by using pictorial structure as well. In this model, the detected text object in each video frame is a part and two neighboring text objects of a text event are connected by a spring-like link. Inter-frame link energy is computed for each link based on the character energy, similarity of neighboring text objects, and motion information. After refining the model using inter-frame link energy, the remaining text event models are marked as text events. At character level, because the proposed method is based on the assumption that the strokes of a character have uniform thickness, it can detect and localize characters from different languages in different styles, such as typewritten text or handwriting text, if the characters have approximately uniform stroke thickness. At text level, however, because the spatial relationship between two neighboring characters is used to localize text objects, the proposed method may fail to detect and localize the characters with multiple separate strokes or connected characters. For example, some East Asian language characters, such as Chinese, Japanese, and Korean, have many strokes of a single character. We need to group the strokes first to form single characters and then group characters to form text objects. While, the characters of some languages, such Arabic and Hindi, are connected together, we cannot extract spatial information between neighboring characters since they are detected as a single character. Therefore, in current stage the proposed method can detect and localize the text objects that are composed of separate characters with connected strokes with approximately uniform thickness. We evaluated our method comprehensively using three English language-based image and video datasets: ICDAR 2003/2005 text locating dataset (258 training images and 251 test images), Microsoft Street View text detection dataset (307 street view images), and VACE video dataset (50 broadcast news videos from CNN and ABC). The experimental results demonstrate that the proposed text detection method can capture the inherent properties of text and discriminate text from other objects efficiently.
APA, Harvard, Vancouver, ISO, and other styles
10

Zipstein, Marc. "Les Méthodes de compression de textes : algorithmes et performances." Paris 7, 1990. http://www.theses.fr/1990PA077107.

Full text
Abstract:
La compression de textes a pour but de réduire le nombre de symboles nécessaires à la représentation d'un texte. L'objet de cette thèse est l'étude, la mise au point et la comparaison de méthodes de compression universelles, c'est-à-dire capables de traiter de manière efficace n'importe quel type de textes. Nous montrons que l'utilisation d'automates permet d'augmenter l'efficacité des méthodes de compression classiques et nous présentons une nouvelle méthode basée sur l'utilisation de l'automate des facteurs. Nous présentons les deux grandes classes d'algorithmes de compression de données : les algorithmes de codages statistiques et les algorithmes de codage par facteurs. Les algorithmes de codage statistique traitent les textes par blocs de longueur fixe, un bloc fréquent ayant une traduction courte. Nous présentons les codages de Huffman statique et adaptatif ainsi que le codage arithmétique. Nous proposons une représentation du codage arithmétique à l'aide d'un transducteur ce qui garantit un traitement en temps réel. Les codages par facteur traduisent les textes en utilisant leurs propres facteurs. Nous présentons les algorithmes de codage dus à Ziv et Lempel, et nous décrivons une nouvelle méthode basée sur l'utilisation de l'automate des facteurs. Ce travail se termine par la comparaison des performances des algorithmes décrits
APA, Harvard, Vancouver, ISO, and other styles
11

Štindlová, Marie. "Museli to založit." Master's thesis, Vysoké učení technické v Brně. Fakulta výtvarných umění, 2015. http://www.nusl.cz/ntk/nusl-232451.

Full text
APA, Harvard, Vancouver, ISO, and other styles
12

Sjölund, Jonathan. "Detection of Frozen Video Subtitles Using Machine Learning." Thesis, Linköpings universitet, Datorseende, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-158239.

Full text
Abstract:
When subtitles are burned into a video, an error can sometimes occur in the encoder that results in the same subtitle being burned into several frames, resulting in subtitles becoming frozen. This thesis provides a way to detect frozen video subtitles with the help of an implemented text detector and classifier. Two types of classifiers, naïve classifiers and machine learning classifiers, are tested and compared on a variety of different videos to see how much a machine learning approach can improve the performance. The naïve classifiers are evaluated using ground truth data to gain an understanding of the importance of good text detection. To understand the difficulty of the problem, two different machine learning classifiers are tested, logistic regression and random forests. The result shows that machine learning improves the performance over using naïve classifiers by improving the specificity from approximately 87.3% to 95.8% and improving the accuracy from 93.3% to 95.5%. Random forests achieve the best overall performance, but the difference compared to when using logistic regression is small enough that more computationally complex machine learning classifiers are not necessary. Using the ground truth shows that the weaker naïve classifiers would be improved by at least 4.2% accuracy, thus a better text detector is warranted. This thesis shows that machine learning is a viable option for detecting frozen video subtitles.
APA, Harvard, Vancouver, ISO, and other styles
13

Wolf, Christian. "Détection de textes dans des images issues d'un flux vidéo pour l'indexation sémantique." Lyon, INSA, 2003. http://theses.insa-lyon.fr/publication/2003ISAL0074/these.pdf.

Full text
Abstract:
Ce travail entre dans le cadre de l'indexation d'images et de vidéos. Les systèmes disponibles pour chercher dans les bases des documents audiovisuels travaillent sans connaissance, ils utilisent des méthodes de traitement d'image pour extraire des caractéristiques de bas niveau. Nous utilisons le texte présent dans les images et les vidéos. Les méthodes de détection de texte présentées dans la littérature sont très simples : la plupart sont basées sur l'estimation de la texture ou sur la détection des contours suivie par l'accumulation de ces caractéristiques. Nous proposons la prise en compte des caractéristiques géométriques directement dans la phase de détection. Une première détection grossière sert à calculer une image de probabilité de texte : ensuite, pour chaque pixel, nous calculons une estimation robuste des caractéristiques géométriques de la boite de texte de laquelle elle fait éventuellement partie. Ces caractéristiques sont rajoutées aux caractéristiques de la première étape de détection. L'apprentissage se fait avec un classificateur de type "Support Vector Machines". Pour la segmentation des caractères nous proposons deux algorithmes différents : le premier algorithme est basé sur la maximisation d'un critère de contraste ; la deuxième approche exploite des connaissances a priori sur la répartition locale des pixels "texte" et "non-texte" pour aider à la décision de seuillage. Un modèle statistique (en utilisant un modèle de champs de Markov) est élaboré et intégré dans un modèle bayésien d'estimation pour obtenir une estimation de l'image originale binaire
This work situates within the framework of image and video indexation. A way to include semantic knowledge into the indexing process is to use the text included in the images and video sequences. It is rich in information but easy to use. Existing methods for text detection are simple: most of them are based on texture estimation or edge detection followed by an accumulation of these characteristics. We suggest the usage of geometrical features very early in the detection chain: a first coarse detection calculates a text "probability" image. Afterwards, for each pixel we calculate geometrical properties of the eventual surrounding text rectangle, which are added to the features of the first step and fed into a support vector machine classifier. For the application to video sequences, we propose an algorithm which detects text on a frame by frame basis, tracking the found text rectangles across multiple frames and integrating the frame robustly into a single image. We tackle the character segmentation problem and suggest two different methods: the first algorithm maximizes a criterion based on the local contrast in the image. The second approach exploits a priori knowledge on the spatial binary distribution of the pixels. This prior knowledge in the form of a Markov random field model is integrated into Bayesian estimation framework in order to obtain an estimation of the original binary image
APA, Harvard, Vancouver, ISO, and other styles
14

Bird, Paul. "Elementary students' comprehension of computer presented text." Thesis, University of British Columbia, 1990. http://hdl.handle.net/2429/29187.

Full text
Abstract:
The study investigated grade 6 students' comprehension of narrative text when presented on a computer and as printed words on paper. A set of comprehension tests were developed for three stories of varying length (382 words, 1047 words and 1933 words) using a skills hierarchy protocol. The text for each story was prepared for presentation on a Macintosh computer using a program written for the study and as print in the form of exact copies of the computer screen. Students from two grade 6 classes in a suburban elementary school were randomly assigned to read one of the stories in either print form or on the computer and subsequently completed a comprehension test as well as a questionnaire concerning attitude and personal information. The responses from the comprehension tests were evaluated by graduate students in Language Education. The data evolved from the tests and questionnaires were analysed to determine measures of test construct validity, inter-rater reliability, and any significant difference in the means of comprehension scores for the two experimental groups for each story. The results indicated small but insignificant differences between the means of the three comprehension test scores for computer and print. A number of students reading from the computer complained of eye fatigue. The scores of subjects reading the longest story and complaining of eye fatigue were significantly lower.
Education, Faculty of
Curriculum and Pedagogy (EDCP), Department of
Graduate
APA, Harvard, Vancouver, ISO, and other styles
15

Chen, Datong. "Text detection and recognition in images and video sequences /." [S.l.] : [s.n.], 2003. http://library.epfl.ch/theses/?display=detail&nr=2863.

Full text
APA, Harvard, Vancouver, ISO, and other styles
16

Sharma, Nabin. "Multi-lingual Text Processing from Videos." Thesis, Griffith University, 2015. http://hdl.handle.net/10072/367489.

Full text
Abstract:
Advances in digital technology have produced low priced portable imaging devices such as digital cameras attached to mobile phones, camcorders, PDA’s etc. which are highly portable. These devices can be used to capture videos and images at ease, which can be shared through the internet and other communication media. In the commercial do- main, cameras are used to create news, advertisement videos and other forms of material for information communication. The use of multiple languages to create information for targeted audiences is quite common in countries having multiple official languages. Trans- mission of news, advertisement videos and images across various communication channels has created large databases of videos and these are increasing exponentially. Effective management of such databases requires proper indexing for the retrieval of relevant in- formation. Text information is dominant in most of the videos and images, which can be used as keywords for retrieval of relevant video and images. Automatic annotation of videos and images to extract keywords requires the text to be converted to an editable form. This thesis addresses the problem of multi-lingual text processing from video frames. Multi-lingual text processing involves text detection, word segmentation, script identification, and text recognition. Additionally, text frame classification is required to avoid processing a video frame which does not contain text information. A new multi-lingual video word dataset was created and published as a part of the current research. The dataset comprises words of ten scripts, namely English (Roman), Hindi (Devanagari), Bengali (Bangla), Arabic, Oriya, Gujrathi, Punjabi, Kannada, Tamil and Telugu. This dataset was created to facilitate future research on multi-lingual text recognition.
Thesis (PhD Doctorate)
Doctor of Philosophy (PhD)
School of Information and Communication Technology.
Science, Environment, Engineering and Technology
Full Text
APA, Harvard, Vancouver, ISO, and other styles
17

Minetto, Rodrigo. "Reconnaissance de zones de texte et suivi d'objets dans les images et les vidéos." Paris 6, 2012. http://www.theses.fr/2012PA066108.

Full text
Abstract:
Dans cette thèse, nous abordons trois problèmes de vision par ordinateur: (1) la détection et la reconnaissance d'objets de texte dans des images de scènes réelles; (2) le suivi de ces objets de texte dans une vidéo numérique, et (3) le suivi d'objets 3D rigides et arbitraires avec des amers connus dans une vidéo numérique. Pour chaque problème, nous avons développé des algorithmes innovants, qui sont au moins aussi précis et robustes que les algorithmes de l'état de l’art. Plus précisément, pour la reconnaissance de texte, nous avons développé (et largement évalué) un nouveau descripteur basé sur HOG, et dédié au traitement du texte Roman, baptisé T-HOG. Nous avons montré sa valeur en tant que post-filtre pour un détecteur de texte existant (SnooperText). Nous avons également amélioré l’algorithme SnooperText en développant une approche multi-échelle pour traiter des caractères de taille très différentes tout en limitant la sensibilité de l'algorithme aux différents artéfacts. Pour le suivi des objets de textes, nous avons décrit quatre manières de combiner la détection et le suivi, et nous avons développé un tracker particulier, basé sur un filtre particulaire exploitant le T-HOG. Pour le suivi des objets rigides, nous avons développé un nouvel algorithme précis et robuste (AFFTRACK) qui combine le KLT tracker avec une calibration améliorée de la caméra. Nous avons largement testé nos algorithmes sur plusieurs bases de données de la littérature. Nous avons également créé plusieurs bases de données (publiquement disponibles) pour l’évaluation des algorithmes de détection, suivi de textes et de suivi d'objets rigides dans les vidéos
In this thesis we address three computer vision problems: (1) the detection and recognition of flat text objects in images of real scenes; (2) the tracking of such text objects in a digital video; and (3) the tracking an arbitrary three-dimensional rigid object with known markings in a digital video. For each problem we developed innovative algorithms, which are at least as accurate and robust as other state-of-the-art algorithms. Specifically, for text recognition we developed (and extensively evaluated) a new HOG-based descriptor specialized for Roman script, which we call T-HOG, and showed its value as a post-filter for an existing text detector (SnooperText). We also improved the SnooperText algorithm by using the multi-scale technique to handle widely different letter sizes while limiting the sensitivity of the algorithm to various artifacts. For text tracking, we describe four basic ways of combining a text detector and a text tracker, and we developed a specific tracker based on a particle-filter which exploits the T-HOG recognizer. For rigid object tracking we developed a new accurate and robust algorithm (AFFTrack) that combines the KLT feature tracker with an improved camera calibration procedure. We extensively tested our algorithms on several benchmarks well-known in the literature. We also created benchmarks (publicly available) for the evaluation of text detection and tracking and rigid object tracking algorithms
APA, Harvard, Vancouver, ISO, and other styles
18

Fraz, Muhammad. "Video content analysis for intelligent forensics." Thesis, Loughborough University, 2014. https://dspace.lboro.ac.uk/2134/18065.

Full text
Abstract:
The networks of surveillance cameras installed in public places and private territories continuously record video data with the aim of detecting and preventing unlawful activities. This enhances the importance of video content analysis applications, either for real time (i.e. analytic) or post-event (i.e. forensic) analysis. In this thesis, the primary focus is on four key aspects of video content analysis, namely; 1. Moving object detection and recognition, 2. Correction of colours in the video frames and recognition of colours of moving objects, 3. Make and model recognition of vehicles and identification of their type, 4. Detection and recognition of text information in outdoor scenes. To address the first issue, a framework is presented in the first part of the thesis that efficiently detects and recognizes moving objects in videos. The framework targets the problem of object detection in the presence of complex background. The object detection part of the framework relies on background modelling technique and a novel post processing step where the contours of the foreground regions (i.e. moving object) are refined by the classification of edge segments as belonging either to the background or to the foreground region. Further, a novel feature descriptor is devised for the classification of moving objects into humans, vehicles and background. The proposed feature descriptor captures the texture information present in the silhouette of foreground objects. To address the second issue, a framework for the correction and recognition of true colours of objects in videos is presented with novel noise reduction, colour enhancement and colour recognition stages. The colour recognition stage makes use of temporal information to reliably recognize the true colours of moving objects in multiple frames. The proposed framework is specifically designed to perform robustly on videos that have poor quality because of surrounding illumination, camera sensor imperfection and artefacts due to high compression. In the third part of the thesis, a framework for vehicle make and model recognition and type identification is presented. As a part of this work, a novel feature representation technique for distinctive representation of vehicle images has emerged. The feature representation technique uses dense feature description and mid-level feature encoding scheme to capture the texture in the frontal view of the vehicles. The proposed method is insensitive to minor in-plane rotation and skew within the image. The capability of the proposed framework can be enhanced to any number of vehicle classes without re-training. Another important contribution of this work is the publication of a comprehensive up to date dataset of vehicle images to support future research in this domain. The problem of text detection and recognition in images is addressed in the last part of the thesis. A novel technique is proposed that exploits the colour information in the image for the identification of text regions. Apart from detection, the colour information is also used to segment characters from the words. The recognition of identified characters is performed using shape features and supervised learning. Finally, a lexicon based alignment procedure is adopted to finalize the recognition of strings present in word images. Extensive experiments have been conducted on benchmark datasets to analyse the performance of proposed algorithms. The results show that the proposed moving object detection and recognition technique superseded well-know baseline techniques. The proposed framework for the correction and recognition of object colours in video frames achieved all the aforementioned goals. The performance analysis of the vehicle make and model recognition framework on multiple datasets has shown the strength and reliability of the technique when used within various scenarios. Finally, the experimental results for the text detection and recognition framework on benchmark datasets have revealed the potential of the proposed scheme for accurate detection and recognition of text in the wild.
APA, Harvard, Vancouver, ISO, and other styles
19

Zheng, Yilin. "Text-Based Speech Video Synthesis from a Single Face Image." The Ohio State University, 2019. http://rave.ohiolink.edu/etdc/view?acc_num=osu1572168353691788.

Full text
APA, Harvard, Vancouver, ISO, and other styles
20

Gokturk, Ozkan Ziya. "Metadata Extraction From Text In Soccer Domain." Master's thesis, METU, 2008. http://etd.lib.metu.edu.tr/upload/12609871/index.pdf.

Full text
Abstract:
Video databases and content based retrieval in these databases have become popular with the improvements in technology. Metadata extraction techniques are used for providing data to video content. One popular metadata extraction technique for mul- timedia is information extraction from text. For some domains, it is possible to &
#64257
nd accompanying text with the video, such as soccer domain, movie domain and news domain. In this thesis, we present an approach of metadata extraction from match reports for soccer domain. The UEFA Cup and UEFA Champions League Match Reports are downloaded from the web site of UEFA by a web-crawler. These match reports are preprocessed by using regular expressions and then important events are extracted by using hand-written rules. In addition to hand-written rules, two di&
#64256
erent machine learning techniques are applied on match corpus to learn event patterns and automatically extract match events. Extracted events are saved in an MPEG-7 &
#64257
le. A user interface is implemented to query the events in the MPEG-7 match corpus and view the corresponding video segments.
APA, Harvard, Vancouver, ISO, and other styles
21

Gasser, Wolfgang. "„Das Ende (m)einer Kindheit?“: Wissenschaft und Selbstbezüge – Jugendliche analysieren Texte und Video-Interviews zu Kindertransporten." HATiKVA e.V. – Die Hoffnung Bildungs- und Begegnungsstätte für Jüdische Geschichte und Kultur Sachsen, 2015. https://slub.qucosa.de/id/qucosa%3A34939.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

Martin, Thomas. "Vers une reconnaissance multimodale du texte et de la parole pour l'analyse de documents vidéos pédagogiques." La Rochelle, 2009. http://www.theses.fr/2009LAROS264.

Full text
Abstract:
Cette thèse s'intéresse à la mise en œuvre de méthodes de reconnaissance multimodale du texte et de la parole dans des contenus audiovisuels. Elle se focalise en particulier sur les enregistrements de cours présentiels, dans lesquels est fait un usage intensif de l'écrit et de la parole. Avec l'augmentation massive de la production de données multimédias, l'accès à ces dernières devient problématique et doit passer par une indexation efficace des contenus. En particulier, il est nécessaire de tenir compte de la nature hétérogène de l'information présente. C'est à cet objectif que tentent de répondre le paradigme de l'analyse multimodale et les méthodes d'analyse qui s'y rapportent. Il convient cependant de constater qu'en raison de l'émergence récente de l'analyse multimodale, il n'y a eu que peu de tentatives de définition de ce domaine. Par ailleurs, peu de travaux se sont intéressés à l'interaction entre texte et parole dans les flux multimédias et à la prise en compte de cette interaction pour leur extraction. Notre contribution s'axe donc sur deux points. En premier lieu, nous souhaitons combler le manque de définition en proposant un modèle de l'analyse multimodale. Son objectif est de disposer d'un cadre permettant une meilleure description des applications recourant à l'analyse multimodale, notamment en définissant précisément les concepts de modalité et de multimodalité. Le second point de notre contribution est axé sur la reconnaissance multimodale du texte et de la parole. Nous procédons pour cela à une comparaison des processus de reconnaissance du texte et de la parole pour ensuite étudier deux cas de collaboration texte-parole. Le premier n'implique pas un processus de reconnaissance du texte mais repose sur la constitution d'un modèle de langage thématisé à partir des ressources textuelles du cours pour améliorer la reconnaissance de la parole. Malgré la petite taille de ce corpus, nous montrons une amélioration significative des résultats de reconnaissance. Nous expérimentons enfin une méthode de reconnaissance croisée du texte et de la parole basée sur la complémentarité des représentations écrite et phonétique du langage. Nous montrons que cela permet d'améliorer les résultats de reconnaissance du texte, voire même, à terme, de faire ressortir le vocabulaire spécialisé du cours
This work focuses on the implementation of methods for multimodal recognition of text and speech in audiovisual content. It focuses in particular on lecture records, in which text and speech are extensively used. As the production of multimedia data is massively increasing, the access to these data becomes problematic and requires an efficient content indexing. It is necessary to take into account the heterogeneous nature of this information which is the aim of the paradigm of multimodal analysis. It should however be noted that due to the recent emergence of multimodal analysis, there have been only few attempts to define this field. In addition, only few studies have focused on the interaction between text and speech in multimedia stream and the use of this interaction for their extraction. Our contribution focuses on two points. First, we wish to address the lack of definition by proposing a model of the multimodal analysis. Its goal is to propose a framework for a better description of applications using multimodal analysis, including clearly defining the concepts of modality and multimodality. The second point of our contribution is focused on the multimodal recognition of text and speech. We compare text and speech recognition processes then consider two cases of text-speech collaboration. The first one doesn't involve text recognition. It aims to improve speech recognition by using a thematized language model based on textual resources of the course. Despite the small size of this corpus, we show a significant improvement in recognition results. We are experiencing also a text and speech cross-recognition method based on the complementarity phonetic and written representations of language. We show that this approach improves the text recognition results and could be used to emphasize the specialized vocabulary of the course
APA, Harvard, Vancouver, ISO, and other styles
23

Hekimoglu, M. Kadri. "Video-text processing by using Motorola 68020 CPU and its environment." Thesis, Monterey, California. Naval Postgraduate School, 1991. http://hdl.handle.net/10945/26833.

Full text
APA, Harvard, Vancouver, ISO, and other styles
24

Demirtas, Kezban. "Automatic Video Categorization And Summarization." Master's thesis, METU, 2009. http://etd.lib.metu.edu.tr/upload/3/12611113/index.pdf.

Full text
Abstract:
In this thesis, we make automatic video categorization and summarization by using subtitles of videos. We propose two methods for video categorization. The first method makes unsupervised categorization by applying natural language processing techniques on video subtitles and uses the WordNet lexical database and WordNet domains. The method starts with text preprocessing. Then a keyword extraction algorithm and a word sense disambiguation method are applied. The WordNet domains that correspond to the correct senses of keywords are extracted. Video is assigned a category label based on the extracted domains. The second method has the same steps for extracting WordNet domains of video but makes categorization by using a learning module. Experiments with documentary videos give promising results in discovering the correct categories of videos. Video summarization algorithms present condensed versions of a full length video by identifying the most significant parts of the video. We propose a video summarization method using the subtitles of videos and text summarization techniques. We identify significant sentences in the subtitles of a video by using text summarization techniques and then we compose a video summary by finding the video parts corresponding to these summary sentences.
APA, Harvard, Vancouver, ISO, and other styles
25

Saidane, Zohra. "Reconnaissance de texte dans les images et les vidéos en utilisant les réseaux de neurones à convolutions." Phd thesis, Télécom ParisTech, 2008. http://pastel.archives-ouvertes.fr/pastel-00004685.

Full text
APA, Harvard, Vancouver, ISO, and other styles
26

Tarczyńska, Anna. "Methods of Text Information Extraction in Digital Videos." Thesis, Blekinge Tekniska Högskola, Sektionen för datavetenskap och kommunikation, 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-2656.

Full text
Abstract:
Context The huge amount of existing digital video files needs to provide indexing to make it available for customers (easier searching). The indexing can be provided by text information extraction. In this thesis we have analysed and compared methods of text information extraction in digital videos. Furthermore, we have evaluated them in the new context proposed by us, namely usefulness in sports news indexing and information retrieval. Objectives The objectives of this thesis are as follows: providing a better understanding of the nature of text extraction; performing a systematic literature review on various methods of text information extraction in digital videos of TV sports news; designing and executing an experiment in the testing environment; evaluating available and promising methods of text information extraction from digital video files in the proposed context associated with video sports news indexing and retrieval; providing an adequate solution in the proposed context described above. Methods This thesis consists of three research methods: Systematic Literature Review, Video Content Analysis with the checklist, and Experiment. The Systematic Literature Review has been used to study the nature of text information extraction, to establish the methods and challenges, and to specify the effective way of conducting the experiment. The video content analysis has been used to establish the context for the experiment. Finally, the experiment has been conducted to answer the main research question: How useful are the methods of text information extraction for indexation of video sports news and information retrieval? Results Through the Systematic Literature Review we identified 29 challenges of the text information extraction methods, and 10 chains between them. We extracted 21 tools and 105 different methods, and analyzed the relations between them. Through Video Content Analysis we specified three groups of probability of text extraction from video, and 14 categories for providing video sports news indexation with the taxonomy hierarchy. We have conducted the Experiment on three videos files, with 127 frames, 8970 characters, and 1814 words, using the only available MoCA tool. As a result, we reported 10 errors and proposed recommendations for each of them. We evaluated the tool according to the categories mentioned above and offered four advantages, and nine disadvantages of the Tool mentioned above. Conclusions It is hard to compare the methods described in the literature, because the tools are not available for testing, and they are not compared with each other. Furthermore, the values of recall and precision measures highly depend on the quality of the text contained in the video. Therefore, performing the experiments on the same indexed database is necessary. However, the text information extraction is time consuming (because of huge amount of frames in video), and even high character recognition rate gives low word recognition rate. Therefore, the usefulness of text information extraction for video indexation is still low. Because most of the text information contained in the videos news is inserted in post-processing, the text extraction could be provided in the root: during the processing of the original video, by the broadcasting company (e.g. by automatically saving inserted text in separate file). Then the text information extraction will not be necessary for managing the new video files
The huge amount of existing digital video files needs to provide indexing to make it available for customers (easier searching). The indexing can be provided by text information extraction. In this thesis we have analysed and compared methods of text information extraction in digital videos. Furthermore, we have evaluated them in the new context proposed by us, namely usefulness in sports news indexing and information retrieval.
APA, Harvard, Vancouver, ISO, and other styles
27

Bartlett, Melissa Ellis. "High School Students Reading Informational Texts| A Comparison of Written and Video Response Modalities." Thesis, North Carolina State University, 2015. http://pqdtopen.proquest.com/#viewpdf?dispub=3690206.

Full text
APA, Harvard, Vancouver, ISO, and other styles
28

Hay, Richard. "Views and perceptions of the use of text and video in English teaching." Thesis, Högskolan i Gävle, Avdelningen för humaniora, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:hig:diva-25400.

Full text
Abstract:
This essay investigates how students studying at upper secondary level perceive the use of text and video as teaching aids as part of their English studies. Students that came from both vocational and preparatory programs completed an online survey and seven of them subsequently took part in interviews. Both the survey and the interviews show that a large majority of students would like much more video based teaching material as part of their English courses - some would even prefer all the course material to be video based. The results also show that even though the students want more video, opinion is divided when it comes to how much, and in what way video is best used or incorporated into English teaching. Many of the students that asked for more video said that they found it difficult to read and to understand longer texts, furthermore they found texts to be boring. They pointed out that video was more interesting and motivating. Video was generally seen as being the preferred choice when it came to authentic language, help with pronunciation and access to the culture of different English speaking countries. Text, on the other hand, was seen to provide a much richer and more detailed information which was especially helpful when it came to spelling and grammar. It was also clear that the preference for video was greater among the students from the vocational classes. There was also a general agreement that, although video is used as a teaching aid, it is more usually used by their teachers as a time filler or reward. Finally, even if learning English continues to be based on text and course books, there is a broad consensus among the students that more video should be used, as it is seen as a valuable and an effective complement to traditional text based material.
APA, Harvard, Vancouver, ISO, and other styles
29

Castro, Adriana Petito de Almeida Silva. "Desempenho termico de vidros utilizados na construção civil : estudo em celulas-teste." [s.n.], 2006. http://repositorio.unicamp.br/jspui/handle/REPOSIP/257744.

Full text
Abstract:
Orientador: Lucila Chebel Labaki
Tese (doutorado) - Universidade Estadual de Campinas, Faculdade de Engenharia Civil, Arquitetura e Urbanismo
Made available in DSpace on 2018-08-08T09:29:44Z (GMT). No. of bitstreams: 1 Castro_AdrianaPetitodeAlmeidaSilva_D.pdf: 2534532 bytes, checksum: 4ea70ab9e519a13a5ddf61c29646f7b7 (MD5) Previous issue date: 2006
Resumo: O desempenho térmico de uma edificação depende de fatores como implantação, orientação, materiais e componentes construtivos, que devem ser adequadamente definidos para diferentes condições climáticas. O edifício atua como mecanismo de controle das variáveis do clima, através de sua envoltória (paredes, piso, cobertura e aberturas) e dos elementos do entorno, e deve ser projetado de modo a proporcionar conforto e eficiência energética. Entre os componentes da construção, os vidros funcionam como um dos elementos de maior potencialidade de aquecimento interno. Devido à sua transparência à radiação solar, possibilitam facilmente o ingresso de grande parte dessa energia no ambiente. No presente rabalho estudou-se o comportamento de superfícies transparentes em fachadas, em situação real, através de medições em seis células-teste, de dimensões 2,00 x 2,50m. Analisou-se in loco o comportamento térmico de vidros, comercialmente disponíveis no mercado, cuja caracterização, do ponto de vista da transmissão espectral, já era conhecida através de técnica espectrofotométrica. Foram selecionados 14 tipos de vidros: cinco planos, quatro refletivos pirolíticos, três refletivos metalizados a vácuo e dois laminados, os quais foram instalados em aberturas de 1,00 X 1,20m nas células-teste, localizadas em fachadas com orientação norte e oeste. A análise foi realizada separadamente para as duas orientações. Avaliou-se o desempenho térmico de cada vidro, tendo o incolor, de espessura 4mm, como referência. A metodologia utilizada consistiu na aquisição de dados mensuráveis de parâmetros ambientais: temperaturas superficiais internas dos vidros, temperaturas de bulbo seco no ambiente interno e temperaturas do ar externo. O fator de ganho solar foi calculado a partir dos resultados de absortância dos diferentes materiais, obtidos por análise espectrofotométrica, e diferenças entre as temperaturas interna e externa. Os resultados mostram o alto ganho de calor através dos vidros planos, sendo o incolor o de pior desempenho térmico, com maior ganho, seguido pelo bronze, cinza e verde. Os vidros refletivos metalizados a vácuo apresentam o melhor desempenho térmico, dentre os materiais analisados, quando se tem como objetivo atenuar o ganho de calor e projetar edificações com menor consumo energético para refrigeração
Abstract: Thermal performance of buildings depends on several factors, such as implantation, orientation, materials and building components, which should be appropriately defined for different climate conditions. The building acts as a controller of the climatic variables, through the building envelope (walls, floor, roof and openings) and the nearby elements. Building design must provide indoor comfort and energy efficiency. Glazing easily allows the penetration of solar radiation into buildings, due to its transparency to solar radiation. So glasses must be carefully considered in building design, having in mind their potential for internal heating. In this work studies about the behavior of transparent façades, in real conditions, were carried out through measurements in six test-cells with dimensions 2,00 x 2,50m. Fourteen types of glasses were selected: five float glasses, four reflective glasses produced by pyrolitic process, three reflective ones obtained by vacuum metal deposition and two laminated glasses. Their spectral behavior was known from previous spectrophotometric studies. The glasses were installed in 1,00 X 1,20m openings in two façades facing north and west, separately. The colorless 4mm float glass was taken as reference. Internal surface temperatures of glasses, internal dry bulb temperatures and outdoor temperatures were collected. Solar gain factor was calculated on the basis of absortance values, obtained from spectrophotometrical nalysis, and temperature differences inside and outside the cells. Results show a high heat gain through float glasses, with the worst thermal behavior for the colorless one, followed by bronze, gray and green. Furthermore, reflective glasses obtained by vacuum metal deposition present the best thermal performance for the purpose of heat gain attenuation and to design buildings with the least energy consumption for cooling
Doutorado
Edificações
Doutor em Engenharia Civil
APA, Harvard, Vancouver, ISO, and other styles
30

Schwarz, Katharina [Verfasser], and Hendrik P. A. [Akademischer Betreuer] Lensch. "Text–to–Video : Image Semantics and NLP / Katharina Schwarz ; Betreuer: Hendrik P. A. Lensch." Tübingen : Universitätsbibliothek Tübingen, 2019. http://d-nb.info/1182985963/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Uggerud, Nils. "AnnotEasy: A gesture and speech-to-text based video annotation tool for note taking in pre-recorded lectures in higher education." Thesis, Linnéuniversitetet, Institutionen för datavetenskap och medieteknik (DM), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-105962.

Full text
Abstract:
This paper investigates students’ attitudes towards using gestures and speech-to- text (GaST) to take notes while watching recorded lectures. A literature review regarding video based learning, an expert interview, and a background survey regarding students’ note taking habits led to the creation of the prototype AnnotEasy, a tool that allows students to use GaST to take notes. AnnotEasy was tested in three iterations against 18 students, and was updated after each iteration.  The students watched a five minute long lecture and took notes by using AnnotEasy. The participants’ perceived ease of use (PEU) and perceived usefulness (PU) was evaluated based on the TAM. Their general attitudes were evaluated in semi structured interviews.  The result showed that the students had a high PEU and PU of AnnotEasy. They were mainly positive towards taking notes by using GaST. Further, the result suggests that AnnotEasy could facilitate the process of structuring a lecture’s content. Lastly, even though students had positive attitudes towards using speech to create notes, observations showed that this was problematic when the users attempted to create longer notes. This indicates that speech could be more beneficial for taking shorter notes.
APA, Harvard, Vancouver, ISO, and other styles
32

Memmi, Paul Joseph. "Etude sémiolinguistique du sous-titrage pour une écriture concise assistée par ordinateur (ECAO) avec application à l'audiovisuel." Paris 10, 2005. http://www.theses.fr/2005PA100069.

Full text
Abstract:
Intelligentiæ pauca – À l'intelligence peu (suffit). Cet adage qu'aimait Stendhal montre par sa forme elliptique, par le plaisir qu'il suscite et l'esprit qu'il requiert ce qu'est l'écriture concise. Cette thèse vise à concevoir un logiciel de traitement de texte, l'ÉCAO (Écriture Concise Assistée par Ordinateur) qui, appliqué à l'audiovisuel, servirait pour Internet, la traduction sous-titrée et le sous-titrage unilingue à l'usage des malentendants. L'étude sémiolinguistique du sous-titrage, cas d'écriture concise en environnement verbal et audiovisuel, aboutit à proposer une méthode de référencement et de désambiguïsation des informations source et des opérateurs de concision phrastique. Certains sont programmables, d'autres révèlent les carences de l'automate face à des constructions de sens pourtant capitales. Par là, se révèle le propos essentiel de cette recherche : l'étude de l'intégration cognitive des communications complexes et de la concision comme mode de représentation
Intelligentiæ pauca – To intelligence, little (is enough). Through its elliptic form, the pleasure it arouses and the wit it calls for, this phrase praised by Stendhal points out what concise writing is. This thesis aims at conceiving a word processor ÉCAO (French for Automatically Processed Concise Writing – APCW) which, in its audiovisual application, should find uses also for Internet, subtitled translations and subtitling for the hearing-impaired. A semiolinguistic study of the subtitling, an example of concise writing in a verbal and audiovisual environment, leads to coming up with a method for referencing and disambiguating the source information and with a set of phrastic concision operators. Some are programmable, others reveal the automaton's deficiencies faced with sense constructions which are yet of capital importance. There lies the essential purpose of this research: the study of cognitive integration of complex communications and of concision as a mode of representation
APA, Harvard, Vancouver, ISO, and other styles
33

Ulvbäck, Gustav, and Wingårdh Rickard Eriksson. "Förmedla information med animerad text : Blir textbaserad information på sociala medier mer intressant om det sker i rörlig bild med animerad text?" Thesis, Södertörns högskola, Institutionen för naturvetenskap, miljö och teknik, 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:sh:diva-34509.

Full text
Abstract:
Syftet med studien är att undersöka om textbaserad information på sociala medier blir mer intressant om det sker i rörlig bild med animerad text. Studien bygger på en målgrupp bestående av unga Facebookanvändare där användare och deras digitala konsumtion på den sociala plattformen undersöks. Empirisk datainsamling har genomförts i form av enkäter som digitalt distribuerats till den utvalda målgruppen. Djupgående, kvalitativa och kompletterande intervjuer med slumpmässigt utvalda respondenter har genomförts för att säkerställa en kvalitativ ansats i studien. Studiens resultat visar på en jämn fördelning för föredragen form av nyhetsrapportering i form av rörlig bild och bild med kompletterande text där det tydligt framkommer spetsegenskaper för de olika alternativen. Undersökningen visar att de olika alternativen för informationsförmedling har tydliga kopplingar till olika motiveringar. Tillgänglighet och intressant är synonymt med rörlig bild medans informativt och tydligt är synonymt med bild med kompletterande text. Studien visar även på att det är ett område som kräver vidare forskning.
APA, Harvard, Vancouver, ISO, and other styles
34

Bachman, Kathryn M. "Using Videos versus Traditional Written Texts in the Classroom to Enhance Student Learning." Ohio Dominican University Honors Theses / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=oduhonors1449441013.

Full text
APA, Harvard, Vancouver, ISO, and other styles
35

Jaroňová, Eva. "Od ideálu k utopii (zítřek, co už byl)." Master's thesis, Vysoké učení technické v Brně. Fakulta výtvarných umění, 2012. http://www.nusl.cz/ntk/nusl-232359.

Full text
Abstract:
My work contains five videos. Installation presents videos on five separate tv screens. Each video duration is max 3 min. Videos are based on text. Text exists as installation in space and continues in form of performance. Content of text is simple wisdom which is twist off into nonsense. Realisation contains element of ephemerality. Former work is destroed by nature or human incidence.
APA, Harvard, Vancouver, ISO, and other styles
36

Wells, Emily Jean. "The effects of luminance contrast, raster modulation, and ambient illumination on text readability and subjective image quality." Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-07102009-040235/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Stokes, Charlotte Ellenor. "Investigating the Efficacy of Video versus Text Instruction for the Recall of Food Safety Information." Digital Archive @ GSU, 2009. http://digitalarchive.gsu.edu/nutrition_theses/28.

Full text
Abstract:
Purpose: Teaching consumers proper home food safety practices is an important strategy to combat foodborne illness. Food safety educators with limited resources must do a cost-versus-benefit analysis before choosing the optimum medium to reach their target audiences. The objectives of this research were to determine whether presenting food safety information in a video format was more effective than text-only in terms of audience recall of the information one week later; to determine whether an intervention in text or video form increased students’ knowledge of food safety information as compared to no intervention at all; and to identify certain demographic factors that could have influenced performance on a food safety quiz. Methods: One hundred thirty-three Georgia State University undergraduate students were assigned to one of three groups. One group viewed a food safety video (n=59), a second group received the same information in text-only form (n=41), and the third group (n=33) served as the control and received no intervention. Students filled out a demographic questionnaire and completed a pre-intervention and post-intervention food safety knowledge test. Average scores were calculated, and the data were analyzed using SPSS 16.0 for Windows. Results: There was no significant difference between pre-intervention test scores among the three groups (p=.057). The video group scored significantly higher on the post-intervention test (p=.006) than the text group and the control group (p<.001). The video group (p<.001) and text group (p<.001) both scored significantly higher on the post-intervention quiz than the pre-intervention quiz, but the control group did not (p=.466). Video was superior to text overall and in conveying basic food safety principles; however, students in the text group demonstrated a better recall of more detailed food safety information such as proper internal cooking temperatures for poultry and ground beef. Previous food safety education in the classroom or online was found to be the only significant predictor of better performance on the pre-intervention test (p=.004). Conclusion: Video is more effective than text when used to deliver simple, direct food safety messages. More detailed information, such as proper internal cooking temperatures, might be best delivered in text form. Consumers are likely to benefit most from a multimedia approach to food safety education that includes videos, accompanying brochures, and Web site content.
APA, Harvard, Vancouver, ISO, and other styles
38

Tran, Anh Xuan. "Identifying latent attributes from video scenes using knowledge acquired from large collections of text documents." Thesis, The University of Arizona, 2014. http://pqdtopen.proquest.com/#viewpdf?dispub=3634275.

Full text
Abstract:

Peter Drucker, a well-known influential writer and philosopher in the field of management theory and practice, once claimed that “the most important thing in communication is hearing what isn't said.” It is not difficult to see that a similar concept also holds in the context of video scene understanding. In almost every non-trivial video scene, most important elements, such as the motives and intentions of the actors, can never be seen or directly observed, yet the identification of these latent attributes is crucial to our full understanding of the scene. That is to say, latent attributes matter.

In this work, we explore the task of identifying latent attributes in video scenes, focusing on the mental states of participant actors. We propose a novel approach to the problem based on the use of large text collections as background knowledge and minimal information about the videos, such as activity and actor types, as query context. We formalize the task and a measure of merit that accounts for the semantic relatedness of mental state terms, as well as their distribution weights. We develop and test several largely unsupervised information extraction models that identify the mental state labels of human participants in video scenes given some contextual information about the scenes. We show that these models produce complementary information and their combination significantly outperforms the individual models, and improves performance over several baseline methods on two different datasets. We present an extensive analysis of our models and close with a discussion of our findings, along with a roadmap for future research.

APA, Harvard, Vancouver, ISO, and other styles
39

Macindoe, Annie C. "Melancholy and the memorial: Representing loss, grief and affect in contemporary visual art." Thesis, Queensland University of Technology, 2018. https://eprints.qut.edu.au/119695/1/Annie_Macindoe_Thesis.pdf.

Full text
Abstract:
Melancholy and the Memorial: Representing Loss, Grief and Affect in Contemporary Visual Art is a practice-led project that explores how contemporary art can respond to the limitations of traditional forms of language in the representation of trauma, loss and grief. The project reflects on the work of theorists and artists who also explore the ineffability of these memories and experiences. The creative outcomes have investigated how text, moving image, sound and space can be combined to reframe the dialogue around public and private expressions of trauma and open up discussion of the potential for shared, affectual experiences through art.
APA, Harvard, Vancouver, ISO, and other styles
40

Wolf, Christian Jolion Jean-Michel. "Détection de textes dans des images issues d'un flux vidéo pour l'indexation sémantique." Villeurbanne : Doc'INSA, 2005. http://docinsa.insa-lyon.fr/these/pont.php?id=wolf.

Full text
Abstract:
Thèse doctorat : Informatique : Villeurbanne, INSA : 2003.
Thèse rédigée en anglais. Introduction et conclusion générale en français. En 2ème partie, choix d'articles en français avec résumés, mots-clef et réf. bibliogr. Titre provenant de l'écran-titre. Bibliogr. p. 147-154. Publications de l'auteur p. 155-157.
APA, Harvard, Vancouver, ISO, and other styles
41

Бикова, О. Д. "Відеовербальний текст німецькомовного вербального дискурсу." Thesis, Сумський державний університет, 2013. http://essuir.sumdu.edu.ua/handle/123456789/30524.

Full text
Abstract:
Дослідивши питання відеовербального тексту німецькомовного рекламного дискурсу, можна зробити висновки, що реклама є багатогранним поняттям і єдиного визначення цього явища не існує. При цитуванні документа, використовуйте посилання http://essuir.sumdu.edu.ua/handle/123456789/30524
APA, Harvard, Vancouver, ISO, and other styles
42

Ryrå, Landgren Isabella. "Samspel i det berättartekniska : text, bild och effekter i musikvideor." Thesis, Högskolan Väst, Avd för medier och design, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:hv:diva-8965.

Full text
Abstract:
Music videos have under the last 50 years become entertainment in our society. Some are created to reflect feelings while others are a kind of showcase for the artist. There are also those based off of the lyrics and thus creates a short film or an illustration of the lyrics. Through the use of technologies such as visual effects is it possible to bring the impossible worlds and stories alive.Videos with these effects are the kind of videos I've been analyzing in this essay with the purposeto explore how much the visual effects affect the narration. To achieve this I have chosen to make a semiotic study focused on analysis and interpretations of five chosen music videos created during orafter the year 2000. CGI, slow-motion and metaphors are techniques I've been looking at and they have proved to contribute to how the story of the video is told and how it's understood. The interplay between image and text is another thing I've been studying and in the chosen videos it's been varying between interpretations and literal translation to the other.
Musikvideor har under de senaste 50 åren varit en form av underhållning för vårt samhälle. Somliga formas för att spegla känslor medan andra visar upp artisten. Det finns de som baserar sig på låttexten för att skapa en kortare film eller gestalta låttextens innehåll. Med hjälp av tekniker som visuella effekter kan dessa drömlika och omöjliga världar och historier komma till liv. Det är videor med sådana effekter jag valt att analysera i denna uppsats med syftet att ta reda påhur stor roll de visuella effekterna spelar i berättandet. För att komma fram till detta har jag gjort en semiotisk studie fokuserad på analys och tolkningar av fem valda videor skapade under eller efter 2000-talet. CGI, slow-motion och metaforer är tekniker jag kollat på och det har visat sig att de alla bidrar till hur berättandet utspelas och uppfattas. Sambandet mellan bild och text i de valda videorna har pendlat mellan tolkning till bokstavligt översatt till varandra.
APA, Harvard, Vancouver, ISO, and other styles
43

Bayar, Mujdat. "Event Boundary Detection Using Web-cating Texts And Audio-visual Features." Master's thesis, METU, 2011. http://etd.lib.metu.edu.tr/upload/12613755/index.pdf.

Full text
Abstract:
We propose a method to detect events and event boundaries in soccer videos by using web-casting texts and audio-visual features. The events and their inaccurate time information given in web-casting texts need to be aligned with the visual content of the video. Most match reports presented by popular organizations such as uefa.com (the official site of Union of European Football Associations) provide the time information in minutes rather than seconds. We propose a robust method which is able to handle uncertainties in the time points of the events. As a result of our experiments, we claim that our method detects event boundaries satisfactorily for uncertain web-casting texts, and that the use of audio-visual features improves the performance of event boundary detection.
APA, Harvard, Vancouver, ISO, and other styles
44

Saracoglu, Ahmet. "Localization And Recognition Of Text In Digital Media." Master's thesis, METU, 2007. http://etd.lib.metu.edu.tr/upload/2/12609028/index.pdf.

Full text
Abstract:
Textual information within digital media can be used in many areas such as, indexing and structuring of media databases, in the aid of visually impaired, translation of foreign signs and many more. This said, mainly text can be separated into two categories in digital media as, overlay-text and scene-text. In this thesis localization and recognition of video text regardless of its category in digital media is investigated. As a necessary first step, framework of a complete system is discussed. Next, a comparative analysis of feature vector and classification method pairs is presented. Furthermore, multi-part nature of text is exploited by proposing a novel Markov Random Field approach for the classification of text/non-text regions. Additionally, better localization of text is achieved by introducing bounding-box extraction method. And for the recognition of text regions, a handprint based Optical Character Recognition system is thoroughly investigated. During the investigation of text recognition, multi-hypothesis approach for the segmentation of background is proposed by incorporating k-Means clustering. Furthermore, a novel dictionary-based ranking mechanism is proposed for recognition spelling correction. And overall system is simulated on a challenging data set. Also, a through survey on scene-text localization and recognition is presented. Furthermore, challenges are identified and discussed by providing related work on them. Scene-text localization simulations on a public competition data set are also provided. Lastly, in order to improve recognition performance of scene-text on signs that are affected from perspective projection distortion, a rectification method is proposed and simulated.
APA, Harvard, Vancouver, ISO, and other styles
45

Srsen, Kenney Kristen Laura. "CRITICAL VIDEO PROJECTS: UNDERSTANDING NINE STUDENTS’ EXPERIENCES WITH CRITICAL LITERACY AS THEY RE-IMAGINE CANONICAL TEXTS THROUGH FILMS." Kent State University / OhioLINK, 2019. http://rave.ohiolink.edu/etdc/view?acc_num=kent1572546051237628.

Full text
APA, Harvard, Vancouver, ISO, and other styles
46

Ramírez, Díaz José Fernando. "Formación de imagen completa de una página con texto impreso mediante procesamiento de imágenes obtenidas de un video." Bachelor's thesis, Pontificia Universidad Católica del Perú, 2020. http://hdl.handle.net/20.500.12404/17644.

Full text
Abstract:
En la presente tesis se aborda el diseño e implementación de un algoritmo que permite formar la imagen completa de un documento con texto impreso partiendo de un video que contiene fragmentos de la página en cuestión. Dicho algoritmo recibe como entrada un video registrado empleando la cámara de un teléfono móvil y como resultado retornará la imagen del documento con texto completo; esta imagen puede ser empleada posteriormente en un algoritmo de reconocimiento óptico de caracteres (u OCR por sus siglas en inglés) para recuperar el texto en forma digital. El enfoque del desarrollo de esta propuesta es el de brindar una solución alternativa, en cuanto a adquisición de imágenes, para las existentes aplicaciones móviles de OCR enfocadas en apoyar a personas con ceguera parcial o total. Para abarcar el planteamiento y cumplimiento de los objetivos de este proyecto, se ha estructurado el mismo en 4 capítulos. En el capítulo 1 se aborda la actual situación de personas con distintos grados de discapacidad visual en nuestro país y diversos sistemas que buscan apoyarlos en recuperar su autonomía informativa y educativa. Además, se trata detalles sobre el estado del arte en adquisición de imágenes para las aplicaciones OCR existentes en la actualidad y sus falencias. En el capítulo 2 se presenta el marco teórico que avala el desarrollo del algoritmo propuesto, desde la teoría necesaria en procesamiento de imágenes y, también, sobre el registro de vídeos. En el capítulo 3 se trata el diseño e implementación del algoritmo en dos plataformas: inicialmente en Python 3.6 para la etapa de calibración de parámetros en una computadora de escritorio, y en C++ para las pruebas finales en un teléfono con SO Android. En dicho capítulo también se hace presente consideraciones planteadas para la creación del conjunto de videos de pruebas en Python. Finalmente, en el capítulo 4 se exponen las pruebas y resultados obtenidos de la aplicación del algoritmo, en Python, sobre la base de muestras creadas, y los resultados finales del uso de la aplicación en Android. Para estimar el grado de conformidad de la imagen resultante se hará uso de la métrica de Levenshtein o distancia de edición, la cual señala cuántos caracteres detectados en la imagen compuesta son diferentes a los caracteres del texto original.
Tesis
APA, Harvard, Vancouver, ISO, and other styles
47

Miana, Anna Christina. "Avaliação do desempenho térmico de brises transparentes: ensaio em células-teste." Universidade de São Paulo, 2005. http://www.teses.usp.br/teses/disponiveis/18/18141/tde-06032006-120003/.

Full text
Abstract:
O objetivo desta pesquisa foi avaliar o desempenho térmico de brises transparentes, por meio de medidas em células-teste construídas em escala 1:1, na cidade de Campinas, São Paulo. Trata-se de um estudo que faz avaliação comparativa das temperaturas internas e superficiais dos brises e do vidro da janela em seis células-teste, quatro delas protegidas com brises transparentes horizontais e verticais combinados, constituídos por vidros, outra protegida com brise-soleil metálico horizontal e uma sem dispositivo de proteção solar, sendo esta considerada de referência. Foram selecionados quatro tipos de vidro com diferentes características ópticas: vidro float comum incolor, vidro impresso mini-boreal, vidro float comum azul e vidro refletivo metalizado a vácuo prata médio. Foi possível concluir que os brises de vidro refletivo prata, float azul e impresso possuem um desempenho térmico similar entre si e ao brise metálico estudado, e que o brise de vidro float incolor não apresenta um resultado de desempenho térmico satisfatório. Com relação aos procedimentos de medidas de campo, foram verificadas as modificações necessárias, para trabalhos futuros que tenham o mesmo objetivo dessa pesquisa, ou similares a esse. Além da avaliação do desempenho térmico, foi também iniciada uma avaliação do desempenho luminoso dos brises. Por meio das medidas dos níveis de iluminação natural dentro das células-teste constatou-se que o brise de vidro impresso apresentou um ótimo desempenho luminoso e que o brise de vidro refletivo prata reduziu significativamente a iluminância dentro da célula-teste, obtendo um desempenho luminoso inferior até mesmo ao brise metálico
This research intended to evaluate thermal performance of transparent solar protections using measurements in full scale test cells, located in the city of Campinas, São Paulo, Brazil. Surface (window and shading device) and internal air temperatures were compared for six test cells. One of them was unprotected, for reference purposes, another was obstructed with a metallic shading device and the other four had transparent glass horizontal and vertical shades installed. Four different types of glass with different optical properties were selected: float clear glass, mini-boreal printed glass, float blue glass and metallic silver reflective glass. The results of the thermal appraisal showed that silver reflective glass, float blue and printed glass shadings achieved similar performance, not very different to the ones obtained for the metallic shading device. Therefore the float clear glass shading did not attain a satisfactory result. The field measurements procedures were also evaluated and test cells characteristics problems were identified, in order to suggest changes for future research in this area. This research began to evaluate light performance of the same solar protections. For light performance evaluation were measured the daylight in the center of each test cell and outside. It was concluded that printed glass shadings presented very good results and silver reflective glass reduced the daylight inside the test cell
APA, Harvard, Vancouver, ISO, and other styles
48

Nguyen, Chu Duc. "Localization and quality enhancement for automatic recognition of vehicle license plates in video sequences." Thesis, Ecully, Ecole centrale de Lyon, 2011. http://www.theses.fr/2011ECDL0018.

Full text
Abstract:
La lecture automatique de plaques d’immatriculation de véhicule est considérée comme une approche de surveillance de masse. Elle permet, grâce à la détection /localisation ainsi que la reconnaissance optique, d’identifier un véhicule dans les images ou les séquences d’images. De nombreuses applications comme le suivi du trafic, la détection de véhicules volés, le télépéage ou la gestion d’entrée / sortie des parkings utilise ce procédé. Or malgré d’important progrès enregistré depuis l’apparition des premiers prototypes en 1979 accompagné d’un taux de reconnaissance parfois impressionnant, notamment grâce aux avancés en recherche scientifique et en technologie des capteurs, les contraintes imposés pour le bon fonctionnement de tels systèmes en limitent les portées. En effet, l’utilisation optimale des techniques de localisation et de reconnaissance de plaque d’immatriculation dans les scénarii opérationnels nécessite des conditions d’éclairage contrôlées ainsi qu’une limitation dans de la pose, de vitesse ou tout simplement de type de plaque. La lecture automatique de plaques d’immatriculation reste alors un problème de recherche ouvert. La contribution majeure de cette thèse est triple. D’abord une nouvelle approche robuste de localisation de plaque d’immatriculation dans des images ou des séquences d’images est proposée. Puis, l’amélioration de la qualité des plaques localisées est traitée par une adaptation de technique de super-résolution. Finalement, un modèle unifié de localisation et de super-résolution est proposé permettant de diminuer la complexité temporelle des deux approches combinées
Automatic reading of vehicle license plates is considered an approach to mass surveillance. It allows, through the detection / localization and optical recognition to identify a vehicle in the images or video sequences. Many applications such as traffic monitoring, detection of stolen vehicles, the toll or the management of entrance/ exit parking uses this method. Yet in spite of important progress made since the appearance of the first prototype sin 1979, with a recognition rate sometimes impressive thanks to advanced science and sensor technology, the constraints imposed for the operation of such systems limit laid. Indeed, the optimal use of techniques for localizing and recognizing license plates in operational scenarios requiring controlled lighting conditions and a limitation of the pose, velocity, or simply type plate. Automatic reading of vehicle license plates then remains an open research problem. The major contribution of this thesis is threefold. First, a new approach to robust license plate localization in images or image sequences is proposed. Then, improving the quality of the plates is treated with a localized adaptation of super-resolution technique. Finally, a unified model of location and super-resolution is proposed to reduce the time complexity of both approaches combined
APA, Harvard, Vancouver, ISO, and other styles
49

Escobar, Mayte. "The Body As Border: El Cuerpo Como Frontera." CSUSB ScholarWorks, 2015. https://scholarworks.lib.csusb.edu/etd/247.

Full text
Abstract:
Being First generation born Mexican American I am looking into the blend of the two cultures and the disparity between them. The border is the core of my investigation; by traveling across the border I have become conscious of the differences among both sides and duality within myself. My identity has developed from a synthesis of these two cultures, and my wok explores these two factions that cannot be one without the other. fusion is apparent in my self-portraits where I dress up with the colors from both sides of the border. But I also take a personal look into understanding the history and identity of each nation. I create a juxtaposition with these two identities that become one and explore the social, cultural, and political issues we face in the everyday. I recreate my “investigation,” by trying to dig deeper, exposing the layers, and facing my own identity crisis in the process.
APA, Harvard, Vancouver, ISO, and other styles
50

Hansen, Simon. "TEXTILE - Augmenting Text in Virtual Space." Thesis, Malmö högskola, Fakulteten för kultur och samhälle (KS), 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-23172.

Full text
Abstract:
Three-dimensional literature is a virtually non-existent or in any case very rare and emergent digital art form, defined by the author as a unit of text, which is not confined to the two-dimensional layout of print literature, but instead mediated across all three axes of a virtual space. In collaboration with two artists the author explores through a bodystorming workshop how writers and readers could create and experience three-dimensional literature in mixed reality, by using mobile devices that are equipped with motion sensors, which enable users to perform embodied interactions as an integral part of the literary experience.For documenting the workshop, the author used body-mounted action cameras in order to record the point-of-view of the participants. This choice turned out to generate promising knowledge on using point-of-view footage as an integral part of the methodological approach. The author has found that by engaging creatively with such footage, the designer gains a profound understanding and vivid memory of complex design activities.As the outcome the various design activities, the author developed a concept for an app called TEXTILE. It enables users to build three-dimensional texts by positioning words in a virtual bubble of space around the user and to share them, either on an online platform or at site-specific places. A key finding of this thesis is that the creation of three-dimensional literature on a platform such as TEXTILE is not just an act of writing – it is an act of sculpture and an act of social performance.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography