Dissertations / Theses on the topic 'Video annotation'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'Video annotation.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Chaudhary, Ahmed. "Video annotation tools." Thesis, Texas A&M University, 2008. http://hdl.handle.net/1969.1/85911.
Full textHartley, Edward. "Automating video annotation." Thesis, Lancaster University, 2004. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.435884.
Full textClement, Michael David. "Obstacle Annotation by Demonstration." Diss., CLICK HERE for online access, 2007. http://contentdm.lib.byu.edu/ETD/image/etd1722.pdf.
Full textMahmood, Muhammad Habib. "Motion annotation in complex video datasets." Doctoral thesis, Universitat de Girona, 2018. http://hdl.handle.net/10803/667583.
Full textLa segmentació del moviment es refereix al procés de separar regions i trajectòries d'una seqüència de vídeo en subconjunts coherents d'espai i de temps. En aquesta tesi hem creat un nou i multifacètic dataset amb seqüències de la vida real que inclou diferent número de moviments i fotogrames per seqüència i distorsions amb dades incomplertes. A més, inclou ground-truth en tots els fotogrames basat en mesures de trajectòria i regió. Hem proposat també una nova eina semiautomàtica per delinear les trajectòries en vídeos complexos, fins i tot en vídeos capturats amb càmeres mòbils. Amb una mínima anotació manual dels objectes, l'algoritme és capaç de propagar-la en tots els fotogrames. Durant les oclusions, la correcció de les etiquetes es realitza aplicant el seguiment de la màscara per a cada ordre de profunditat. Els resultats obtinguts mostren que el nostre enfocament ofereix resultats reeixits en una àmplia varietat de seqüències de vídeo.
Aydinlilar, Merve. "Semi-automatic Semantic Video Annotation Tool." Master's thesis, METU, 2011. http://etd.lib.metu.edu.tr/upload/12613966/index.pdf.
Full textFoley-Fisher, Zoltan. "A pursuit method for video annotation." Thesis, University of British Columbia, 2012. http://hdl.handle.net/2429/43613.
Full textSalway, Andrew. "Video annotation : the role of specialist text." Thesis, University of Surrey, 1999. http://epubs.surrey.ac.uk/843350/.
Full textSilva, João Miguel Ferreira da. "People and object tracking for video annotation." Master's thesis, Faculdade de Ciências e Tecnologia, 2012. http://hdl.handle.net/10362/8953.
Full textObject tracking is a thoroughly researched problem, with a body of associated literature dating at least as far back as the late 1970s. However, and despite the development of some satisfactory real-time trackers, it has not yet seen widespread use. This is not due to a lack of applications for the technology, since several interesting ones exist. In this document, it is postulated that this status quo is due, at least in part, to a lack of easy to use software libraries supporting object tracking. An overview of the problems associated with object tracking is presented and the process of developing one such library is documented. This discussion includes how to overcome problems like heterogeneities in object representations and requirements for training or initial object position hints. Video annotation is the process of associating data with a video’s content. Associating data with a video has numerous applications, ranging from making large video archives or long videos searchable, to enabling discussion about and augmentation of the video’s content. Object tracking is presented as a valid approach to both automatic and manual video annotation, and the integration of the developed object tracking library into an existing video annotator, running on a tablet computer, is described. The challenges involved in designing an interface to support the association of video annotations with tracked objects in real-time are also discussed. In particular, we discuss our interaction approaches to handle moving object selection on live video, which we have called “Hold and Overlay” and “Hold and Speed Up”. In addition, the results of a set of preliminary tests are reported.
project “TKB – A Transmedia Knowledge Base for contemporary dance” (PTDC/EA /AVP/098220/2008 funded by FCT/MCTES), the UTAustin – Portugal, Digital Media Program (SFRH/BD/42662/2007 FCT/MCTES) and by CITI/DI/FCT/UNL (Pest-OE/EEI/UI0527/2011)
Dye, Brigham R. "Reliability of Pre-Service Teachers Coding of Teaching Videos Using Video-Annotation Tools." BYU ScholarsArchive, 2007. https://scholarsarchive.byu.edu/etd/990.
Full textWang, Yang. "Digital video segmentation and annotation in news programs." Hong Kong : University of Hong Kong, 2001. http://sunzi.lib.hku.hk/hkuto/record.jsp?B23273082.
Full textGoldman, Daniel R. "A framework for video annotation, visualization, and interaction /." Thesis, Connect to this title online; UW restricted, 2007. http://hdl.handle.net/1773/6994.
Full textDemirdizen, Goncagul. "An Ontology-driven Video Annotation And Retrieval System." Master's thesis, METU, 2010. http://etd.lib.metu.edu.tr/upload/12612592/index.pdf.
Full text("
, "
)"
, "
AND"
and "
OR"
operators. For all these query types, the system supports both general and video specific query processing. By this means, the user is able to pose queries on all videos in the video databases as well as the details of a specific video of interest.
Adam, Jameel. "Video annotation wiki for South African sign language." Thesis, University of the Western Cape, 2011. http://etd.uwc.ac.za/index.php?module=etd&action=viewtitle&id=gen8Srv25Nme4_1540_1304499135.
Full textThe SASL project at the University of the Western Cape aims at developing a fully automated translation system between English and South African Sign Language (SASL). Three important aspects of this system require SASL documentation and knowledge. These are: recognition of SASL from a video sequence, linguistic translation between SASL and English and the rendering of SASL. Unfortunately, SASL documentation is a scarce resource and no official or complete documentation exists. This research focuses on creating an online collaborative video annotation knowledge management system for SASL where various members of the community can upload SASL videos to and annotate them in any of the sign language notation systems, SignWriting, HamNoSys and/or Stokoe. As such, knowledge about SASL structure is pooled into a central and freely accessible knowledge base that can be used as required. The usability and performance of the system were evaluated. The usability of the system was graded by users on a rating scale from one to five for a specific set of tasks. The system was found to have an overall usability of 3.1, slightly better than average. The performance evaluation included load and stress tests which measured the system response time for a number of users for a specific set of tasks. It was found that the system is stable and can scale up to cater for an increasing user base by improving the underlying hardware.
Bodin, Erik. "Video content annotation automation using machine learning techniques." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-155751.
Full textAnnotering av semantiskt innehall av videomaterial ar kritiskt for eektiv sokning inomsadant material, vilket i sin tur mojliggor t.ex. att forfragningar till sokmotorer kanreturerna endast relevanta segment av videoklipp inom stora videohanteringssystem.Manuell annotering ar dock en trakig och tidsodande uppgift, vilket medfor lag kvalite ochliten mangd av sadana annoteringar. I denna uppsats foreslas ett system for attautomatisera det mesta av den processen. Systemet lar sig fran manuellt annoteratvideomaterial att inferera annoteringar for nytt material automatiskt, utan att kravaandringar pa systemet mellan olika anvandarskapta koncept att annotera. Prototypen somar presenterad i denna uppsats och utvarderad pa ett antal koncept visar lovande resultatfor koncept som har hogt inytande pa scenmiljoerna.
Lindström, Lucas. "Towards a Video Annotation System using Face Recognition." Thesis, Umeå universitet, Institutionen för datavetenskap, 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-85251.
Full textWang, Yang, and 王揚. "Digital video segmentation and annotation in news programs." Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 2001. http://hub.hku.hk/bib/B31225305.
Full textMarkatopoulou, Foteini. "Machine learning architectures for video annotation and retrieval." Thesis, Queen Mary, University of London, 2018. http://qmro.qmul.ac.uk/xmlui/handle/123456789/44693.
Full textLamia, Louis M. "Video annotation for choreographers on the NB platform." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/105994.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (page 71).
The NB Platform, originally developed to host discussion on class materials, has been adapted to suit the needs of dance choreographers. A Video Annotator was developed, and features requested by choreographers were added to help suit the platform to their needs. The updated annotator was then subjected to user testing to determine more ways in which software can be developed for the dance community and video annotation software can be improved.
by Louis M. Lamia.
M. Eng.
Li, Honglin. "Hierarchical video semantic annotation the vision and techniques /." Connect to this title online, 2003. http://rave.ohiolink.edu/etdc/view?acc%5Fnum=osu1071863899.
Full textTitle from first page of PDF file. Document formatted into pages; contains xv, 146 p.; also includes graphics. Includes bibliographical references (p. 136-146).
VON, WITTING DANIEL. "Annotation and indexing of video content basedon sentiment analysis." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-156387.
Full textTack vare tekniska framsteg inom mobilitet och tillgänglighet, kan media såsom film distribueras till flertalet olika plattformar, i form avströmning eller liknande tjänster. Det enorma utbudet av TV-serier och film utgör svårigheter för hur materialet ska lagras, sorteras och katalogiseras. Ofta är det dessutom användarna som ställer krav på vad somär relevant i en sökning. Det påvisar vikten av lämplig notation och indexering.I dag används oftast text som beskrivning av videoinnehållet, i form av antingen genre eller nyckelord. Det här arbetet är ett försök till att automatiskt kunna indexera film och serier, beroende på det semantiska innehållet. Att istället beskriva videomaterialet beroende på hur det uppfattas, samt de känslor som väcks, innebär en mer karaktäristisk skildring. Ett sådant signalement skulle beskriva det faktiska innehållet på ett annat sätt, som är mer lämpligt för jämförelser mellan två videoproduktioner. Eftersom skapandet av film anpassar sig till hur människor uppfattar videomaterial, kommer denna undersökning utnyttja de regler och praxis som används, som hjälp för maskininlärningen. Hur en film uppfattas, eller de känslor som framkallas, utgör en bas för inlärningen, då de används för att beteckna de olika koncept som ska klassificeras. En video representeras som en sekvens av klipp, med avsikt att fånga de tidsmässiga egenskaperna. Metoden som används för denna övervakade inlärning är en SVM som kan hantera data i form av strängar. Förutom de teknikaliteter som krävs för att förstå inlärningen,tar rapporten upp relevanta andra områden, t.ex. hur information ska extraheras och videosegmentering. Resultaten visar att det finns mönster i video, lämpliga för inlärning. På grund av för lite data, är det inte möjligt att avgöra hur metoden presterar. Det vore därför intressant med vidare analys, med mer data samt smärre modifikationer.
Diakopoulos, Nicholas A. "Collaborative annotation, analysis, and presentation interfaces for digital video." Diss., Atlanta, Ga. : Georgia Institute of Technology, 2009. http://hdl.handle.net/1853/29680.
Full textCommittee Chair: Essa, Irfan; Committee Member: Abowd, Gregory; Committee Member: Bolter, Jay; Committee Member: Lampe, Cliff; Committee Member: Stasko, John. Part of the SMARTech Electronic Thesis and Dissertation Collection.
Shao, Wenbin. "Automatic annotation of digital photos." Access electronically, 2007. http://www.library.uow.edu.au/adt-NWU/public/adt-NWU20080403.120857/index.html.
Full textSchroeter, Ronald. "Collaborative video indexing, annotation and discussion over high-bandwidth networks /." St. Lucia, Qld, 2004. http://www.library.uq.edu.au/pdfserve.php?image=thesisabs/absthe18014.pdf.
Full textAl-Athel, Mohammed S. "The role of terminology and local grammar in video annotation." Thesis, University of Surrey, 2008. http://epubs.surrey.ac.uk/1018/.
Full textYan, Fei. "Tennis ball tracking for automatic annotation of broadcast tennis video." Thesis, University of Surrey, 2007. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.441906.
Full textSkowronek, Ondřej. "Anotace obrazu a videa formou hry." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2014. http://www.nusl.cz/ntk/nusl-236063.
Full textVrochidis, Stefanos. "Interactive video retrieval using implicit user feedback." Thesis, Queen Mary, University of London, 2013. http://qmro.qmul.ac.uk/xmlui/handle/123456789/8729.
Full textOliveira, Laura Sofia Martins. "Interfaces for television content sharing and annotation." Master's thesis, Faculdade de Ciências e Tecnologia, 2013. http://hdl.handle.net/10362/10973.
Full textThe ways of television consumption and production are changing significantly, with the viewers moving away from the traditional linear model. The various devices for accessing content have a significant role in these changes and suggest new paradigms of access. Social experience has also changed and takes on new forms molded by technology. Content sharing and production from users are some of the trends that globally influence how we relate to audiovisual content. Therefore the aim is to develop ways to access television content, that allow commenting and sharing, through multimodal annotations. These annotations include text, sketches, handwriting and images. Our solution provides users a way to watch and annotate television content, in real-time and in a collaborative environment. Using a mobile device, users can annotate content together with other users, while watching both content and annotations on a TV. These annotations can also be shared through social networks or saved on other platforms. Finally, the system also uses content provided by the users to search and link to television content.
Fundação para a Ciência e a Tecnologia - UTA-Est/MAI/0010/2009 - project ImTV (On-Demand Immersive-TV for Communities of Media Producers and Consumers)
Kucuk, Dilek. "Exploiting Information Extraction Techniques For Automatic Semantic Annotation And Retrieval Of News Videos In Turkish." Phd thesis, METU, 2011. http://etd.lib.metu.edu.tr/upload/12613043/index.pdf.
Full textVolkmer, Timo, and timovolkmer@gmx net. "Semantics of Video Shots for Content-based Retrieval." RMIT University. Computer Science and Information Technology, 2007. http://adt.lib.rmit.edu.au/adt/public/adt-VIT20090220.122213.
Full textChen, Juan. "Content-based Digital Video Processing. Digital Videos Segmentation, Retrieval and Interpretation." Thesis, University of Bradford, 2009. http://hdl.handle.net/10454/4256.
Full textUggerud, Nils. "AnnotEasy: A gesture and speech-to-text based video annotation tool for note taking in pre-recorded lectures in higher education." Thesis, Linnéuniversitetet, Institutionen för datavetenskap och medieteknik (DM), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-105962.
Full textCabral, Diogo Nuno Crespo Ribeiro. "Video interaction using pen-based technology." Doctoral thesis, Faculdade de Ciências e Tecnologia, 2014. http://hdl.handle.net/10362/11503.
Full textVideo can be considered one of the most complete and complex media and its manipulating is still a difficult and tedious task. This research applies pen-based technology to video manipulation, with the goal to improve this interaction. Even though the human familiarity with pen-based devices, how they can be used on video interaction, in order to improve it, making it more natural and at the same time fostering the user’s creativity is an open question. Two types of interaction with video were considered in this work: video annotation and video editing. Each interaction type allows the study of one of the interaction modes of using pen-based technology: indirectly, through digital ink, or directly, trough pen gestures or pressure. This research contributes with two approaches for pen-based video interaction: pen-based video annotations and video as ink. The first uses pen-based annotations combined with motion tracking algorithms, in order to augment video content with sketches or handwritten notes. It aims to study how pen-based technology can be used to annotate a moving objects and how to maintain the association between a pen-based annotations and the annotated moving object The second concept replaces digital ink by video content, studding how pen gestures and pressure can be used on video editing and what kind of changes are needed in the interface, in order to provide a more familiar and creative interaction in this usage context.
This work was partially funded by the UTAustin-Portugal, Digital Media, Program (Ph.D. grant: SFRH/BD/42662/2007 - FCT/MCTES); by the HP Technology for Teaching Grant Initiative 2006; by the project "TKB - A Transmedia Knowledge Base for contemporary dance" (PTDC/EAT/AVP/098220/2008 funded by FCT/MCTES); and by CITI/DI/FCT/UNL (PEst-OE/EEI/UI0527/2011)
Jakob, Persson. "How to annotate in video for training machine learning with a good workflow." Thesis, Umeå universitet, Institutionen för tillämpad fysik och elektronik, 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-187078.
Full textArtificiell intelligens och maskininlärning används inom många olika områden, ett av dessa områden är bildigenkänning. Vid produktionen av ett TV-program eller av en film kan bildigenkänning användas för att hjälpa redigerarna att hitta specifika objekt, scener eller personer i videoinnehållet, vilket påskyndar produktionen. Men bildigenkänningsprogram fungerar inte alltid helt perfekt och kan inte användas i produktionen av ett TV-program eller film som det är tänkt att användas i det sammanhanget. För att förbättra bildigenkänningsprogram så behöver dess algoritm tränas på stora datasets av bilder och labels. Men att skapa dessa datasets tar tid och det behövs program som kan skapa datasets och återträna algoritmer för bildigenkänning så att de fungerar bättre. Syftet med detta examensarbete var att undersöka om det var möjligt att skapa ett verktyg som kan markera(annotera) objekt och personer i video och använda datat som träningsdata för algoritmer. Men även att skapa ett verktyg som kan återträna algoritmer för bildigenkänning så att de blir bättre utifrån datat man får från ett bildigenkänningprogram. Det var också viktigt att dessa verktyg hade ett bra arbetsflöde för användarna. Studien bestod av en teoretisk studie för att få mer kunskap om annoteringar i video och hur man skapar bra UX-design med ett bra arbetsflöde. Intervjuer hölls också för att få mer kunskap om kraven på produkten och vilka som skulle använda den. Det resulterade i ett användarscenario och ett arbetsflöde som användes tillsammans med kunskapen från den teoretiska studien för att skapa en hi-fi prototyp, där en iterativ process med användbarhetstestning användes. Detta resulterade i en slutlig hi-fi prototyp med bra design och ett bra arbetsflöde för användarna där det är möjligt att markera(annotera) objekt och personer med en bounding box och där det är möjligt att återträna algoritmer för bildigenkänning som har körts på video.
Ren, Jinchang. "Semantic content analysis for effective video segmentation, summarisation and retrieval." Thesis, University of Bradford, 2009. http://hdl.handle.net/10454/4251.
Full textGokturk, Ozkan Ziya. "Metadata Extraction From Text In Soccer Domain." Master's thesis, METU, 2008. http://etd.lib.metu.edu.tr/upload/12609871/index.pdf.
Full text#64257
nd accompanying text with the video, such as soccer domain, movie domain and news domain. In this thesis, we present an approach of metadata extraction from match reports for soccer domain. The UEFA Cup and UEFA Champions League Match Reports are downloaded from the web site of UEFA by a web-crawler. These match reports are preprocessed by using regular expressions and then important events are extracted by using hand-written rules. In addition to hand-written rules, two di&
#64256
erent machine learning techniques are applied on match corpus to learn event patterns and automatically extract match events. Extracted events are saved in an MPEG-7 &
#64257
le. A user interface is implemented to query the events in the MPEG-7 match corpus and view the corresponding video segments.
Nielsen, Ryler Jay. "Japanese Vocabulary Learning Through an Interactive Video Platform: Comparative Effects of L1 Versus L2 Definitions and Kana Versus Kanji Presentation." BYU ScholarsArchive, 2016. https://scholarsarchive.byu.edu/etd/6096.
Full textDerathé, Arthur. "Modélisation de la qualité de gestes chirurgicaux laparoscopiques." Thesis, Université Grenoble Alpes, 2020. https://thares.univ-grenoble-alpes.fr/2020GRALS021.pdf.
Full textSous cœlioscopie, le traitement chirurgical permet une meilleure prise en charge du patient, et sa pratique est de plus en plus fréquente en routine clinique. Cette pratique présente néanmoins ses difficultés propres pour le chirurgien, et nécessite une formation prolongée pendant l’internat et en post-internat. Pour faciliter cette formation, il est notamment possible de développer des outils d’évaluation et d’analyse de la pratique chirurgicale.Dans cette optique, l’objectif de ce travail de thèse est d’étudier la faisabilité d’une méthodologie proposant, à partir d’un traitement algorithmique, des analyses à portée clinique pertinente pour le chirurgien. J’ai donc traité les problèmes suivants : Il m’a fallu recueillir et annoter un jeu de données, implémenter un environnement d’apprentissage dédié à la prédiction d’un aspect spécifique de la pratique chirurgicale, et proposer une approche permettant de traduire mes résultats algorithmiques sous une forme pertinente pour le chirurgien. Dès que cela était possible, nous avons cherché à valider ces différentes étapes de la méthodologie
Tran, Hoang Tung. "Automatic tag correction in videos : an approach based on frequent pattern mining." Thesis, Saint-Etienne, 2014. http://www.theses.fr/2014STET4028/document.
Full textThis thesis presents a new system for video auto tagging which aims at correcting the tags provided by users for videos uploaded on the Internet. Most existing auto-tagging systems rely mainly on the textual information and learn a great number of classifiers (on per possible tag) to tag new videos. However, the existing user-provided video annotations are often incorrect and incomplete. Indeed, users uploading videos might often want to rapidly increase their video’s number-of-view by tagging them with popular tags which are irrelevant to the video. They can also forget an obvious tag which might greatly help an indexing process. In this thesis, we limit the use this questionable textual information and do not build a supervised model to perform the tag propagation. We propose to compare directly the visual content of the videos described by different sets of features such as SIFT-based Bag-Of-visual-Words or frequent patterns built from them. We then propose an original tag correction strategy based on the frequency of the tags in the visual neighborhood of the videos. We have also introduced a number of strategies and datasets to evaluate our system. The experiments show that our method can effectively improve the existing tags and that frequent patterns build from Bag-Of-visual-Words are useful to construct accurate visual features
Tarakci, Hilal. "An Ontology-based Multimedia Information Management System." Master's thesis, METU, 2008. http://etd.lib.metu.edu.tr/upload/12609865/index.pdf.
Full textHeggland, Jon. "OntoLog : Flexible Management of Semantic Video Content Annotations." Doctoral thesis, Norwegian University of Science and Technology, Faculty of Information Technology, Mathematics and Electrical Engineering, 2005. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-707.
Full textTo encode, query and present the semantic content of digital video precisely and flexibly is very useful for many kinds of knowledge work: system analysis and evaluation, documentation and education, to name a few. However, that kind of video management is not a trivial matter. The traditional stratified annotation model has quite poor facilities for specifying the meaning – the structure and relationships – of the strata. Because of this, it may also be troublesome to present the annotations to the users in a clear and flexible manner.
This thesis presents OntoLog, a system for managing the semantic content of video. It extends the stratified annotation model by defining the strata as objects and classes in ontologies, thereby making their semantic meaning more explicit and relating them to each other in a semantic network. The same ontologies are also used to define properties and objects for describing both the strata, individual video intervals and entire videos. This constitutes a very customisable, expressive and precise description model, without sacrificing simplicity and conceptual integrity.
Arranging the annotation strata in a near-hierarchical network with specified semantics (classes, subclasses and instances) also enables reasoning about the annotations during query and browsing. In particular, it enables visual aggregation of traditional timeline-based strata graphics. Using this to create compact content visualisations, the OntoLog system is able to present tens of videos on screen at the same time, thus providing inter-video browsing. By judiciously disaggregating selected parts of the strata hierarchy, users can focus on relevant strata at their preferred level of detail – overview-and-zoom functionality for semantic annotations, in other words.
The OntoLog system has been implemented in the form of six Java applications and web services – together covering annotation editing, browsing, analysis, search, query and presentation with various approaches – built on top of an RDF database founded on SQL. The system has been tested under realistic conditions in several real-world projects, with good results. A novel information gathering interface for OntoLog data, Savanta, has been created. This is based on an iterative interaction paradigm featuring inter-video browsing, filtering, navigation and context-sensitive temporal analysis of the annotations. In a comparative usability evaluation, Savanta is shown to outperform more traditional user interfaces for video search/browsing with regard to expressive power, straightforwardness and user satisfaction.
Yilmazturk, Mehmet Celaleddin. "Online And Semi-automatic Annotation Of Faces In Personal Videos." Master's thesis, METU, 2010. http://etd.lib.metu.edu.tr/upload/12611936/index.pdf.
Full textYaprakkaya, Gokhan. "Face Identification, Gender And Age Groups Classifications For Semantic Annotation Of Videos." Thesis, METU, 2010. http://etd.lib.metu.edu.tr/upload/12612848/index.pdf.
Full textZilevu, Kobla Setor. "Interactive Interfaces for Capturing and Annotating Videos of Human Movement." Thesis, Virginia Tech, 2019. http://hdl.handle.net/10919/91424.
Full textMaster of Science
In this thesis, I describe the iterative service design process I used in identifying and understanding the needs of diverse stakeholders, the development of technologies to support their mutually beneficial needs, and the evaluation of the end-user experience with these technologies. Over three iterative design cycles, the set of identified end-user customers expanded to include the patient, the supervising therapist, the annotating therapist, and other members of the development team. Multiple versions of interactive movement capture and annotation tools were developed as the needs of these stakeholders were clarified and evolved, and the optimal data forms and structures became evident Interactions between the stakeholders and the developed technologies operating in various environments were evaluated and assessed to help improve and optimize the entire service ecosystem. Results and findings from these three design cycles are being used to direct and shape my ongoing and future doctoral research.
Rocque, Ryan K. "A Study of the Effectiveness of Annotations in Improving the Listening Comprehension of Intermediate ESL Learners." Diss., CLICK HERE for online access, 2008. http://contentdm.lib.byu.edu/ETD/image/etd2370.pdf.
Full textZhao, Yang. "Person Retrieval with Deep Learning." Thesis, Griffith University, 2022. http://hdl.handle.net/10072/411526.
Full textThesis (PhD Doctorate)
Doctor of Philosophy (PhD)
School of Eng & Built Env
Science, Environment, Engineering and Technology
Full Text
Prié, Yannick. "Vers une phénoménologie des inscriptions numériques. Dynamique de l'activité et des structures informationnelles dans les systèmes d'interprétation." Habilitation à diriger des recherches, Université Claude Bernard - Lyon I, 2011. http://tel.archives-ouvertes.fr/tel-00655574.
Full textQiu, Shun-Qun, and 邱順群. "Video Bookmarks – Prototype of Video Annotation Tool." Thesis, 2018. http://ndltd.ncl.edu.tw/handle/a6ddv7.
Full text實踐大學
工業產品設計學系碩士班
106
"Video Bookmarks" is a supplementary information on the timeline in a movie screen. It has two display types. The first type is unexpanded and is presented on the timeline in the form of a dialog bubble; after being clicked, it is rendered in a dialog window, providing the comment content of the video paragraph.This study is to explore the possibility of the development of video bookmarker annotation tool, which is based on the learn and discussion of video. The research is divided into four stages: "using requirements exploration", "designing with users", "usability testing", "product prototype description", and so on. According to the interview data and the three versions of the transformation process to show the results of the study. Since the timeline of the video player has both the "status display" and "accept adjustments" features. Adding a subsidy message around the timeline will change the "status display" of the timeline to help the user pre-identify the selected video paragraph. Therefore, film bookmarks can provide three kinds of services: time axis supplementary information, annotated content to strengthen specific information, annotated content to bring up a total of contrasting information, and these three services have become the guiding signs for users in the process of non-linear browsing. According to the interview data, this design is suitable for "closed discussion" and knowledge representation belongs to "procedural knowledge" educational context, in order to help teachers and students to communicate and review the blind spot in the learning process. There are two kinds of operation behavior: reading video bookmark and adding video bookmark, which correspond to " video bookmark" and " annotation tool " and so on. Enables users to use annotation tool to edit video bookmark to clearly communicate the message. The tagging tool should retain the text string, graffiti, voice input and combined with the film control instructions to remove the pressure of time for the user to facilitate editing operations. The size of the video bookmark, display location and interactive type are important design conditions of this product. These are the development points of the basic version of the Movie Bookmarker-tagging tool. From the data from the usability test, we can infer the mapping effect of user identity "video bookmark" on the timeline and consider it to be a nudge. In the future, the "annotation tool " can be adapted in detail to help teachers and students edit " video bookmarks" so that both sides can have a cross-space-time exchange and discussion based on the video.
Wu, Wen-Kai, and 吳文凱. "Wiki-based Collaborative Video Annotation System." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/11740303788819181923.
Full text國立暨南國際大學
資訊工程學系
97
With the rapid development of the Web 2.0, sharing information on the social network is common and popular like Wikipedia and Blog. Wikipedia is the representative collaborative platform to allow users to write and edit the web content. Furthermore, Streaming and compression technology allow users to access multimedia data through network more easily than before. YouTube is the famous website to share the video. Our system offers an easy-to-use interface to annotate and add information on video. By integrating the Wiki spirit with sharing YouTube video, the user can collaboratively annotate and edit the video using our system platform. Finally, using the equation to determine the highest credit, the system provides the recommended annotated video to users.
Hong, Guo-Xiang, and 洪國翔. "Optimal Training Set Selection for Video Annotation." Thesis, 2009. http://ndltd.ncl.edu.tw/handle/28133091433393507547.
Full text國立清華大學
電機工程學系
97
Most learning-based video semantic analysis methods hope to obtain the good semantic model that require a large training set to achieve good performances. However, annotating a large video is labor-intensive and the training data set collection is not easy either. Generally, most of training set selection schemes adopted a random selection or select parts of the video data, they neglect similar and extensive characteristics of the training set. In this thesis, we propose several different methods to construct the training set and reduce user involvement. They are clustering-based, spatial dispersiveness, temporal dispersiveness, and sample-based. Using those selection schemes, we hope to construct a small size and effective training set by using spatial and temporal distribution, clustering information of the whole video data. Therefore if the selected training data can represent the characteristic of the whole video data, the classification performance will be better even when the size of the training set is smaller than that of the whole video data. We can choose the best samples for training a semantic model and use SVM to classify the category of each sample. This thesis intends to classify the shots of the semantic into the five categories: person, landscape, cityscape, map and others. Experimental results show that these methods are effective for training set selection in video annotation, and outperform random selection.