Dissertations / Theses on the topic 'HTML documents'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 dissertations / theses for your research on the topic 'HTML documents.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Xie, Wei University of Ballarat. "Classification of HTML Documents." University of Ballarat, 2006. http://archimedes.ballarat.edu.au:8080/vital/access/HandleResolver/1959.17/12774.
Full textMaster of Computing
Xie, Wei. "Classification of HTML Documents." University of Ballarat, 2006. http://archimedes.ballarat.edu.au:8080/vital/access/HandleResolver/1959.17/15628.
Full textMaster of Computing
Levering, Ryan Reed. "Multi-stage modeling of HTML documents." Diss., Online access via UMI:, 2004.
Find full textStachowiak, Maciej 1976. "Automated extraction of structured data from HTML documents." Thesis, Massachusetts Institute of Technology, 1998. http://hdl.handle.net/1721.1/9896.
Full textIncludes bibliographical references (leaf 45).
by Maciej Stachowiak.
M.Eng.
Nálevka, Petr. "Compound XML documents." Master's thesis, Vysoká škola ekonomická v Praze, 2007. http://www.nusl.cz/ntk/nusl-1746.
Full textTemelkuran, Baris 1980. "Hap-Shu : a language for locating information in HTML documents." Thesis, Massachusetts Institute of Technology, 2003. http://hdl.handle.net/1721.1/87882.
Full textMeziane, Souad. "Analyse et conversion de documents : du pixel au langage HTML." Lyon, INSA, 1998. http://www.theses.fr/1998ISAL0128.
Full textThis work is part of the thematic "Document Analysis" in the Laboratory Reconnaissance de Forme et Vision(RFV). To achieve an analysis system ables to, interpret documents and to restore its structure, the Methodologies we have chosen lean on several approaches and particularly on the syntactic and structural approach of the Pattern Recognition. The aim in this work is to convert some paper documents into HTML documents because these documents are more used on the Internet. The application domain of such systems could be general; however, we concentrate us on a particular type of documents with a rich typography: the summaries. In this context, we have realized a system that exploits on one hand the information about content of the document such as its physical and logical structures, and on the other hand on two level grammars. It is composed with two grammars: a meta-grammar and a hyper-grammar. In our system, the role of the meta-grammar is to describe the physical and logical structures of the document. The hyper-grammar is constituted with a set of calculus rules and describes the treatments to do in order to convert the document in HTML. The summary analysis is done in two steps: analysis and identification of the document, and then translation into HTML. During of the first step, the system constructs a learning base by using the grammatical inference. This base contains several patterns of synopses to identify. An unknown document, submitted to the system is identified by matching with the patterns of the base by using all the attributes obtained in the analysis step. The layout of HTML document construction is based on the grammatical analysis of the hyper-grammar. The last is obtained by translation of the logical labels and some typographic parameters into HTML commands. The result of the grammatical analysis of the hyper-grammar produces the structured HTML document corresponding to the studied document. This last will be visualized by software of navigation
Mohammadzadeh, Hadi. "Improving Retrieval Accuracy in Main Content Extraction from HTML Web Documents." Doctoral thesis, Universitätsbibliothek Leipzig, 2013. http://nbn-resolving.de/urn:nbn:de:bsz:15-qucosa-130500.
Full textDas rasante Wachstum von textbasierten Informationen im World Wide Web und die Vielfalt der Anwendungen, die diese Daten nutzen, macht es notwendig, effiziente und effektive Methoden zu entwickeln, die den Hauptinhalt identifizieren und von den zusätzlichen Inhaltsobjekten wie z.B. Navigations-Menüs, Anzeigen, Design-Elementen oder Haftungsausschlüssen trennen. Zunächst untersuchen, entwickeln und evaluieren wir in dieser Arbeit R2L, DANA, DANAg und AdDANAg, eine Familie von neuartigen Algorithmen zum Extrahieren des Inhalts von Web-Dokumenten. Das grundlegende Konzept hinter R2L, das auch zur Entwicklung der drei weiteren Algorithmen führte, nutzt die Besonderheiten der Rechts-nach-links-Sprachen aus, um den Hauptinhalt von Webseiten zu extrahieren. Da der lateinische Zeichensatz und die Rechts-nach-links-Zeichensätze durch verschiedene Abschnitte des Unicode-Zeichensatzes kodiert werden, lassen sich die Rechts-nach-links-Zeichen leicht von den lateinischen Zeichen in einer HTML-Datei unterscheiden. Das erlaubt dem R2L-Ansatz, Bereiche mit einer hohen Dichte von Rechts-nach-links-Zeichen und wenigen lateinischen Zeichen aus einer HTML-Datei zu erkennen. Aus diesen Bereichen kann dann R2L die Rechts-nach-links-Zeichen extrahieren. Die erste Erweiterung, DANA, verbessert die Wirksamkeit des Baseline-Algorithmus durch die Verwendung eines HTML-Parsers in der Nachbearbeitungsphase des R2L-Algorithmus, um den Inhalt aus Bereichen mit einer hohen Dichte von Rechts-nach-links-Zeichen zu extrahieren. DANAg erweitert den Ansatz des R2L-Algorithmus, so dass eine Sprachunabhängigkeit erreicht wird. Die dritte Erweiterung, AdDANAg, integriert eine neue Vorverarbeitungsschritte, um u.a. die Weblinks zu normalisieren. Die vorgestellten Ansätze werden in Bezug auf Effizienz und Effektivität analysiert. Im Vergleich mit mehreren etablierten Hauptinhalt-Extraktions-Algorithmen zeigen wir, dass sie in diesen Punkten überlegen sind. Darüber hinaus findet die Extraktion der Überschriften aus Web-Artikeln vielfältige Anwendungen. Hierzu entwickeln wir mit TitleFinder einen sich nur auf den Textinhalt beziehenden und sprachabhängigen Ansatz. Das vorgestellte Verfahren ist in Bezug auf Effektivität und Effizienz besser als bekannte Ansätze, die auf strukturellen und visuellen Eigenschaften der HTML-Datei beruhen
Yerra, Rajiv. "Detecting Similar HTML Documents Using A Sentence-Based Copy Detection Approach." Diss., CLICK HERE for online access, 2005. http://contentdm.lib.byu.edu/ETD/image/etd977.pdf.
Full textSinger, Ron. "Comparing machine learning and hand-crafted approaches for information extraction from HTML documents." Thesis, McGill University, 2003. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=79127.
Full textJABOUR, IAM VITA. "THE IMPACT OF STRUCTURAL ATTRIBUTES TO IDENTIFY TABLES AND LISTS IN HTML DOCUMENTS." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2010. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=17247@1.
Full textA segmentação de documentos HTML tem sido essencial para as tarefas de extração de informações, como mostram vários estudos na área. Nesta dissertação investigamos a relação entre o documento HTML e sua representação visual, mostrando como esta ligação ajuda na abordagem estrutural para a identificação de segmentos. Também investigamos como utilizar algoritmos de distância de edição em árvores para encontrar padrões na árvore DOM, tornando possível resolver duas tarefas de identificação de segmentos. A primeira tarefa é a identificação de tabelas genuínas, aonde foi obtido 90,40% de F1 utilizando o corpus fornecido por (Wang e Hu, 2002). Mostramos através de um estudo experimental que este resultado é competitivo com os melhores resultados da área. A segunda tarefa que consideramos é a identificação de listas de produtos em sites de comércio eletrônico, nessa obtivemos 94,95% de F1 utilizando um corpus com 1114 documentos HTML, criado a partir de 8 sites. Concluímos que os algoritmos de similaridade estrutural ajudam na resolução de ambas às tarefas e acreditamos que possam ajudar na identificação de outros tipos de segmentos.
The segmentation of HTML documents has been essential to information extraction tasks, as showed by several works in this area. This paper studies the link between an HTML document and its visual representation to show how it helps segments identification using a structural approach. For this, we investigate how tree edit distance algorithms can find structural similarities in a DOM tree, using two tasks to execute our experiments. The first one is the identification of genuine tables where we obtained a 90.40% F1 score using the corpus provided by (Wang e Hu, 2002). We show through an experimental study that this result is competitive with the best results in the area. The second task studied is the identification of product listings in e-commerce sites. Here we get a 94.95% F1 score using a corpus with 1114 HTML documents from 8 distinct sites. We conclude that algorithms to calculate trees similarity provide competitive results for both tasks, making them also good candidates to identify other types of segments.
Mysore, Gopinath Abhijith Athreya. "Automatic Detection of Section Title and Prose Text in HTML Documents Using Unsupervised and Supervised Learning." University of Cincinnati / OhioLINK, 2018. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1535371714338677.
Full textAl, Assimi Abdel-Basset. "Gestion de l'évolution non centralisée de documents parallèles multilingues." Université Joseph Fourier (Grenoble), 2000. http://www.theses.fr/2000GRE10127.
Full textBukovčák, Jakub. "Extrakce informací z webových stránek." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2019. http://www.nusl.cz/ntk/nusl-403184.
Full textMohammadzadeh, Hadi [Verfasser], Gerhard [Akademischer Betreuer] Heyer, Gerhard [Gutachter] Heyer, and Jinan [Gutachter] Fiaidhi. "Improving Retrieval Accuracy in Main Content Extraction from HTML Web Documents / Hadi Mohammadzadeh ; Gutachter: Gerhard Heyer, Jinan Fiaidhi ; Betreuer: Gerhard Heyer." Leipzig : Universitätsbibliothek Leipzig, 2013. http://d-nb.info/1237818303/34.
Full textCheriat, Ahmed. "Une Méthode de correction de la structure de documents XML dans le cadre d'une validation incrémentale." Tours, 2006. http://www.theses.fr/2006TOUR4022.
Full textXML becomes the main tool used to exchange data on the web. In this context, XML document should respect schema constraints that describe the structural form of XML documents. The validation of an XML document wrt a schema constraints consists in testing whether the document verifies the set of structural specifications described by this schema. Supposing that updates are applied to the document, an incremental validator is the one that verifies whether the updated document complies with the schema, by validating only the parts of the document involved in the updates (to reduce the cost of a validation from scratch of the whole XML document). In this thesis we associate the validation process with correctoin proposals. During the execution of our validation method, if a constraint violation is found, a correction routine is called in order to propose local solutions capable of allowing the validation process to continue. Firstly, we are interested in a special case of this problem, which consists in correcting a word wrt another word (the initial valid word) and wrt a regular language. Indeed, the correction of an XML document having only a root and its sons corresponds to the correction of a word (composed by the children of the root) wrt a regular language (the constraint associated tothe root). In a second time, we extended this idea to the correction of XML tree (the structure of documents) wrt tree languages (the schema associated to an XML document). This correction is done by applying the minimum of modifications (by insertion, deletion or by replacement of some elements) on an invalid XML document in order to obtain a valid XML document. The algorithms presented in this thesis were implemented (in Java) and the experimental result are shown
Grabs, Torsten. "Storage and retrieval of XML documents with a cluster of database systems /." Berlin : Aka, 2003. http://www.loc.gov/catdir/toc/fy0713/2007435297.html.
Full textEeg-Tverbakk, Camilla. "Theatre-ting : toward a materialist practice of staging documents." Thesis, University of Roehampton, 2016. https://pure.roehampton.ac.uk/portal/en/studentthesis/theatre-–-ting(dd5f299e-6fdc-4c69-bed6-7ae690de6a8d).html.
Full textKocman, Radim. "Podpora dynamického DOM v zobrazovacím stroji HTML." Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2014. http://www.nusl.cz/ntk/nusl-236139.
Full textMilosevic, Nikola. "A multi-layered approach to information extraction from tables in biomedical documents." Thesis, University of Manchester, 2018. https://www.research.manchester.ac.uk/portal/en/theses/a-multilayered-approach-to-information-extraction-from-tables-in-biomedical-documents(c2edce9c-ae7f-48fa-81c2-14d4bb87423e).html.
Full textMerkl-Davies, Doris. "The obfuscation hypothesis re-examined : analyzing impression management in corporate narrative report documents." Thesis, Bangor University, 2007. https://research.bangor.ac.uk/portal/en/theses/the-obfuscation-hypothesis-reexamined--analyzing-impression-management-in-corporate-narrative-report-documents(3fd58e2c-790a-44b7-80c8-2c4b41ef72c3).html.
Full textTao, Cui. "Schema Matching and Data Extraction over HTML Tables." Diss., CLICK HERE for online access, 2003. http://contentdm.lib.byu.edu/ETD/image/etd279.pdf.
Full textHaber, Renato Ribeiro. "Uma Ferramenta de Importação de Documentos HTML para um Ambiente de Ensino." Universidade de São Paulo, 1999. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-09032018-141601/.
Full textThis work presents a tool prototype, the Html2flip, that provides an importation and adaptation environment of documents described in HTML (HyperText Markup Language) standard for the internai representation of the SASHE (Hypermeclia System for Authorship and Supporting Educational Applications), that is based on the structural organization of multimeclia objects, proposed by MCA (Nested Contexts Model). Moreover, this work extended the capacity of the information text node editor of the previous prototype conceming the processing of text-files described in RTF (Rich Text Formar) standard. This way, the SASHE became capable to process and to organize instructional materiais prepared in its proper environment, in the WWW (World- Wide Web) environment, as well as in common word processors.
Silva, Patrick Pedreira. "ExtraWeb: um sumarizador de documentos Web baseado em etiquetas HTML e ontologia." Universidade Federal de São Carlos, 2006. https://repositorio.ufscar.br/handle/ufscar/322.
Full textFinanciadora de Estudos e Projetos
This dissertation presents an automatic summarizer of Web documents based on both HTML tags and ontological knowledge. It has been derived from two independent approaches: one that focuses solely upon HTML tags, and another that focuses only on ontological knowledge. The three approaches were implemented and assessed, indicating that associating both knowledge types have a promising descriptive power for Web documents. The resulting prototype has been named ExtraWeb. The ExtraWeb system explores the HTML structure of Web documents in Portuguese and semantic information using the Yahoo ontology in Portuguese. This has been enriched with additional terms extracted from both a thesaurus, Diadorim and the Wikipedia. In a simulated Web search, ExtraWeb achieved a similar utility degree to Google one, showing its potential to signal through extracts the relevance of the retrieved documents. This has been an important issue recently. Extracts may be particularly useful as surrogates of the current descriptions provided by the existing search engines. They may even substitute the corresponding source documents. In the former case, those descriptions do not necessarily convey relevant content of the documents; in the latter, reading full documents demands a substantial overhead of Web users. In both cases, extracts may improve the search task, provided that they actually signal relevant content. So, ExtraWeb is a potential plug-in of search engines, to improve their descriptions. However, its scability and insertion in a real setting have not yet been explored.
Esta dissertação propõe um sumarizador de documentos Web baseado em etiquetas HTML e conhecimento ontológico, derivado de outras duas abordagens independentes: uma que contempla somente etiquetas HTML e outra, somente conhecimento ontológico. As três abordagens foram implementadas e avaliadas, indicando que a composição desses dois tipos de conhecimento tem um bom potencial descritivo de documentos Web. O protótipo resultante é denominado ExtraWeb. O ExtraWeb explora a estrutura de marcação de documentos em português e informações de nível semântico usando a ontologia do Yahoo em português, enriquecida com vocabulário extraído de um thesaurus, Diadorim, e da Wikipédia. Em uma tarefa simulada por internautas, de busca de documentos, o ExtraWeb obteve um grau de utilidade próximo ao do Google, evidenciando seu potencial para indicar, por meio de extratos, a relevância de documentos recuperados na Web. Esse foco é de grande interesse atualmente, pois os extratos podem ser particularmente úteis como substitutos das descrições atuais das ferramentas de busca ou, mesmo, como substitutos dos documentos correspondentes completos. No primeiro caso, as descrições nem sempre contemplam as informações mais relevantes dos documentos; no segundo, sua leitura implica um esforço considerável por parte do internauta. Em ambos os casos, extratos podem otimizar essa tarefa, se comprovada sua utilidade para a indicação da relevância dos documentos. Assim, o ExtraWeb tem potencial para ser um acessório das ferramentas de busca, para melhorar a forma como os resultados são apresentados, muito embora sua escalabilidade e implantação em um ambiente real ainda não tenham sido exploradas.
Oliver, Robert W. "The vocation of the laity to evangelization an ecclesiological inquiry into the Synod on the laity (1987), Christifideles laici (1989), and documents of the NCCB (1987-1996) /." Roma : Editrice Pontificia Università Gregoriana, 1997. http://catalog.hathitrust.org/api/volumes/oclc/37849170.html.
Full textChen, Xueqi. "Query Rewriting for Extracting Data behind HTML Forms." Diss., CLICK HERE for online access, 2004. http://contentdm.lib.byu.edu/ETD/image/etd406.Chen.
Full textStewart, Jeffrey D. "An XML-based knowledge management system of port information for U.S. Coast Guard Cutters." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 2003. http://library.nps.navy.mil/uhtbin/hyperion-image/03Mar%5FStewart.pdf.
Full textThesis advisor(s): Magdi N. Kamel, Gordon H. Bradley. Includes bibliographical references (p. 101-103). Also available online.
Chen, Benfeng. "Transforming Web pages to become standard-compliant through reverse engineering /." View abstract or full-text, 2006. http://library.ust.hk/cgi/db/thesis.pl?COMP%202006%20CHEN.
Full textMull, Randall Franklin. "Teaching web design at the higher education level." Morgantown, W. Va. : [West Virginia University Libraries], 2001. http://etd.wvu.edu/templates/showETD.cfm?recnum=1954.
Full textTitle from document title page. Document formatted into pages; contains iii, 47 p. Vita. Includes abstract. Includes bibliographical references (p. 36-37).
Nicoletti, Alberto. "Conversione di documenti DOCX in formato RASH." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2016. http://amslaurea.unibo.it/12298/.
Full textMajdzadeh, Khandani Kourosh. "Rights and liabilities of the consignees/endorsees : a comparative study of the Rotterdam Rules and English Law." Thesis, University of Manchester, 2018. https://www.research.manchester.ac.uk/portal/en/theses/rights-and-liabilities-of-the-consigneesendorsees-a-comparative-study-of-the-rotterdam-rules-and-english-law(aa10e154-facf-4573-a10f-30786c51e4c0).html.
Full textParker, Rembert N. "An introduction to computer programming for complete beginners using HTML, JavaScript, and C#." CardinalScholar 1.0, 2008. http://liblink.bsu.edu/uhtbin/catkey/1465970.
Full textDepartment of Computer Science
Cohen, Eric Joseph. "An investigation into World Wide Web publishing with the Hypertext Markup Language /." Online version of thesis, 1995. http://hdl.handle.net/1850/12229.
Full textHan, Wei. "Wrapper application generation for semantic web." Diss., Georgia Institute of Technology, 2003. http://hdl.handle.net/1853/5407.
Full textPaolucci, Francesco. "A Fitting Algorithm: applicazione automatica di vincoli tipografici per la stampa di documenti testuali su browser." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2020. http://amslaurea.unibo.it/20534/.
Full textTandon, Seema Amit. "Web Texturizer: Exploring intra web document dependencies." CSUSB ScholarWorks, 2004. https://scholarworks.lib.csusb.edu/etd-project/2539.
Full textMoura, Antonio Gilberto de. "Proposta de um sistema para geração de Applets Java para animação de paginas HTML destinadas a educação a distancia." [s.n.], 2002. http://repositorio.unicamp.br/jspui/handle/REPOSIP/260239.
Full textDissertação (mestrado) - Universidade Estadual de Campinas, Faculdade de Engenharia Eletrica e de Computação
Made available in DSpace on 2018-08-03T15:06:30Z (GMT). No. of bitstreams: 1 Moura_AntonioGilbertode_M.pdf: 9110464 bytes, checksum: 7065ce0ccb8423e69937937ed11d6b03 (MD5) Previous issue date: 2002
Mestrado
Al-Dallal, Ammar Sami. "Enhancing recall and precision of web search using genetic algorithm." Thesis, Brunel University, 2012. http://bura.brunel.ac.uk/handle/2438/7379.
Full textRubano, Vincenzo. "L'(in)accessibilità degli articoli scientifici sul Web e l'uso di RASH e EPUB." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2016. http://amslaurea.unibo.it/12281/.
Full textWest, Philip. "A framework for responsive content adaptation in electronic display networks." Thesis, Rhodes University, 2006. http://hdl.handle.net/10962/d1004824.
Full textTicona, Quispe Miguel. "La conservacion preventiva y curativa de los documentos publicos oficiales en la Biblioteca Central de la Universidad Mayor de San Andres." Universidad Mayor de San Andrés. Programa Cybertesis BOLIVIA, 2003. http://www.cybertesis.umsa.bo:8080/umsa/2007/ticona_qm/html/index-frames.html.
Full textGonzalez-Ayala, Sofia Natalia. "Black, Afro-Colombian, Raizal and Palenquero communities at the National Museum of Colombia : a reflexive ethnography of (in)visibility, documentation and participatory collaboration." Thesis, University of Manchester, 2016. https://www.research.manchester.ac.uk/portal/en/theses/black-afrocolombian-raizal-and-palenquero-communities-at-the-national-museum-of-colombia-a-reflexive-ethnography-of-invisibility-documentation-and-participatory-collaboration(e40c8594-35c7-49b9-af1c-ccca82cb335f).html.
Full textNagrath, Vineet. "Software architectures for cloud robotics : the 5 view Hyperactive Transaction Meta-Model (HTM5)." Thesis, Dijon, 2015. http://www.theses.fr/2015DIJOS005/document.
Full textSoftware development for cloud connected robotic systems is a complex software engineeringendeavour. These systems are often an amalgamation of one or more robotic platforms, standalonecomputers, mobile devices, server banks, virtual machines, cameras, network elements and ambientintelligence. An agent oriented approach represents robots and other auxiliary systems as agents inthe system.Software development for distributed and diverse systems like cloud robotic systems require specialsoftware modelling processes and tools. Model driven software development for such complexsystems will increase flexibility, reusability, cost effectiveness and overall quality of the end product.The proposed 5-view meta-model has separate meta-models for specifying structure, relationships,trade, system behaviour and hyperactivity in a cloud robotic system. The thesis describes theanatomy of the 5-view Hyperactive Transaction Meta-Model (HTM5) in computation independent,platform independent and platform specific layers. The thesis also describes a domain specificlanguage for computation independent modelling in HTM5.The thesis has presented a complete meta-model for agent oriented cloud robotic systems and hasseveral simulated and real experiment-projects justifying HTM5 as a feasible meta-model
Baroni, Andrea. "Pattern design per documenti strutturati: il problema della conversione da e per formati documentali tradizionali." Bachelor's thesis, Alma Mater Studiorum - Università di Bologna, 2017. http://amslaurea.unibo.it/14422/.
Full textSebastião, Cláudio Barradas. "Proposta de um modelo conceitual de ferramenta para monitoramento de documento na web." Florianópolis, SC, 2003. http://repositorio.ufsc.br/xmlui/handle/123456789/84744.
Full textMade available in DSpace on 2012-10-20T12:07:30Z (GMT). No. of bitstreams: 1 198175.pdf: 1739707 bytes, checksum: 290e2c7ab594e697dc4fabda07e89adb (MD5)
A Web pode ser vista de duas formas: serviços e conteúdo. Conteúdo é o conjunto das informações eletrônicas que podem ser publicadas através do meio Web e por serviços designamos o conjunto de funcionalidades que possibilitam a extração, integração, publicação e visualização do conteúdo. Com esta visão, este estudo comtempla uma grande estruturação de como desenvolver páginas Web e gerencia-las de uma forma prática, segura e responsável, utilizando-se de todas as opções que as inúmeras ferramentas de desenvolvimento Web nos proporcionam.
Sire, Guillaume. "La production journalistique et Google : chercher à ce que l’information soit trouvée." Thesis, Paris 2, 2013. http://www.theses.fr/2013PA020040/document.
Full textIn this thesis, we aim to disentangle the cooperative but also competitive relationship between Google and news publishers, which is at the same time technical, economic, legal, social, political and certainly communicational. In order to do so, we trace the historical development of two singular universes, describing what publishers can do to overcome the search engine and optimize their ranking. We then analyse how Google can influence publishers’ conduct, by studying power relations, respective incentives, aims, and informational and socio-economic backgrounds. Finally, we report on actual practices of French traditional news publishers: what they communicate to Google, by which means and at what price, for which expected results, after which concessions, detours and controversies. Thus, we explain how search engine optimization is likely to affect the way content is valued, its production organisation, the website’s structure, journalists’ prac tice an editorial policy. We show a back and forth movement between performative utterances and performed circumstances, having an effect on and by texts, architexts and hypertexts. To sum up, this thesis is dedicated to understanding what happens to news and publishers once they strive for their information to be found by Google's users
Björk, Linus. "Avancerad webbteknologi i mobila webbläsare." Thesis, Linköpings universitet, Institutionen för datavetenskap, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-65921.
Full textThe web develops fast and web applications are getting more advanced. At the same time the mobile browsers develop at a rapid pace. However, it still differs a lot between a mobile browser and a standard web browser. You also interact with a mobile phone in a different way than what you do with a computer. This thesis examines whether it is possible to create advanced web applications that by utilizing the latest web technologies can replace ordinary mobile applications. The investigation is done by creating a lightweight version of a phone application, Mobile Documents on Symbian S60, which is an application that manages documents, emails and attachments. The development is done in Google Web Toolkit and technologies such as AJAX and Comet are both used. As the number of different types of phones with touch screens is very large the investigation only will target a small number of phones running web browsers as Mobile Safari, microB and Android Browser. The conclusions of this report is that the JavaScript support of today's browsers is enough to run advanced web applications. However, it differs a lot between browsers and the main problem is to create a functional user interface that works equally well on all phones and with all the different interaction possibilities that a mobile phone gives.
Hendges, Graciela Rabuske. "Tackling genre classification." Florianópolis, SC, 2007. http://repositorio.ufsc.br/xmlui/handle/123456789/90448.
Full textMade available in DSpace on 2012-10-23T10:39:26Z (GMT). No. of bitstreams: 1 249271.pdf: 3171345 bytes, checksum: 00f207cece278de30d1f5b7fd246c496 (MD5)
Pesquisas recentes sobre comunicação científica têm revelado que desde o final dos anos de 1990 o uso de periódicos acadêmicos passou da mídia impressa para o mídia eletrônica (Tenopir, 2002, 2003; Tenopir & King, 2001, 2002) e, conseqüentemente, há previsões de que por volta de 2010 cerca de 80% dos periódicos terão apenas versões online (Harnad, 1998). Todavia, essas pesquisas mostram também que nem todas as disciplinas estão migrando para a Internet com a mesma velocidade. Enquanto que áreas como as Ciências da Informação, Arquivologia, Web design e Medicina têm mostrado interesse e preocupação em entnder e explicar esse fenômeno, em Lingüística Aplicada, particularmente em Análise de Gênero, os estudos ainda são escassos. Neste trabalho, portanto, procuro investigar em que medida o meio eletrônico (Internet) afeta o gênero artigo acadêmico no seu processo de mudança da mídia impressa para a mídia eletrônica. Mais especificamente, examino artigos acadêmicos em HTML nas áreas de Lingüística e Medicina com vistas a verificar se esse hypertexto é um gênero novo ou não. A abordagem metodológica adotada nesta pesquisa deriva da proposta de Askehave e Swales (2001) e de Swales (2004), na qual o critéro predominante para a classificação de um gênero é o propósito comunicativo, o qual só pode ser definido com base em uma análise textual tanto quanto em uma análise contextual. Dessa forma, neste estudo foram coletados e analisados dados textuais e contextuais e os resultados de ambas análises revelam que o artigo acadêmico em HTML é um gênero novo, cujo propósito comunicativo é realizado por hiperlinks e portanto, esse gênero é profundamente dependente da mídia eletrônica.
Costa, José Henrique Calenzo. "Filtered-page ranking." reponame:Repositório Institucional da UFSC, 2016. https://repositorio.ufsc.br/xmlui/handle/123456789/167840.
Full textMade available in DSpace on 2016-09-20T04:25:42Z (GMT). No. of bitstreams: 1 341906.pdf: 4935734 bytes, checksum: 5630ca8c10871314b7f54120d18ae335 (MD5) Previous issue date: 2016
Algoritmos de ranking de páginas Web podem ser criados usando técnicas baseadas em elementos estruturais da página Web, em segmentação da página ou na busca personalizada. Esta pesquisa aborda um método de ranking de documentos previamente filtrados, que segmenta a página Web em blocos de três categorias para delas eliminar conteúdo irrelevante. O método de ranking proposto, chamado Filtered-Page Ranking (FPR), consta de duas etapas principais: (i) segmentação da página web e eliminação de conteúdo irrelevante e (ii) ranking de páginas Web. O foco da extração de conteúdo irrelevante é eliminar conteúdos não relacionados à consulta do usuário, através do algoritmo proposto Query-Based Blocks Mining (QBM), para que o ranking considere somente conteúdo relevante. O foco da etapa de ranking é calcular quão relevante cada página Web é para determinada consulta, usando critérios considerados em estudos de recuperação da informação. Com a presente pesquisa pretende-se demonstrar que o QBM extrai eficientemente o conteúdo irrelevante e que os critérios utilizados para calcular quão próximo uma página Web é da consulta são relevantes, produzindo uma média de resultados de ranking de páginas Web de qualidade melhor que a do clássico modelo vetorial.
Abstract : Web page ranking algorithms can be created using content-based, structure-based or user search-based techniques. This research addresses an user search-based approach applied over previously filtered documents ranking, which relies in a segmentation process to extract irrelevante content from documents before ranking. The process splits the document into three categories of blocks in order to fragment the document and eliminate irrelevante content. The ranking method, called Page Filtered Ranking, has two main steps: (i) irrelevante content extraction; and (ii) document ranking. The focus of the extraction step is to eliminate irrelevante content from the document, by means of the Query-Based Blocks Mining algorithm, creating a tree that is evaluated in the ranking process. During the ranking step, the focus is to calculate the relevance of each document for a given query, using criteria that give importance to specific parts of the document and to the highlighted features of some HTML elements. Our proposal is compared to two baselines: the classic vectorial model, and the CETR noise removal algorithm, and the results demonstrate that our irrelevante content removal algorithm improves the results and our relevance criteria are relevant to the process.
Maddipudi, Koushik. "Efficient Architectures for Retrieving Mixed Data with Rest Architecture Style and HTML5 Support." TopSCHOLAR®, 2013. http://digitalcommons.wku.edu/theses/1251.
Full text