To see the other types of publications on this topic, follow the link: Web page.

Dissertations / Theses on the topic 'Web page'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Web page.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Krupp, Brian. "Exploration of Dynamic Web Page Partitioning for Increased Web Page Delivery Performance." Cleveland State University / OhioLINK, 2010. http://rave.ohiolink.edu/etdc/view?acc_num=csu1290629377.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Chiew, Thiam Kian. "Web page performance analysis." Thesis, University of Glasgow, 2009. http://theses.gla.ac.uk/658/.

Full text
Abstract:
Computer systems play an increasingly crucial and ubiquitous role in human endeavour by carrying out or facilitating tasks and providing information and services. How much work these systems can accomplish, within a certain amount of time, using a certain amount of resources, characterises the systems’ performance, which is a major concern when the systems are planned, designed, implemented, deployed, and evolve. As one of the most popular computer systems, the Web is inevitably scrutinised in terms of performance analysis that deals with its speed, capacity, resource utilisation, and availability. Performance analyses for the Web are normally done from the perspective of the Web servers and the underlying network (the Internet). This research, on the other hand, approaches Web performance analysis from the perspective of Web pages. The performance metric of interest here is response time. Response time is studied as an attribute of Web pages, instead of being considered purely a result of network and server conditions. A framework that consists of measurement, modelling, and monitoring (3Ms) of Web pages that revolves around response time is adopted to support the performance analysis activity. The measurement module enables Web page response time to be measured and is used to support the modelling module, which in turn provides references for the monitoring module. The monitoring module estimates response time. The three modules are used in the software development lifecycle to ensure that developed Web pages deliver at worst satisfactory response time (within a maximum acceptable time), or preferably much better response time, thereby maximising the efficiency of the pages. The framework proposes a systematic way to understand response time as it is related to specific characteristics of Web pages and explains how individual Web page response time can be examined and improved.
APA, Harvard, Vancouver, ISO, and other styles
3

Sanoja, Vargas Andrés. "Segmentation de pages web, évaluation et applications." Thesis, Paris 6, 2015. http://www.theses.fr/2015PA066004/document.

Full text
Abstract:
Les pages web sont devenues plus complexes que jamais, principalement parce qu'elles sont générées par des systèmes de gestion de contenu (CMS). Il est donc difficile de les analyser, c'est-à-dire d'identifier et classifier automatiquement les différents éléments qui les composent. La segmentation de pages web est une des solutions à ce problème. Elle consiste à décomposer une page web en segments, visuellement et sémantiquement cohérents, appelés blocs. La qualité d'une segmentation est mesurée par sa correction et sa généricité, c'est-à-dire sa capacité à traiter des pages web de différents types. Notre recherche se concentre sur l'amélioration de la segmentation et sur une mesure fiable et équitable de la qualité des segmenteurs. Nous proposons un modèle pour la segmentation ainsi que notre segmenteur Block-o-Matic (BoM). Nous définissons un modèle d'évaluation qui prend en compte le contenu ainsi que la géométrie des blocs pour mesurer la correction d'un segmenteur par rapport à une vérité de terrain. Ce modèle est générique, il permet de tester tout algorithme de segmentation et observer ses performances sur différents types de page. Nous l'avons testé sur quatre segmenteurs et quatre types de pages. Les résultats montrent que BOM surpasse ses concurrents en général et que la performance relative d'un segmenteur dépend du type de page. Enfin, nous présentons deux applications développées au dessus de BOM. Pagelyzer compare deux versions de pages web et décide si elles sont similaires ou pas. C'est la principale contribution de notre équipe au projet européen Scape (FP7-IP). Nous avons aussi développé un outil de migration de pages HTML4 vers le nouveau format HTML5
Web pages are becoming more complex than ever, as they are generated by Content Management Systems (CMS). Thus, analyzing them, i.e. automatically identifying and classifying different elements from Web pages, such as main content, menus, among others, becomes difficult. A solution to this issue is provided by Web page segmentation which refers to the process of dividing a Web page into visually and semantically coherent segments called blocks.The quality of a Web page segmenter is measured by its correctness and its genericity, i.e. the variety of Web page types it is able to segment. Our research focuses on enhancing this quality and measuring it in a fair and accurate way. We first propose a conceptual model for segmentation, as well as Block-o-Matic (BoM), our Web page segmenter. We propose an evaluation model that takes the content as well as the geometry of blocks into account in order to measure the correctness of a segmentation algorithm according to a predefined ground truth. The quality of four state of the art algorithms is experimentally tested on four types of pages. Our evaluation framework allows testing any segmenter, i.e. measuring their quality. The results show that BoM presents the best performance among the four segmentation algorithms tested, and also that the performance of segmenters depends on the type of page to segment.We present two applications of BoM. Pagelyzer uses BoM for comparing two Web pages versions and decides if they are similar or not. It is the main contribution of our team to the European project Scape (FP7-IP). We also developed a migration tool of Web pages from HTML4 format to HTML5 format in the context of Web archives
APA, Harvard, Vancouver, ISO, and other styles
4

SOUZA, CRISTON PEREIRA DE. "EFFICIENT WEB PAGE REFRESH POLICIES." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2010. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=15893@1.

Full text
Abstract:
CONSELHO NACIONAL DE DESENVOLVIMENTO CIENTÍFICO E TECNOLÓGICO
Uma máquina de busca precisa constantemente revisitar páginas Web para manter seu repositório local atualizado. Uma política de revisitação deve ser empregada para construir um escalonamento de revisitações que mantenha o repositório o mais atualizado possível utilizando os recursos disponíveis. Para evitar sobrecarga de servidores Web, a política de revisitação deve respeitar um tempo mínimo entre requisições consecutivas a um mesmo servidor. Esta regra é chamada restrição de politeness. Devido ao porte do problema, consideramos que uma política de revisitação é eficiente se o tempo médio para escalonar uma revisitação é sublinear no número de páginas do repositório. Neste sentido, quando a restrição de politeness é considerada, não conhecemos política eficiente com garantia teórica de qualidade. Nesta pesquisa investigamos três políticas eficientes que respeitam a restrição de politeness, chamadas MERGE, RANDOM e DELAYED. Fornecemos fatores de aproximação para o nível de atualização do repositório quando empregamos as política MERGE ou RANDOM. Demonstramos que 0,77 é um limite inferior para este fator de aproximação quando empregamos a política RANDOM, e apresentamos uma conjectura de que 0,927 é um limite inferior para este fator de aproximação quando empregamos a política MERGE. As políticas também são avaliadas através da simulação da execução destas políticas para manter o nível de atualização de um repositório contendo 14,5 milhões de páginas Web. Um repositório contendo artigos da Wikipedia também é utilizado nos experimentos, onde podemos observar que a política MERGE apresenta melhores resultados que uma estratégia gulosa natural para este repositório. A principal conclusão desta pesquisa é que existem políticas simples e eficientes para o problema de revisitação de páginas Web, que perdem pouco em termos do nível de atualização do repositório mesmo quando consideramos a restrição de politeness.
A search engine needs to continuously revisit web pages in order to keep its local repository up-to-date. A page revisiting schedule must be defined to keep the repository up-to-date using the available resources. In order to avoid web server overload, the revisiting policy must respect a minimum amount of time between consecutive requests to the same server. This rule is called politeness constraint. Due to the large number of web pages, we consider that a revisiting policy is efficient when the mean time to schedule a revisit is sublinear on the number of pages in the repository. Therefore, when the politeness constraint is considered, there are no existing efficient policies with theoretical quality guarantees. We investigate three efficient policies that respect the politeness constraint, called MERGE, RANDOM and DELAYED. We provide approximation factors for the repository’s up-to-date level for the MERGE and RANDOM policies. Based on these approximation factors, we devise a 0.77 lower bound for the approximation factor provided by the RANDOM policy and we present a conjecture that 0.927 is a lower bound for the approximation factor provided by the MERGE policy. We evaluate these policies through simulation experiments which try to keep a repository with 14.5 million web pages up-to-date. Additional experiments based on a repository with Wikipedia’s articles concluded that the MERGE policy provides better results than a natural greedy strategy. The main conclusion of this research is that there are simple and efficient policies that can be applied to this problem, even when the politeness constraint must be respected, resulting in a small loss of repository’s up-to-date level.
APA, Harvard, Vancouver, ISO, and other styles
5

Hou, Jingyu. "Discovering web page communities for web-based data management." University of Southern Queensland, Faculty of Sciences, 2002. http://eprints.usq.edu.au/archive/00001447/.

Full text
Abstract:
The World Wide Web is a rich source of information and continues to expand in size and complexity. Mainly because the data on the web is lack of rigid and uniform data models or schemas, how to effectively and efficiently manage web data and retrieve information is becoming a challenge problem. Discovering web page communities, which capture the features of the web and web-based data to find intrinsic relationships among the data, is one of the effective ways to solve this problem. A web page community is a set of web pages that has its own logical and semantic structures. In this work, we concentrate on the web data in web page format and exploit hyperlink information to discover (construct) web page communities. Three main web page communities are studied in this work: the first one is consisted of hub and authority pages, the second one is composed of relevant web pages with respect to a given page (URL), and the last one is the community with hierarchical cluster structures. For analysing hyperlinks, we establish a mathematical framework, especially the matrix-based framework, to model hyperlinks. Within this mathematical framework, hyperlink analysis is placed on a solid mathematic base and the results are reliable. For the web page community that is consisted of hub and authority pages, we focus on eliminating noise pages from the concerned page source to obtain another good quality page source, and in turn improve the quality of web page communities. We propose an innovative noise page elimination algorithm based on the hyperlink matrix model and mathematic operations, especially the singular value decomposition (SVD) of matrix. The proposed algorithm exploits hyperlink information among the web pages, reveals page relationships at a deeper level, and numerically defines thresholds for noise page elimination. The experiment results show the effectiveness and feasibility of the algorithm. This algorithm could also be used solely for web-based data management systems to filter unnecessary web pages and reduce the management cost. In order to construct a web page community that is consisted of relevant pages with respect to a given page (URL), we propose two hyperlink based relevant page finding algorithms. The first algorithm comes from the extended co-citation analysis of web pages. It is intuitive and easy to be implemented. The second one takes advantage of linear algebra theories to reveal deeper relationships among the web pages and identify relevant pages more precisely and effectively. The corresponding page source construction for these two algorithms can prevent the results from being affected by malicious hyperlinks on the web. The experiment results show the feasibility and effectiveness of the algorithms. The research results could be used to enhance web search by caching the relevant pages for certain searched pages. For the purpose of clustering web pages to construct a community with its hierarchical cluster structures, we propose an innovative web page similarity measurement that incorporates hyperlink transitivity and page importance (weight).Based on this similarity measurement, two types of hierarchical web page clustering algorithms are proposed. The first one is the improvement of the conventional K-mean algorithms. It is effective in improving page clustering, but is sensitive to the predefined similarity thresholds for clustering. Another type is the matrix-based hierarchical algorithm. Two algorithms of this type are proposed in this work. One takes cluster-overlapping into consideration, another one does not. The matrix-based algorithms do not require predefined similarity thresholds for clustering, are independent of the order in which the pages are presented, and produce stable clustering results. The matrix-based algorithms exploit intrinsic relationships among web pages within a uniform matrix framework, avoid much influence of human interference in the clustering procedure, and are easy to be implemented for applications. The experiments show the effectiveness of the new similarity measurement and the proposed algorithms in web page clustering improvement. For applying above mathematical algorithms better in practice, we generalize the web page discovering as a special case of information retrieval and present a visualization system prototype, as well as technical details on visualization algorithm design, to support information retrieval based on linear algebra. The visualization algorithms could be smoothly applied to web applications. XML is a new standard for data representation and exchange on the Internet. In order to extend our research to cover this important web data, we propose an object representation model (ORM) for XML data. A set of transformation rules and algorithms are established to transform XML data (DTD and XML documents with DTD or without DTD) into this model. This model capsulizes elements of XML data and data manipulation methods. DTD-Tree is also defined to describe the logical structure of DTD. It also can be used as an application program interface (API) for processing DTD, such as transforming a DTD document into the ORM. With this data model, semantic meanings of the tags (elements) in XML data can be used for further research in XML data management and information retrieval, such as community construction for XML data.
APA, Harvard, Vancouver, ISO, and other styles
6

Myers, Paul Thomas. "The Cucamonga Middle School web page: Using parent input to redesign an existing school web page." CSUSB ScholarWorks, 2001. https://scholarworks.lib.csusb.edu/etd-project/2008.

Full text
Abstract:
This project focused on the enhancement of an existing school web page. Schools have jumped on the bandwagon in record numbers in the past couple of years publishing web pages about their schools. Cucamonga Middle School is no exception, having first published a web site in 1995. Utilizing information gathered from a survey, the author/webmaster created a redesign of the existing school webpage. The wants and needs of parents, as noted in the survey were the driving force behind the redesign.
APA, Harvard, Vancouver, ISO, and other styles
7

Metikurke, Seema Sreenivasamurthy. "Grid-Enabled Automatic Web Page Classification." Digital Archive @ GSU, 2006. http://digitalarchive.gsu.edu/cs_theses/23.

Full text
Abstract:
Much research has been conducted on the retrieval and classification of web-based information. A big challenge is the performance issue, especially for a classification algorithm returning results for a large set of data that is typical when accessing the Web. This thesis describes a grid-enabled approach for automatic web page classification. The basic approach is first described that uses a vector space model (VSM). An enhancement of the approach through the use of a genetic algorithm (GA) is then described. The enhanced approach can efficiently process candidate web pages from a number of web sites and classify them. A prototype is implemented and empirical studies are conducted. The contributions of this thesis are: 1) Application of grid computing to improve performance of both VSM and GA using VSM based web page classification; 2) Improvement of the VSM classification algorithm by applying GA that uniquely discovers a set of training web pages while also generating a near optimal parameter values set for VSM.
APA, Harvard, Vancouver, ISO, and other styles
8

Sanoja, Vargas Andrés. "Segmentation de pages web, évaluation et applications." Electronic Thesis or Diss., Paris 6, 2015. http://www.theses.fr/2015PA066004.

Full text
Abstract:
Les pages web sont devenues plus complexes que jamais, principalement parce qu'elles sont générées par des systèmes de gestion de contenu (CMS). Il est donc difficile de les analyser, c'est-à-dire d'identifier et classifier automatiquement les différents éléments qui les composent. La segmentation de pages web est une des solutions à ce problème. Elle consiste à décomposer une page web en segments, visuellement et sémantiquement cohérents, appelés blocs. La qualité d'une segmentation est mesurée par sa correction et sa généricité, c'est-à-dire sa capacité à traiter des pages web de différents types. Notre recherche se concentre sur l'amélioration de la segmentation et sur une mesure fiable et équitable de la qualité des segmenteurs. Nous proposons un modèle pour la segmentation ainsi que notre segmenteur Block-o-Matic (BoM). Nous définissons un modèle d'évaluation qui prend en compte le contenu ainsi que la géométrie des blocs pour mesurer la correction d'un segmenteur par rapport à une vérité de terrain. Ce modèle est générique, il permet de tester tout algorithme de segmentation et observer ses performances sur différents types de page. Nous l'avons testé sur quatre segmenteurs et quatre types de pages. Les résultats montrent que BOM surpasse ses concurrents en général et que la performance relative d'un segmenteur dépend du type de page. Enfin, nous présentons deux applications développées au dessus de BOM. Pagelyzer compare deux versions de pages web et décide si elles sont similaires ou pas. C'est la principale contribution de notre équipe au projet européen Scape (FP7-IP). Nous avons aussi développé un outil de migration de pages HTML4 vers le nouveau format HTML5
Web pages are becoming more complex than ever, as they are generated by Content Management Systems (CMS). Thus, analyzing them, i.e. automatically identifying and classifying different elements from Web pages, such as main content, menus, among others, becomes difficult. A solution to this issue is provided by Web page segmentation which refers to the process of dividing a Web page into visually and semantically coherent segments called blocks.The quality of a Web page segmenter is measured by its correctness and its genericity, i.e. the variety of Web page types it is able to segment. Our research focuses on enhancing this quality and measuring it in a fair and accurate way. We first propose a conceptual model for segmentation, as well as Block-o-Matic (BoM), our Web page segmenter. We propose an evaluation model that takes the content as well as the geometry of blocks into account in order to measure the correctness of a segmentation algorithm according to a predefined ground truth. The quality of four state of the art algorithms is experimentally tested on four types of pages. Our evaluation framework allows testing any segmenter, i.e. measuring their quality. The results show that BoM presents the best performance among the four segmentation algorithms tested, and also that the performance of segmenters depends on the type of page to segment.We present two applications of BoM. Pagelyzer uses BoM for comparing two Web pages versions and decides if they are similar or not. It is the main contribution of our team to the European project Scape (FP7-IP). We also developed a migration tool of Web pages from HTML4 format to HTML5 format in the context of Web archives
APA, Harvard, Vancouver, ISO, and other styles
9

Khalil, Faten. "Combining web data mining techniques for web page access prediction." University of Southern Queensland, Faculty of Sciences, 2008. http://eprints.usq.edu.au/archive/00004341/.

Full text
Abstract:
[Abstract]: Web page access prediction gained its importance from the ever increasing number of e-commerce Web information systems and e-businesses. Web page prediction, that involves personalising the Web users’ browsing experiences, assists Web masters in the improvement of the Web site structure and helps Web users in navigating the site and accessing the information they need. The most widely used approach for this purpose is the pattern discovery process of Web usage mining that entails many techniques like Markov model, association rules and clustering. Implementing pattern discovery techniques as such helps predict the next page tobe accessed by theWeb user based on the user’s previous browsing patterns. However, each of the aforementioned techniques has its own limitations, especiallywhen it comes to accuracy and space complexity. This dissertation achieves better accuracy as well as less state space complexity and rules generated by performingthe following combinations. First, we combine low-order Markov model and association rules. Markov model analysis are performed on the data sets. If the Markov model prediction results in a tie or no state, association rules are used for prediction. The outcome of this integration is better accuracy, less Markov model state space complexity and less number of generated rules than using each of the methods individually. Second, we integrate low-order Markov model and clustering. The data sets are clustered and Markov model analysis are performed oneach cluster instead of the whole data sets. The outcome of the integration is better accuracy than the first combination with less state space complexity than higherorder Markov model. The last integration model involves combining all three techniques together: clustering, association rules and low-order Markov model. The data sets are clustered and Markov model analysis are performed on each cluster. If the Markov model prediction results in close accuracies for the same item, association rules are used for prediction. This integration model achievesbetter Web page access prediction accuracy, less Markov model state space complexity and less number of rules generated than the previous two models.
APA, Harvard, Vancouver, ISO, and other styles
10

Eriksson, Tobias. "Automatic web page categorizationusing text classication methods." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-142424.

Full text
Abstract:
Over the last few years, the Web has virtually exploded with an enormous amount of web pages of dierent types of content. With the current size of Web, it has become cumbersome to try and manually index and categorize all of its content. Evidently, there is a need for automatic web page categorization. This study explores the use of automatic text classication methods for categorization of web pages. The results in this paper is shown to be comparable to results in other papers on automatic web page categorization, however not as good as results on pure text classication.
Under de senaste åren så har Webben exploderat i storlek, med miljontals webbsidor av vitt skilda innehåll. Den enorma storleken av Webben gör att det blir ohanterligt att manuellt indexera och kategorisera allt detta innehåll. Uppenbarligen behövs det automatiska metoder för att kategorisera webbsidor. Denna studie undersöker hur metoder för automatiskt textklassicering kan användas för kategorisering av hemsidor. De uppnådda resultatet i denna rapport är jämförbara med resultat i annan litteratur på samma område, men når ej upp till resultatet i studier på ren textklassicering.
APA, Harvard, Vancouver, ISO, and other styles
11

Das, Somak R. "Evaluation of QUIC on web page performance." Thesis, Massachusetts Institute of Technology, 2014. http://hdl.handle.net/1721.1/91444.

Full text
Abstract:
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2014.
19
Title as it appears in MIT commencement exercises program, June 6, 2014: Designing a better transport protocol for the web. Cataloged from PDF version of thesis.
Includes bibliographical references (pages 53-54).
This work presents the first study of a new protocol, QUIC, on Web page performance. Our experiments test the HTTP/1.1, SPDY, and QUIC multiplexing protocols on the Alexa U.S. Top 500 websites, across 100+ network configurations of bandwidth and round-trip time (both static links and cellular networks). To do so, we design and implement QuicShell, a tool for measuring QUIC's Web page performance accurately and reproducibly. Using QuicShell, we evaluate the strengths and weaknesses of QUIC. Due to its design of stream multiplexing over UDP, QUIC outperforms its predecessors over low-bandwidth links and high-delay links by 10 - 60%. It also helps Web pages with small objects and HTTPS-enabled Web pages. To improve QUIC's performance on cellular networks, we implement the Sprout-EWMA congestion control protocol and find that it improves QUIC's performance by > 10% on high-delay links.
by Somak R. Das.
M. Eng.
APA, Harvard, Vancouver, ISO, and other styles
12

Mereuta, Alina. "Smart web accessibility platform : dichromacy compensation and web page structure improvement." Thesis, Tours, 2014. http://www.theses.fr/2014TOUR4032/document.

Full text
Abstract:
Dans ces travaux, nous nous concentrons sur l’amélioration de l’accessibilité web pourdes utilisateurs handicapés visuels à travers d’outils s’intègrant dans la plateforme SmartWeb Accessibility Platform (SWAP). Après une synthèse sur l’accessibilité, un panoramade SWAP est présenté. Notre première contribution concerne la compensation de la pertede contraste textuel dans les pages web pour les daltoniens tout en maintenant les intentionsde l’auteur retranscrites par les couleurs. Ce problème est modélisé sous la forme d’unproblème de minimisation d’une fonction numérique dépendante des couleurs initiales, desrelations entre les couleurs de la page et des couleurs modifiées. L’intérêt et l’efficacité detrois méthodes d’optimisation (mass-spring system, CMA-ES, API) ont été évaluées surdeux jeux de données (réel, artificiel). Notre deuxième contribution cherche à améliorer lastructure de pages web pour des utilisateurs de lecteurs d’écran afin de compenser la perted’information due à la linéarisation du contenu. Grâce à l’utilisation d’heuristiques et detechnique d’apprentissage, les zones caractéristiques de la page sont identifiées. La pageest complétée en utilisant des liens supplémentaires et les marqueurs ARIA afin de permettreune meilleure identification des zones par les lecteurs d’écran. Nos expérimentationsmontrent l’efficacité de notre proposition
This thesis works are focused on enhancing web accessibility for users with visual disabilities using tools integrated within the SmartWeb Accessibility Platform (SWAP). After a synthesis on accessibility, SWAP is presented. Our first contribution consists in reducing the contrast loss for textual information in web pages for dichromat users while maintaining the author’s intentions conveyed by colors. The contrast compensation problem is reduced at minimizing a fitness function which depends on the original colors and the relationships between them. The interest and efficiency of three methods (mass-spring system, CMA-ES, API) are assessed on two datasets (real and artificial). The second contribution focuses on enhancing web page structure for screen reader users in order to overcome the effect of contents’linearization. Using heuristics and machine learning techniques, the main zones of the page are identified. The page structure can be enhanced using ARIA statements and access links to improve zone identification by screen readers
APA, Harvard, Vancouver, ISO, and other styles
13

Xu, Jingqian. "Full similarity-based page ranking." Diss., Columbia, Mo. : University of Missouri-Columbia, 2008. http://hdl.handle.net/10355/5773.

Full text
Abstract:
Thesis (M.S.)--University of Missouri-Columbia, 2008.
The entire dissertation/thesis text is included in the research.pdf file; the official abstract appears in the short.pdf file (which also appears in the research.pdf); a non-technical general description, or public abstract, appears in the public.pdf file. Title from title screen of research.pdf file (viewed on August 19, 2009) Includes bibliographical references.
APA, Harvard, Vancouver, ISO, and other styles
14

Mortazavi-Asl, Behzad. "Discovering and mining user Web-page traversal patterns." Thesis, National Library of Canada = Bibliothèque nationale du Canada, 2001. http://www.collectionscanada.ca/obj/s4/f2/dsk3/ftp04/MQ61594.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
15

RODRIGUES, THORAN ARAGUEZ. "A COMPARATIVE STUDY OF WEB PAGE CLASSIFICATION STRATEGIES." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2009. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=13890@1.

Full text
Abstract:
A quantidade de informações na Internet aumenta a cada dia. Embora esta proliferação aumente as chances de que o tema sendo buscado por um usuário esteja presente na rede, ela também torna encontrar a informação desejada mais difícil. A classificação automática de páginas é, portanto, uma importante ferramenta na organização de conteúdo da Web, com aplicações específicas na melhoria dos resultados retornados por máquinas de busca. Nesta dissertação foi realizado um estudo comparativo de diferentes conjuntos de atributos e métodos de classificação aplicados ao problema da classificação funcional de páginas web, com foco em 4 classes: Blogs, Blog Posts, Portais de Notícias e Notícias. Ao longo dos experimentos, foi possível constatar que a melhor abordagem para esta tarefa é a utilização de atributos tanto da estrutura quanto do texto das páginas. Foi apresentada também uma estratégia nova de construção de conjuntos de atributos de texto, que leva em consideração os diferentes estilos de escrita das classes de páginas.
The amount of information on the Internet increases every day. Even though this proliferation increases the chances that the subject being searched for by an user is on the Web, it also makes finding the desired information much harder. The automated classification of pages is, therefore, an important tool for organizing Web content, with specific applications on the improvement of results displayed by search engines. In this dissertation, a comparative study of different attribute sets and classification methods for the functional classification of web pages was made, focusing on 4 classes: Blogs, Blog Posts, News Portals and News. Throughout the experiments, it became evident the best approach for this task is to employ attributes that come both from the structure and the text of the web pages. We also presented a new strategy for extracting and building text attribute sets, that takes into account the different writing styles for each page class.
APA, Harvard, Vancouver, ISO, and other styles
16

Derryberry, Jonathan C. (Jonathan Carlyle) 1979. "Creating a web page recommendation system for Haystack." Thesis, Massachusetts Institute of Technology, 2003. http://hdl.handle.net/1721.1/28472.

Full text
Abstract:
Thesis (M. Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2003.
Includes bibliographical references (p. 105).
The driving goal of this thesis was to create a web page recommendation system for Haystack, capable of tracking a user's browsing behavior and suggesting new, interesting web pages to read based on the past behavior. However, during the course of this thesis, 3 salient subgoals were met. First, Haystack's learning framework was unified so that, for example, different types of binary classifiers could be used with black box access under a single interface, regardless of whether they were text learning algorithms or image classifiers. Second, a tree learning module, capable of using hierarchical descriptions of objects and their labels to classify new objects, was designed and implemented. Third, Haystack's learning framework and existing user history faculties were leveraged to create a web page recommendation system that uses the history of a user's visits to web pages to produce recommendations of unvisited links from user-specified web pages. Testing of the recommendation system suggests that using tree learners with both the URL and tabular location of a web page's link as taxonomic descriptions yields a recommender that significantly outperforms traditional, text-based systems.
by Jonathan C. Derryberry.
M.Eng.
APA, Harvard, Vancouver, ISO, and other styles
17

Yu, Chen-Hsiang. "Web page enhancement on desktop and mobile browsers." Thesis, Massachusetts Institute of Technology, 2012. http://hdl.handle.net/1721.1/79216.

Full text
Abstract:
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, February 2013.
"February 2013." Cataloged from PDF version of thesis.
Includes bibliographical references (p. 154-165).
The Web is a convenient platform to deliver information, but reading web pages is not as easy as it was in 1990s. This thesis focuses on investigating techniques to enhance web pages on desktop and mobile browsers for two specific populations: non-native English readers and mobile users. There are three issues addressed in this thesis: web page readability, web page skimmability and continuous reading support on mobile devices. On today's primarily English-language Web, non-native readers encounter some problems, even if they have some fluency in English. This thesis focuses on content presentation and proposes a new transformation method, Jenga Format, to enhance web page readability. A user study with 30 non-native users showed that Jenga transformation not only improved reading comprehension, but also made the web page reading easier. On the other hand, readability research has found that average reading times for non-native readers has remained the same or even worse. This thesis studies this issue and proposes Froggy GX (Generation neXt) to improve reading under time constraints. A user study with 20 non-native users showed that Froggy GX not only enhanced reading comprehension under time constraints, but also provided higher user satisfaction than reading unaided. When using the Web on mobile devices, the reading situation becomes challenging. Even worse, context switches, such as from walking to sitting, static standing, or hands-free situations like driving, happen in reading in on-the-go situations, but this scenario was not adequately addressed in previous studies. This thesis investigates this scenario and proposes a new mobile browser, Read4Me, to support continuous reading on a mobile device. A user study with 10 mobile users showed that auto-switching not only provided significantly fewer dangerous encounters than visual-reading, but also provided the best reading experience.
by Chen-Hsiang Yu.
Ph.D.
APA, Harvard, Vancouver, ISO, and other styles
18

Andr, Ondřej. "Srovnání on-page SEO faktorů pro mobilní web." Master's thesis, Vysoká škola ekonomická v Praze, 2015. http://www.nusl.cz/ntk/nusl-204028.

Full text
Abstract:
The thesis deals with a topic of SEO onpage signals, which are important for search engines because of sorting pages in a search engine result page. It focuses on importance of these signals for mobile SERP. Main goals of this study are to describe current recommendations for SEO on-page factors for mobile web and experimentally test real importance of these signals. Based on the results I composed an optimal set of factors with the most benefit for SEO. Theoretical part of the study summarizes basic facts about mobile searching, describes specific mobile users behaviour and describes current recommendations for mobile web onpage optimizing from Google and Seznam.cz. In practical part there is a comparative study of chosen on-page signals. For its needs I had to create few one page static websites. Each one has been optimized for on factor. All websites focused on the same very specific topic to ensure the same initial conditions. By keywords rank tracking in a SERP I was able to determine which signal is more important than others for search engines. The study results contribute to actual evaluation of each on-page signals importance for mobile website. The study could be beneficial for smaller companys websites, which need to get more visible on the net. They are able to optimize their costs by choosing the right set of on-page factors.
APA, Harvard, Vancouver, ISO, and other styles
19

Jackson, Lance Douglas Smith Jon M. 1959. "Introduction to the Internet and Web page design." [Cedar City, Utah : Southern Utah University], 2009. http://unicorn.li.suu.edu/ScholarArchive/Communication/JacksonLanceD/IntrototheInternet&WebPageDesign.pdf.

Full text
Abstract:
Thesis (M.S.)--Southern Utah University, 2009.
A workbook CD accompanies this text. For more information contact the author, Lance Jackson, Southern Utah University, 351 W. University Blvd., Cedar city, UT 84720. E-mail: jackson@suu.edu. Telephone: (435) 586-7867. Title from PDF title page. "April 2009." "In partial fulfillment of the requirements for the degree [of] Master of Arts in Professional Communication." "A project presented to the faculty of the Communication Department at Southern Utah University." Dr. Jon Smith, Project Supervisor. Includes bibliographical references (p. 14, 33, 49, 69, 85, 104, 135, 155, 174).
APA, Harvard, Vancouver, ISO, and other styles
20

Goodrich, Brian S. "Extending Web Application Development to the User-Editable Space." Diss., CLICK HERE for online access, 2008. http://contentdm.lib.byu.edu/ETD/image/etd2289.pdf.

Full text
APA, Harvard, Vancouver, ISO, and other styles
21

Xiao, Xiangye. "Slicing*-tree based Web page transformation for small displays /." View abstract or full-text, 2005. http://library.ust.hk/cgi/db/thesis.pl?COMP%202005%20XIAO.

Full text
APA, Harvard, Vancouver, ISO, and other styles
22

Lu, Zhengyang. "Web Page Classification Using Features from Titles and Snippets." Thesis, Université d'Ottawa / University of Ottawa, 2015. http://hdl.handle.net/10393/33177.

Full text
Abstract:
Nowadays, when a keyword is provided, a search engine can return a large number of web pages, which makes it difficult for people to find the right information. Web page classification is a technology that can help us to make a relevant and quick selection of information that we are looking for. Moreover, web page classification is important for companies that provide marketing and analytics platforms, because it can help them to build a healthy mix of listings on search engines and large directories. This will provide more insight into the distribution of the types of web pages their local business listings are found on, and finally will help marketers to make better-informed decisions about marketing campaigns and strategies. In this thesis we perform a literature review that introduces web page classification, feature selection and feature extraction. The literature review also includes a comparison of three commonly used classification algorithms and a description of metrics for performance evaluation. The findings in the literature enable us to extend existing classification techniques, methods and algorithms to address a new web page classification problem faced by our industrial partner SweetIQ (a company that provides location-based marketing services and an analytics platform). We develop a classification method based on SweetIQ's data and business needs. Our method includes typical feature selection and feature extraction methods, but the features we use in this thesis are largely different from traditional ones used in the literature. We test selected features and find that the text extracted from the title and snippet of a web page can help a classifier to achieve good performance. Our classification method does not require the full content of a web page. Thus, it is fast and saves a lot of space.
APA, Harvard, Vancouver, ISO, and other styles
23

Salameh, Lynne. "Towards faster web page loads over multiple network paths." Thesis, University College London (University of London), 2018. http://discovery.ucl.ac.uk/10046643/.

Full text
Abstract:
The rising popularity of mobile devices as the main way people access the web has fuelled a corresponding need for faster web downloads on these devices. Emerging web protocols like HTTP/2 and QUIC employ several features that minimise page load times, but fail to take advantage of the availability of at least two interfaces on today's mobile devices. On the other hand, this spread of devices with access to multiple paths has prompted the design of Multipath TCP, a transport protocol that pools bandwidth across these paths. Although MPTCP was originally evaluated for bandwidth limited bulk transfers, in this work, we determine whether using MPTCP can reduce web page load times, which are often latency bound. To investigate the behaviour of web browsing over MPTCP, we instrumented the Chrome web browser's retrieval of 300 popular web sites in sufficient detail, and computed their dependency graph structure. Furthermore, we implemented PCP, an emulation framework that uses these dependency graphs to ask "what-if" questions about the interactions between a wide range of web site designs, varied network conditions, and different web and transport protocols. Using PCP, we first confirm previous results with respect to the improvements HTTP/2 offers over HTTP/1.1. One obstacle, though, is that many web sites have been sharded to improve performance with HTTP/1.1, spreading their content across multiple subdomains. We therefore examine whether the advice to unshard these domains is beneficial. We find that unsharding is generally advantageous, but is not consistently so. Finally, we examine the behaviour of HTTP/2 over MPTCP. We find that MPTCP can improve web page load times under some regimes; in other cases, using regular TCP on the "best" path is more advantageous. We present enhancements to multipath web browsing that allow it to perform as well as or better than regular TCP on the best path.
APA, Harvard, Vancouver, ISO, and other styles
24

Santos, Aécio Solano Rodrigues. "Learning to schedule web page updates using genetic programming." Universidade Federal de Minas Gerais, 2013. http://hdl.handle.net/1843/ESBF-97GJSQ.

Full text
Abstract:
One of the main challenges endured when designing a scheduling policy regarding freshness is to estimate the likelihood of a previously crawled web page being modified on the web, so that the scheduler can use this estimation to determine the order in which those pages should be visited. A good estimation of which pages have more chance of being modified allows the system to reduce the overall cost of monitoring its crawled web pages for keeping updated versions. In this work we present a novel approach that uses machine learning to generate score functions that produce accurate rankings of pages regarding their probability of being modified on the Web when compared to their previously crawled versions. We propose a flexible framework that uses Genetic Programming to evolve score functions to estimate the likelihood that a web page has been modified. We present a thorough experimental evaluation of the benefits of using the framework over five state-of-the-art baselines. Considering the Change Ratio metric, the values produced by our best evolved function show an improvement from 0.52 to 0.71 on average over the baselines.
Um dos principais desafios enfrentados durante o desenvolvimento de políticas de escalonamento para atualizações de páginas web é estimar a probabilidade de uma página que já foi coletada previamente ser modificada na Web. Esta informação pode ser usada pelo escalonador de um coletor de páginas web para determinar a ordem na qual as páginas devem ser recoletadas, permitindo ao sistema reduzir o custo total de monitoramento das páginas coletadas para mantê-las atualizadas. Nesta dissertação é apresentada uma nova abordagem que usa aprendizado de máquina para gerar funções de score que produzem listas ordenadas de páginas com relação a probabilidade de terem sido modificadas na Web quando comparado com a última versão coletada. É proposto um arcabouço flexível que usa Programação Genética para evoluir funções que estimam a probabilidade de a página ter sido modificada. É apresentado ainda uma avaliação experimental dos benefícios de usar o arcabouço proposto em relação a cinco abordagens estado-da-arte. Considerando a métrica Change Ratio, os valores produzidos pela melhor função gerada pelo arcabouço proposto mostram uma melhora de 0.52 para 0.71, em média, em relação aos baselines.
APA, Harvard, Vancouver, ISO, and other styles
25

Knowlton, Corey Lamoin. "Web page design class curriculum for the secondary level." CSUSB ScholarWorks, 2002. https://scholarworks.lib.csusb.edu/etd-project/2108.

Full text
APA, Harvard, Vancouver, ISO, and other styles
26

Namoune, Abdallah. "Investigating visual attention on the web and the development of a web page analyser." Thesis, University of Manchester, 2008. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.500473.

Full text
APA, Harvard, Vancouver, ISO, and other styles
27

Annadi, Ramakanth Reddy. "Adapting Web Page Tables on Mobile Web Browsers: Results from Two Controlled Empirical Studies." Thesis, North Dakota State University, 2014. https://hdl.handle.net/10365/27281.

Full text
Abstract:
Displaying web page content in mobile screens is a challenging task and users often face difficulty retrieving the relevant data. It can force them to adopt a time-consuming hunt-and-peck strategy. Application of design principles can improve the view of the webpage data content and reduce the time consumption in viewing it. This is especially true with HTML tabular data content. This thesis discusses the background and applications of the gestalt design principle techniques to HTML tabular data content. An empirical study was performed to investigate the usability of two types of the adaptive styles namely, single and multi-layout. This thesis also compared the adaptive styles that use gestalt principles with simple HTML tabular data on mobile screens. A controlled study which involved university students was performed showed that the adaptive layout styles improved the efficiency of finding information in the website by gestalt principles usage and eliminating horizontal scroll.
APA, Harvard, Vancouver, ISO, and other styles
28

Costa, José Henrique Calenzo. "Filtered-page ranking." reponame:Repositório Institucional da UFSC, 2016. https://repositorio.ufsc.br/xmlui/handle/123456789/167840.

Full text
Abstract:
Dissertação (mestrado) - Universidade Federal de Santa Catarina, Centro Tecnológico, Programa de Pós-Graduação em Ciência da Computação, Florianópolis, 2016.
Made available in DSpace on 2016-09-20T04:25:42Z (GMT). No. of bitstreams: 1 341906.pdf: 4935734 bytes, checksum: 5630ca8c10871314b7f54120d18ae335 (MD5) Previous issue date: 2016
Algoritmos de ranking de páginas Web podem ser criados usando técnicas baseadas em elementos estruturais da página Web, em segmentação da página ou na busca personalizada. Esta pesquisa aborda um método de ranking de documentos previamente filtrados, que segmenta a página Web em blocos de três categorias para delas eliminar conteúdo irrelevante. O método de ranking proposto, chamado Filtered-Page Ranking (FPR), consta de duas etapas principais: (i) segmentação da página web e eliminação de conteúdo irrelevante e (ii) ranking de páginas Web. O foco da extração de conteúdo irrelevante é eliminar conteúdos não relacionados à consulta do usuário, através do algoritmo proposto Query-Based Blocks Mining (QBM), para que o ranking considere somente conteúdo relevante. O foco da etapa de ranking é calcular quão relevante cada página Web é para determinada consulta, usando critérios considerados em estudos de recuperação da informação. Com a presente pesquisa pretende-se demonstrar que o QBM extrai eficientemente o conteúdo irrelevante e que os critérios utilizados para calcular quão próximo uma página Web é da consulta são relevantes, produzindo uma média de resultados de ranking de páginas Web de qualidade melhor que a do clássico modelo vetorial.

Abstract : Web page ranking algorithms can be created using content-based, structure-based or user search-based techniques. This research addresses an user search-based approach applied over previously filtered documents ranking, which relies in a segmentation process to extract irrelevante content from documents before ranking. The process splits the document into three categories of blocks in order to fragment the document and eliminate irrelevante content. The ranking method, called Page Filtered Ranking, has two main steps: (i) irrelevante content extraction; and (ii) document ranking. The focus of the extraction step is to eliminate irrelevante content from the document, by means of the Query-Based Blocks Mining algorithm, creating a tree that is evaluated in the ranking process. During the ranking step, the focus is to calculate the relevance of each document for a given query, using criteria that give importance to specific parts of the document and to the highlighted features of some HTML elements. Our proposal is compared to two baselines: the classic vectorial model, and the CETR noise removal algorithm, and the results demonstrate that our irrelevante content removal algorithm improves the results and our relevance criteria are relevant to the process.
APA, Harvard, Vancouver, ISO, and other styles
29

Visser, Eugene Bourbon. "Fusing website usability variables and on-page search engine optimisation elements." Thesis, Cape Peninsula University of Technology, 2011. http://hdl.handle.net/20.500.11838/1407.

Full text
Abstract:
Thesis (DTech (Information Technology))Cape Peninsula University of Technology, 2011
It was concluded in the literature review that small- to medium-sized enterprises (SMME) should prioritise utilising the websites on the Internet, as it provides a low cost infrastructure, unlocking opportunities and allowing small- to medium-sized enterprises to market to the international customer, promoting business activities in a low-risk environment. However, visitors do not know that they do not know, meaning a need for facilitation exists between the Internet user in terms of the information required and the information available on the Internet. Search engines (governed by their organic ranking algorithms) were created for this very purpose, to facilitate users in finding relevant information on the Internet in the shortest time possible. Search engines interpret and evaluate any given indexed web page from a targeted keywords perspective, indicating that web pages must be optimised from a search engine perspective. However, the elements search engines perceive to be important may not always be aligned with what website visitors perceive to be important. Anything on the web page that may remotely impede the visitors’ experience could be detrimental as alternative website options are but a click away. An example would be the excessive use of content on a given web page. The search engine may find the excessive content useful as it may provide contextual interpretation of the web page. However, the excessive content may impede a visitor’s website interaction as it is estimated that the average visitors will often view a web page for 45-60 seconds and read a maximum of 200 words only. During the process of identifying the contradictory search engine optimisation (SEO) elements and website usability (WU) attributes, three journal articles were written, with two journal articles following their own research methodologies and the third journal article utilising all the research results in order to create the fused SEO and WU model. Journal Article 1: Two websites were used as part of the experiment: • Control Website (CW): http://www.copywriters.co.za • Experimental Website (EW): http://www.copywriters.co.za/ppc/. The CW is an existing website with no special emphasis applied to SEO and/or WU. The EW was developed by implementing the WU attributes and ignoring all contradictory SEO elements. In order to ensure integrity of the experiment, search engines were denied access to the EW. The traffic sources for the CW were search engines (organic) traffic, as well as direct and referrer traffic.
APA, Harvard, Vancouver, ISO, and other styles
30

Sundin, Albin. "Word Space Models for Web User Clustering and Page Prefetching." Thesis, Linköpings universitet, Institutionen för datavetenskap, 2012. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-82012.

Full text
Abstract:
This study evaluates methods for clustering web users via vector space models, for the purpose of web page prefetching for possible applications of server optimization. An experiment using Latent Semantic Analysis (LSA) is deployed to investigate whether LSA can reproduce the encouraging results obtained from previous research with Random Indexing (RI) and a chaos based optimization algorithm (CAS-C). This is not only motivated by LSA being yet another vector space model, but also by a study indicating LSA to outperform RI in a task similar to the web user clustering and prefetching task. The prefetching task was used to verify the applicability of LSA, where both RI and CAS-C have shown promising results. The original data set from the RI web user clustering and prefetching task was modeled using weighted (tf-idf) LSA. Clusters were defined using a common clustering algorithm (k-means). The least scattered cluster configuration for the model was identified by combining an internal validity measure (SSE) and a relative criterion validity measure (SD index). The assumed optimal cluster configuration was used for the web page prefetching task.   Precision and recall of the LSA based method is found to be on par with RI and CAS-C, in as much that it solves the web user clustering and web task with similar characteristics as unweighted RI. The hypothesized inherent gains to precision and recall by using LSA was neither confirmed nor conclusively disproved. The effects of different weighting functions for RI are discussed and a number of methodological factors are identified for further research concerning LSA based clustering and prefetching.
APA, Harvard, Vancouver, ISO, and other styles
31

Netravali, Ravi Arun. "Understanding and improving Web page load times on modern networks." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/97765.

Full text
Abstract:
Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Computer Science, 2015.
This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 77-80).
This thesis first presents a measurement toolkit, Mahimahi, that records websites and replays them under emulated network conditions. Mahimahi improves on prior record-and-replay frameworks by emulating the multi-origin nature of Web pages, isolating its network traffic, and enabling evaluations of a larger set of target applications beyond browsers. Using Mahimahi, we perform a case study comparing current multiplexing protocols, HTTP/1.1 and SPDY, and a protocol in development, QUIC, to a hypothetical optimal protocol. We find that all three protocols are significantly suboptimal and their gaps from the optimal only increase with higher link speeds and RTTs. The reason for these trends is the same for each protocol: inherent source-level dependencies between objects on a Web page and browser limits on the number of parallel flows lead to serialized HTTP requests and prevent links from being fully occupied. To mitigate the effect of these dependencies, we built Cumulus, a user-deployable combination of a content-distribution network and a cloud browser that improves page load times when the user is at a significant delay from a Web page's servers. Cumulus contains a "Mini-CDN"-a transparent proxy running on the user's machine-and a "Puppet": a headless browser run by the user on a well-connected public cloud. When the user loads a Web page, the Mini-CDN forwards the user's request to the Puppet, which loads the entire page and pushes all of the page's objects to the Mini-CDN, which caches them locally. Cumulus benefits from the finding that dependency resolution, the process of learning which objects make up a Web page, accounts for a considerable amount of user-perceived wait time. By moving this task to the Puppet, Cumulus can accelerate page loads without modifying existing Web browsers or servers. We find that on cellular, in-flight Wi-Fi, and transcontinental networks, Cumulus accelerated the page loads of Google's Chrome browser by 1.13-2.36×. Performance was 1.19-2.13× faster than Opera Turbo, and 0.99-1.66× faster than Chrome with Google's Data Compression Proxy.
by Ravi Arun Netravali.
S.M.
APA, Harvard, Vancouver, ISO, and other styles
32

Vishwasrao, Saket Dilip. "Performance Evaluation of Web Archiving Through In-Memory Page Cache." Thesis, Virginia Tech, 2017. http://hdl.handle.net/10919/78252.

Full text
Abstract:
This study proposes and evaluates a new method for Web archiving. We leverage the caching infrastructure in Web servers for archiving. Redis is used as the page cache and its persistence mechanism is exploited for archiving. We experimentally evaluate the performance of our archival technique using the Greek version of Wikipedia deployed on Amazon cloud infrastructure. We show that there is a slight increase in latencies of the rendered pages due to archiving. Though the server performance is comparable at larger page cache sizes, the maximum throughput the server can handle decreases significantly at lower cache sizes due to more disk write operations as a result of archiving. Since pages are dynamically rendered and the technology stack of Wikipedia is extensively used in a number of Web applications, our results should have broad impact.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
33

Veis, Richard. "Web page analysis of selected airlines on the czech market." Master's thesis, Vysoká škola ekonomická v Praze, 2010. http://www.nusl.cz/ntk/nusl-16701.

Full text
Abstract:
This thesis is devoted to the web-page analysis and to the analysis of the accessability of financial records, that are usually part of annual reports. The first part is devoted to the theoretical basis and standards, where is stated how an airline corporation should act. In the second part, the theoretical basis is applied to concrete airlines, that are significant on the personal air-transportation market. The individual rating is stated in the conclusion. There are also concrete recommendations that should be used for raising the competitiveness of the given subject.
APA, Harvard, Vancouver, ISO, and other styles
34

Williams, Rewa Colette. "Patterns Of 4th Graders' Literacy Events In Web Page Development." [Tampa, Fla.] : University of South Florida, 2003. http://purl.fcla.edu/fcla/etd/SFE0000203.

Full text
APA, Harvard, Vancouver, ISO, and other styles
35

Wei, Chenjie. "Using Automated Extraction of the Page Component Hierarchy to Customize and Adapt Web Pages to Mobile Devices." The Ohio State University, 2012. http://rave.ohiolink.edu/etdc/view?acc_num=osu1338348757.

Full text
APA, Harvard, Vancouver, ISO, and other styles
36

Grace, Phillip Eulon. "Full-page versus partial-page screen designs in web-based training : their effects on learner satisfaction and performance." [Tampa, Fla] : University of South Florida, 2005. http://purl.fcla.edu/usf/dc/et/SFE0001520.

Full text
APA, Harvard, Vancouver, ISO, and other styles
37

Tian, Ran. "Examining the Complexity of Popular Websites." Thesis, University of Oregon, 2015. http://hdl.handle.net/1794/19347.

Full text
Abstract:
A significant fraction of today's Internet traffic is associated with popular web sites such as YouTube, Netflix or Facebook. In recent years, major Internet websites have become more complex as they incorporate a larger number and more diverse types of objects (e.g. video, audio, code) along with more elaborate ways from multiple servers. These not only affect the loading time of pages but also determine the pattern of resulting traffic on the Internet. In this thesis, we characterize the complexity of major Internet websites through large-scale measurement and analysis. We identify thousands of the most popular Internet websites from multiple locations and characterize their complexities. We examine the effect of the relative popularity ranking and business type of the complexity of websites. Finally we compare and contrast our results with a similar study conducted 4 years earlier and report on the observed changes in different aspects.
APA, Harvard, Vancouver, ISO, and other styles
38

Siva, Sahithi Pokala. "Design and delivery : functional colour web pages." Thesis, University of Liverpool, 2001. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.343620.

Full text
APA, Harvard, Vancouver, ISO, and other styles
39

Abreu, Luís Pedro Borges. "Morphing Web Pages to Preclude Web Page Tampering Threats." Master's thesis, 2016. https://repositorio-aberto.up.pt/handle/10216/90184.

Full text
Abstract:
O número de utilizadores da Internet continua a aumentar todos os anos e a Internet é cada vez mais uma ferramenta diária na vida de cada indivíduo, utilizada como instrumento de trabalho ou de entretenimento. Contudo, ao navegar na Internet, os utilizadores tornam-se possíveis alvos de ataques informáticos um vez que efetuam transações de dados, muitas vezes privados e sensíveis, com servidores remotos.Entre os diferentes ataques informáticos existentes, destaca-se o ataque MitB que deu origem ao tema desta dissertação. Os ataques MitB são realizados com recurso a Malware instalado e em execução nos computadores dos utilizadores, que tem acesso às informações das janelas dos navegadores de Internet - por exemplo através de bibliotecas de funções do sistema operativo ou até recorrendo a extensões dos navegadores de Internet. Estes ataques utilizam âncoras do DOM para identificar as secções de uma página web onde pretendem atacar - recolhendo dados ou modificando a própria página. O resultado do ataque será diretamente influenciado pela capacidade do Malware em identificar os pontos de ataque numa determinada página web.O Polimorfismo é um conceito geral que pode ser aplicado a páginas web como uma ferramenta para neutralizar e derrotar este tipo de ataques informáticos, tal como foi documentado pela empresa Shape Security, Inc. em 2014. Aplicando técnicas de polimorfismo a páginas web, as respostas de um servidor serão textualmente diferentes entre si, mas o resultado visual apresentado ao utilizador será sempre o mesmo. Concretamente, os valores dos atributos estáticos e a estrutura dos documentos HTML poderão ser modificados no servidor, criando assim versões polimorfas de uma página web. Estas transformações podem ser realizadas em tempo real no servidor ou pré-calculadas. Desta forma, nunca dois documentos HTML serão textualmente iguais, tornando as páginas em alvos em movimento, dificultando os ataques MitB. Este nível de proteção é necessário uma vez que todas as alterações da página realizadas pelo atacante são locais e portanto difíceis de detectar pelas estruturas de segurança e controlo implementadas nos servidores dos fornecedores dos serviços.Neste dissertação, pretende-se desenvolver uma ferramenta baseada em polimorfismo para proteger páginas web e os seus utilizadores dos ataques MitB que recorrem a âncoras do DOM. Esta ferramenta será avaliada em precisão e eficiência. A precisão será avaliada comparando a lista de erros e de avisos gerada pela navegador de Internet para a página original e as várias versões polimorfas produzidas, enquanto que a eficiência será calculada recorrendo a tentativas automáticas de alterações indevidas das páginas protegidas com a ferramenta desenvolvida.
The number of Internet users keeps growing every year. Moreover, the Internet is becoming a daily tool, which impacts the individual's lives used either as a work tool or for entertainment purposes. However, by using it, people become possible targets for cyber attacks as they keep exchanging data, sometimes sensitive and private data, with remote servers.Among all the different attacks types, MitB is the reason behind the genesis of this thesis subject. MitB attacks are performed by a computer program running on user's computer that is commonly known as Malware, which has access to what happens inside a browser window. It can be a system library or even a browser extension programmed to, automatically, misrepresent the source code of the client-side server response, and other information stored in user's browsers. They rely on markup and DOM anchors to identify sections of a web page to attack. The end result of an attack will be dictated by the malware's ability to successfully identify the right location on the web page to perform the attack.Polymorphism is a broad concept that can be applied to web pages as a tool to both neutralize and defeat such kind of attacks, as documented by Shape Security, Inc. in 2014. Applying polymorphic techniques to web pages, the server response will be textually different between requests, but the visual display to the user will always be the same. That is, the values of static attributes and the structure of HTML documents may be modified on the server immediately before responses are sent off, creating a polymorphic version of the web page, or by pre-building this new versions on the server to decrease the real time computational costs. Therefore, no two HTML documents will be textually the same, turning web pages in somehow a moving target against MitB attacks. This level of protection is necessary since all changes are made locally, client side, making their detection difficult by control and security structures implemented on the service provider's servers.In this thesis, we aim to develop a tool based on polymorphism to protect web pages and users from MitB attacks based on markup and DOM anchors. This tool will be evaluated by accuracy and efficiency. The first metric will be evaluated by recording and comparing the list of errors and warnings generated by original web pages and by their polymorphic versions created with our tool. The efficiency will be evaluated by running automated attempts for tampering web pages protected by our tool.
APA, Harvard, Vancouver, ISO, and other styles
40

Abreu, Luís Pedro Borges. "Morphing Web Pages to Preclude Web Page Tampering Threats." Dissertação, 2016. https://repositorio-aberto.up.pt/handle/10216/90184.

Full text
Abstract:
O número de utilizadores da Internet continua a aumentar todos os anos e a Internet é cada vez mais uma ferramenta diária na vida de cada indivíduo, utilizada como instrumento de trabalho ou de entretenimento. Contudo, ao navegar na Internet, os utilizadores tornam-se possíveis alvos de ataques informáticos um vez que efetuam transações de dados, muitas vezes privados e sensíveis, com servidores remotos.Entre os diferentes ataques informáticos existentes, destaca-se o ataque MitB que deu origem ao tema desta dissertação. Os ataques MitB são realizados com recurso a Malware instalado e em execução nos computadores dos utilizadores, que tem acesso às informações das janelas dos navegadores de Internet - por exemplo através de bibliotecas de funções do sistema operativo ou até recorrendo a extensões dos navegadores de Internet. Estes ataques utilizam âncoras do DOM para identificar as secções de uma página web onde pretendem atacar - recolhendo dados ou modificando a própria página. O resultado do ataque será diretamente influenciado pela capacidade do Malware em identificar os pontos de ataque numa determinada página web.O Polimorfismo é um conceito geral que pode ser aplicado a páginas web como uma ferramenta para neutralizar e derrotar este tipo de ataques informáticos, tal como foi documentado pela empresa Shape Security, Inc. em 2014. Aplicando técnicas de polimorfismo a páginas web, as respostas de um servidor serão textualmente diferentes entre si, mas o resultado visual apresentado ao utilizador será sempre o mesmo. Concretamente, os valores dos atributos estáticos e a estrutura dos documentos HTML poderão ser modificados no servidor, criando assim versões polimorfas de uma página web. Estas transformações podem ser realizadas em tempo real no servidor ou pré-calculadas. Desta forma, nunca dois documentos HTML serão textualmente iguais, tornando as páginas em alvos em movimento, dificultando os ataques MitB. Este nível de proteção é necessário uma vez que todas as alterações da página realizadas pelo atacante são locais e portanto difíceis de detectar pelas estruturas de segurança e controlo implementadas nos servidores dos fornecedores dos serviços.Neste dissertação, pretende-se desenvolver uma ferramenta baseada em polimorfismo para proteger páginas web e os seus utilizadores dos ataques MitB que recorrem a âncoras do DOM. Esta ferramenta será avaliada em precisão e eficiência. A precisão será avaliada comparando a lista de erros e de avisos gerada pela navegador de Internet para a página original e as várias versões polimorfas produzidas, enquanto que a eficiência será calculada recorrendo a tentativas automáticas de alterações indevidas das páginas protegidas com a ferramenta desenvolvida.
The number of Internet users keeps growing every year. Moreover, the Internet is becoming a daily tool, which impacts the individual's lives used either as a work tool or for entertainment purposes. However, by using it, people become possible targets for cyber attacks as they keep exchanging data, sometimes sensitive and private data, with remote servers.Among all the different attacks types, MitB is the reason behind the genesis of this thesis subject. MitB attacks are performed by a computer program running on user's computer that is commonly known as Malware, which has access to what happens inside a browser window. It can be a system library or even a browser extension programmed to, automatically, misrepresent the source code of the client-side server response, and other information stored in user's browsers. They rely on markup and DOM anchors to identify sections of a web page to attack. The end result of an attack will be dictated by the malware's ability to successfully identify the right location on the web page to perform the attack.Polymorphism is a broad concept that can be applied to web pages as a tool to both neutralize and defeat such kind of attacks, as documented by Shape Security, Inc. in 2014. Applying polymorphic techniques to web pages, the server response will be textually different between requests, but the visual display to the user will always be the same. That is, the values of static attributes and the structure of HTML documents may be modified on the server immediately before responses are sent off, creating a polymorphic version of the web page, or by pre-building this new versions on the server to decrease the real time computational costs. Therefore, no two HTML documents will be textually the same, turning web pages in somehow a moving target against MitB attacks. This level of protection is necessary since all changes are made locally, client side, making their detection difficult by control and security structures implemented on the service provider's servers.In this thesis, we aim to develop a tool based on polymorphism to protect web pages and users from MitB attacks based on markup and DOM anchors. This tool will be evaluated by accuracy and efficiency. The first metric will be evaluated by recording and comparing the list of errors and warnings generated by original web pages and by their polymorphic versions created with our tool. The efficiency will be evaluated by running automated attempts for tampering web pages protected by our tool.
APA, Harvard, Vancouver, ISO, and other styles
41

Lienhard, John. "Rohsenow Symposium web page." 2004. http://hdl.handle.net/1721.1/7307.

Full text
APA, Harvard, Vancouver, ISO, and other styles
42

Tsai, Ming-yung, and 蔡明原. "Related Web Page Retrieval Based on Semantic Concepts and Features of Web Pages." Thesis, 2005. http://ndltd.ncl.edu.tw/handle/fj4qej.

Full text
Abstract:
碩士
朝陽科技大學
資訊管理系碩士班
93
Using search engines to find information on the Internet often fails to satisfy user requirements. Previous search methodologies extended the domain of query keywords by the corresponding domain ontology to find related web pages, but typically they omitted the semantic content of the web pages, resulting in ineffective searches. In this paper, we present a related web page retrieval method that not only considers the corresponding domain ontology but also analyzes the semantic content of web pages. First, the method embeds the corresponding domain ontology of search keyword in order to find web pages from the Internet. Next, the method considers the location of the concept in the web pages, and relationships between concepts in the domain ontology when clustering the web page. Finally, an RDF structure is used to describe the relationships between keywords and web pages. We also used a Latent semantic analysis (LSA) algorithm to find relevant words in order to extend the information in the RDF. Experimental results prove that our method makes queries more effectively.
APA, Harvard, Vancouver, ISO, and other styles
43

Marath, Sathi. "Large-Scale Web Page Classification." Thesis, 2010. http://hdl.handle.net/10222/13130.

Full text
Abstract:
Web page classification is the process of assigning predefined categories to web pages. Empirical evaluations of classifiers such as Support Vector Machines (SVMs), k-Nearest Neighbor (k-NN), and Naïve Bayes (NB), have shown that these algorithms are effective in classifying small segments of web directories. The effectiveness of these algorithms, however, has not been thoroughly investigated on large-scale web page classification of such popular web directories as Yahoo! and LookSmart. Such web directories have hundreds of thousands of categories, deep hierarchies, spindle category and document distributions over the hierarchies, and skewed category distribution over the documents. These statistical properties indicate class imbalance and rarity within the dataset. In hierarchical datasets similar to web directories, expanding the content of each category using the web pages of the child categories helps to decrease the degree of rarity. This process, however, results in the localized overabundance of positive instances especially in the upper level categories of the hierarchy. The class imbalance, rarity and the localized overabundance of positive instances make applying classification algorithms to web directories very difficult and the problem has not been thoroughly studied. To our knowledge, the maximum number of categories ever previously classified on web taxonomies is 246,279 categories of Yahoo! directory using hierarchical SVMs leading to a Macro-F1 of 12% only. We designed a unified framework for the content based classification of imbalanced hierarchical datasets. The complete Yahoo! web directory of 639,671 categories and 4,140,629 web pages is used to setup the experiments. In a hierarchical dataset, the prior probability distribution of the subcategories indicates the presence or absence of class imbalance, rarity and the overabundance of positive instances within the dataset. Based on the prior probability distribution and associated machine learning issues, we partitioned the subcategories of Yahoo! web directory into five mutually exclusive groups. The effectiveness of different data level, algorithmic and architectural solutions to the associated machine learning issues is explored. Later, the best performing classification technologies for a particular prior probability distribution have been identified and integrated into the Yahoo! Web directory classification model. The methodology is evaluated using a DMOZ subset of 17,217 categories and 130,594 web pages and we statistically proved that the methodology of this research works equally well on large and small dataset. The average classifier performance in terms of macro-averaged F1-Measure achieved in this research for Yahoo! web directory and DMOZ subset is 81.02% and 84.85% respectively.
APA, Harvard, Vancouver, ISO, and other styles
44

Hu, Yony-Yi, and 胡永毅. "SharePoint Responsive Web Page Design." Thesis, 2015. http://ndltd.ncl.edu.tw/handle/06091112537820073844.

Full text
Abstract:
碩士
中原大學
應用數學研究所
103
As various hand held devices such as mobile phones and tablets become more and more popular, how to design a web page that can be displayed properly on the screens of all devices becomes a very important question. If we were to design the same web page for each possible device, then the maintenance afterwards will become very troublesome and will cost a lot of money. By using Responsive Web Design, we can design a web page once and have it displayed properly on various devices. In this study, we will experiment how to deliver Responsive Web Design web pages on SharePoint platform using Bootstrap. SharePoint Server is one of Microsoft’s product that we can ever encounter in enterprises. It is a product that can be used as a collaboration platform to promote internal or external communication of an enterprise. By using the web pages that SharePoint provides, users can communicate with any device that they have. Among the numerous functions that SharePoint provide, one of which is to let user edit a web page just like the way we usually type our Word document, and can easily configure different ways to view it and set permissions for it. Users can also design workflows on SharePoint without writing a single line of code. One frequently mentioned of Responsive Web Design is when displaying a web page using different screen resolution, the web page will adapt to the screen resolution and modify its content of display. Bootstrap is a set of tools that can be applied on web sites and web applications. It content includes frameworks for HTML, CSS and JavaScript, providing various effects on typesetting, controls used on web pages and navigations of web pages. For web sites and web applications that need to provide service to various devices and their browsers, Bootstrap provides CSS media query that can save web designers a huge amount of time and work since they will no longer to have a version for each of their client devices.
APA, Harvard, Vancouver, ISO, and other styles
45

許烘祥. "Bidirectional Integrated Web Page System." Thesis, 1997. http://ndltd.ncl.edu.tw/handle/77750544223640859097.

Full text
APA, Harvard, Vancouver, ISO, and other styles
46

Chean, Chao-Nan, and 陳昭男. "Detection of Page Type, Time, and Key Terms of Web Pages." Thesis, 2003. http://ndltd.ncl.edu.tw/handle/49481238888740466198.

Full text
Abstract:
碩士
國立中正大學
資訊工程研究所
91
With the rapid growth of WWW, the amount of online resources is getting richer. Modern search engines not only provide general search service for web pages, but domain-specified or type-specified search service to meet users'' need. To be able to provide type-specified search service, one needs to build up an automatic mechanism for type detection. By statistical analysis of the web pages, we find out some features which are appropriate for type detection. We also propose a scoring method to evaluate which type the web page belongs. Sometimes, the time information described in the content of the web page may be different from the last modified time of the web page. We define some rules to detect the time information from the web page. When extracting key terms, three features are calculated for each term in the web page. They are: location, which is the term''s first appearance; emphatic tag, whether the term is emphasized by some kinds of HTML tag or not; and TFIDF, a generality measure of a term''s frequency in a web page.
APA, Harvard, Vancouver, ISO, and other styles
47

"Sequence-based Web Page Template Detection." Master's thesis, 2011. http://hdl.handle.net/2286/R.I.9268.

Full text
Abstract:
abstract: Templates are wildly used in Web sites development. Finding the template for a given set of Web pages could be very important and useful for many applications like Web page classification and monitoring content and structure changes of Web pages. In this thesis, two novel sequence-based Web page template detection algorithms are presented. Different from tree mapping algorithms which are based on tree edit distance, sequence-based template detection algorithms operate on the Prüfer/Consolidated Prüfer sequences of trees. Since there are one-to-one correspondences between Prüfer/Consolidated Prüfer sequences and trees, sequence-based template detection algorithms identify the template by finding a common subsequence between to Prüfer/Consolidated Prüfer sequences. This subsequence should be a sequential representation of a common subtree of input trees. Experiments on real-world web pages showed that our approaches detect templates effectively and efficiently.
Dissertation/Thesis
M.S. Computer Science 2011
APA, Harvard, Vancouver, ISO, and other styles
48

Videira, António Miguel Baptista. "Web Page Classification using Visual Features." Master's thesis, 2013. http://hdl.handle.net/10316/40388.

Full text
APA, Harvard, Vancouver, ISO, and other styles
49

Dai, Shyh-Ming, and 戴世明. "Link-based Automatic Web Page Classification." Thesis, 2002. http://ndltd.ncl.edu.tw/handle/72672279859286395513.

Full text
Abstract:
碩士
元智大學
資訊工程學系
90
As the Internet rapidly develops, the amount of information accumulates vastly.Web search engines and categories help users to find important information quickly and effectively.Therefore, Web search engines and categories of Web pages have became two important services on the Internet. However, either Web search engines or categories of Web pages need some support mechanisms for precisely classifying Web pages to improve the effectiveness. The automatic Web page classification is one of the mechanisms. Because the amount of Internet information is too huge to be classified manually, the automatic Web page classification is becoming the main stream of Web page classification.However, two problems need to be discussed further: how to improve the classification accuracy and how to reduce the ratio of the pages that can not be classified at all.This thesis proposes a new approach called linked-based automatic Web page classification to relief the problems.We improve a tag-weighted approach (Jenkins&Inman) by incorporating link analysis, which picks out the authority links from the Web page being classified, and analyzing the contents which pointed by the authority links. We have conducted experiments to compare our approach with Jenkins&Inman approach. We used a set of classified Yahoo! Web pages for training and verification.The experiment results show that the linked-based automatic Web pages classification indeed improves the classification correctness rate and reduces the amount of Web pagess which cannot be classified in Jenkins'' approach.
APA, Harvard, Vancouver, ISO, and other styles
50

Pi-Hsien, Chang, and 張碧顯. "Web Structure and Page Relationship Discovery from Web Server Log." Thesis, 2008. http://ndltd.ncl.edu.tw/handle/20161066534968523651.

Full text
Abstract:
碩士
國立暨南國際大學
資訊工程學系
96
Web usage mining which extracts knowledge from Web server log is an application of data mining method. The mining results can be used for improving the Web design, predicating user behavior and personalizing Web site. Web usage mining has three major stages: data preprocessing, pattern discovery and pattern analysis. Data pre-processing, which normally spends more than 60% of the whole mining process, is most time consuming. Cooley divided data preprocessing into four and one optional steps. They were data cleaning, user/session identification, path completion, page view identification and transaction identification which is optional. Until now, the preprocessing of Web usage mining must gather external domain knowledge, such as Web structure and Web content classification, which greatly affects the application of Web usage mining. It takes more time for the analyst to be familiar with Web structure and content. For Web administrator, she/he may have concerns with the confidential Web data when giving the detailed Web structure to the analyst. Thus, we want to solve the problem by creating a platform between analysts and Web administrators to help them better communicate during the Web usage mining progress. In this thesis, we propose a framework that can reconstruct Web structure and discover the page relationship from Web server log’s implicit information. The experimental results showed that Web site reconstruction and page relationship discovery with precision of more than 90%. This method that can be easily embedded in the popular preprocessing stage is a workable and practical substitute method.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography