Tesis sobre el tema "Text Data Streams"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte los 28 mejores tesis para su investigación sobre el tema "Text Data Streams".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Explore tesis sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.
Snowsill, Tristan. "Data mining in text streams using suffix trees". Thesis, University of Bristol, 2012. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.556708.
Texto completoMejova, Yelena Aleksandrovna. "Sentiment analysis within and across social media streams". Diss., University of Iowa, 2012. https://ir.uiowa.edu/etd/2943.
Texto completoHill, Geoffrey. "Sensemaking in Big Data: Conceptual and Empirical Approaches to Actionable Knowledge Generation from Unstructured Text Streams". Kent State University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=kent1433597354.
Texto completoPinho, Roberto Dantas de. "Espaço incremental para a mineração visual de conjuntos dinâmicos de documentos". Universidade de São Paulo, 2009. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-14092009-123807/.
Texto completoVisual representations are often adopted to explore document collections, assisting in knowledge extraction, and avoiding the thorough analysis of thousands of documents. Document maps present individual documents in visual spaces in such a way that their placement reflects similarity relations or connections between them. Building these maps requires, among other tasks, placing each document and identifying interesting areas or subsets. A current challenge is to visualize dynamic data sets. In Information Visualization, adding and removing data elements can strongly impact the underlying visual space. That can prevent a user from preserving a mental map that could assist her/him on understanding the content of a growing collection of documents or tracking changes on the underlying data set. This thesis presents a novel algorithm to create dynamic document maps, capable of maintaining a coherent disposition of elements, even for completely renewed sets. The process is inherently incremental, has low complexity and places elements on a 2D grid, analogous to a chess board. Consistent results were obtained as compared to (non-incremental) multidimensional scaling solutions, even when applied to visualizing domains other than document collections. Moreover, the corresponding visualization is not susceptible to occlusion. To assist users in indentifying interesting subsets, a topic extraction technique based on association rule mining was also developed. Together, they create a visual space where topics and interesting subsets are highlighted and constantly updated as the data set changes
Wu, Yingyu. "Using Text based Visualization in Data Analysis". Kent State University / OhioLINK, 2014. http://rave.ohiolink.edu/etdc/view?acc_num=kent1398079502.
Texto completoYoung, Tom y Mark Wigent. "Dynamic Formatting of the Test Article Data Stream". International Foundation for Telemetering, 2010. http://hdl.handle.net/10150/605948.
Texto completoCrossman, Nathaniel C. "Stream Clustering And Visualization Of Geotagged Text Data For Crisis Management". Wright State University / OhioLINK, 2020. http://rave.ohiolink.edu/etdc/view?acc_num=wright1590957641168863.
Texto completoFranco, Tom. "Performing Frame Transformations to Correctly Stream Position Data". University of Cincinnati / OhioLINK, 2017. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1491562251744704.
Texto completoVickers, Stephen R. "Examining the Duplication of Flight Test Data Centers". International Foundation for Telemetering, 2011. http://hdl.handle.net/10150/595653.
Texto completoAircraft flight test data processing began with on site data analysis from the very first aircraft design. This method of analyzing flight data continued from the early 1900's to the present day. Today each new aircraft program builds a separate data center for post flight processing (PFP) to include operations, system administration, and management. Flight Test Engineers (FTE) are relocated from geographical areas to ramp up the manpower needed to analyze the PFP data center products and when the first phase of aircraft design and development is completed the FTE headcount is reduced with the FTE either relocated to another program or the FTE finds other employment. This paper is a condensed form of the research conducted by the author on how the methodology of continuing to build PFP data centers cost the aircraft company millions of dollars in development and millions of dollars on relocation plus relocation stress effects on FTE which can hinder productivity. This method of PFP data center development can be avoided by the consolidation of PFP data centers using present technology.
Yates, James William. "Mixing Staged Data Flow and Stream Computing Techniques in Modern Telemetry Data Acquisition/Processing Architectures". International Foundation for Telemetering, 1999. http://hdl.handle.net/10150/608707.
Texto completoToday’s flight test processing systems must handle many more complex data formats than just the PCM and analog FM data streams of yesterday. Many flight test programs, and their respective test facilities, are looking to leverage their computing assets across multiple customers and programs. Typically, these complex programs require the ability to handle video, packet, and avionics bus data in real time, in addition to handling the more traditional PCM format. Current and future telemetry processing systems must have an architecture that will support the acquisition and processing of these varied data streams. This paper describes various architectural designs of both staged data flow and stream computing architectures, including current and future implementations. Processor types, bus design, and the effects of varying data types, including PCM, video, and packet telemetry, will be discussed.
Boppudi, Srimanth. "Further Investigation of a New Traction Stress Based Shear Strength Characterization Method with Test Data". ScholarWorks@UNO, 2014. http://scholarworks.uno.edu/td/1847.
Texto completoKříž, Blažej. "Framework pro tvorbu generátorů dat". Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2012. http://www.nusl.cz/ntk/nusl-236623.
Texto completoLuo, Dan y Yajing Ran. "Micro Drivers behind the Changes of CET1 Capital Ratio : An empirical analysis based on the results of EU-wide stress test". Thesis, Internationella Handelshögskolan, Högskolan i Jönköping, IHH, Företagsekonomi, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:hj:diva-44140.
Texto completoFaleiros, Thiago de Paulo. "Propagação em grafos bipartidos para extração de tópicos em fluxo de documentos textuais". Universidade de São Paulo, 2016. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-10112016-105854/.
Texto completoHandling large amounts of data is a requirement for modern text mining algorithms. For some applications, documents are published constantly, which demand a high cost for long-term storage. So it is necessary easily adaptable methods for an approach that considers documents flow, and be capable of analyzing the data in one step without requiring the high cost of storage. Another requirement is that this approach can exploit heuristics in order to improve the quality of results. Several models for automatic extraction of latent information in a collection of documents have been proposed in the literature, among them probabilistic topic models are prominent. Probabilistic topic models achieve good practical results, and have been extended to several models with different types of information included. However, properly describe these models, derive them, and then get appropriate inference algorithms are difficult tasks, requiring a rigorous mathematical treatment for descriptions of operations performed in the latent dimensions discovery process. Thus, for the development of a simple and efficient method to tackle the problem of latent dimensions discovery, a proper representation of the data is required. The hypothesis of this thesis is that by using bipartite graph for representation of textual data one can address the task of latent patterns discovery, present in the relationships between documents and words, in a simple and intuitive way. For validation of this hypothesis, we have developed a framework based on label propagation algorithm using the bipartite graph representation. The framework, called PBG (Propagation in Bipartite Graph) was initially applied to the unsupervised context for a static collection of documents. Then a semi-supervised version was proposed which need only a small amount of labeled documents to the transductive classification task. Finally, it was applied in the dynamic context in which flow of textual data was considered. Comparative analyzes were performed, and the results indicated that the PBG is a viable and competitive alternative for tasks in the unsupervised and semi-supervised contexts.
Al-Ajmi, Adel. "Wellbore stability analysis based on a new true-triaxial failure criterion". Doctoral thesis, Stockholm : Department of Land and Water Resources Engineering, Royal Institute of Technology, 2006. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-4037.
Texto completoGONÇALVES, JÚNIOR Paulo Mauricio. "Multivariate non-parametric statistical tests to reuse classifiers in recurring concept drifting environments". Universidade Federal de Pernambuco, 2013. https://repositorio.ufpe.br/handle/123456789/12226.
Texto completoSubmitted by João Arthur Martins (joao.arthur@ufpe.br) on 2015-03-12T18:02:08Z No. of bitstreams: 2 Tese Paulo Gonçalves Jr..pdf: 2957463 bytes, checksum: de163caadf10cbd5442e145778865224 (MD5) license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5)
Made available in DSpace on 2015-03-12T18:02:08Z (GMT). No. of bitstreams: 2 Tese Paulo Gonçalves Jr..pdf: 2957463 bytes, checksum: de163caadf10cbd5442e145778865224 (MD5) license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5) Previous issue date: 2013-04-23
Fluxos de dados s~ao um modelo de processamento de dados recente, onde os dados chegam continuamente, em grandes quantidades, a altas velocidades, de modo que eles devem ser processados em tempo real. Al em disso, v arias institui c~oes p ublicas e privadas armazenam grandes quantidades de dados que tamb em devem ser processadas. Classi cadores tradicionais n~ao s~ao adequados para lidar com grandes quantidades de dados por basicamente duas raz~oes. Primeiro, eles costumam ler os dados dispon veis v arias vezes at e convergirem, o que e impratic avel neste cen ario. Em segundo lugar, eles assumem que o contexto representado por dados e est avel no tempo, o que pode n~ao ser verdadeiro. Na verdade, a mudan ca de contexto e uma situa c~ao comum em uxos de dados, e e chamado de mudan ca de conceito. Esta tese apresenta o rcd, uma estrutura que oferece uma abordagem alternativa para lidar com os uxos de dados que sofrem de mudan cas de conceito recorrentes. Ele cria um novo classi cador para cada contexto encontrado e armazena uma amostra dos dados usados para constru -lo. Quando uma nova mudan ca de conceito ocorre, rcd compara o novo contexto com os antigos, utilizando um teste estat stico n~ao param etrico multivariado para veri car se ambos os contextos prov^em da mesma distribui c~ao. Se assim for, o classi cador correspondente e reutilizado. Se n~ao, um novo classi cador e gerado e armazenado. Tr^es tipos de testes foram realizados. Um compara o rcd com v arios algoritmos adaptativos (entre as abordagens individuais e de agrupamento) em conjuntos de dados arti ciais e reais, entre os mais utilizados na area de pesquisa de mudan ca de conceito, com mudan cas bruscas e graduais. E observada a capacidade dos classi cadores em representar cada contexto, como eles lidam com as mudan cas de conceito e os tempos de treinamento e teste necess arios para avaliar os conjuntos de dados. Os resultados indicam que rcd teve resultados estat sticos semelhantes ou melhores, em compara c~ao com os outros classi cadores. Nos conjuntos de dados do mundo real, rcd apresentou precis~oes pr oximas do melhor classi cador em cada conjunto de dados. Outro teste compara dois testes estat sticos (knn e Cramer) em suas capacidades de representar e identi car contextos. Os testes foram realizados utilizando classi cadores xi xii RESUMO tradicionais e adaptativos como base do rcd, em conjuntos de dados arti ciais e do mundo real, com v arias taxas de varia c~ao. Os resultados indicam que, em m edia, KNN obteve melhores resultados em compara c~ao com o teste de Cramer, al em de ser mais r apido. Independentemente do crit erio utilizado, rcd apresentou valores mais elevados de precis~ao em compara c~ao com seus respectivos classi cadores base. Tamb em e apresentada uma melhoria do rcd onde os testes estat sticos s~ao executadas em paralelo por meio do uso de um pool de threads. Os testes foram realizados em tr^es processadores com diferentes n umeros de n ucleos. Melhores resultados foram obtidos quando houve um elevado n umero de mudan cas de conceito detectadas, o tamanho das amostras utilizadas para representar cada distribui c~ao de dados era grande, e havia uma alta freq u^encia de testes. Mesmo que nenhuma destas condi c~oes se aplicam, a execu c~ao paralela e seq uencial ainda t^em performances muito semelhantes. Finalmente, uma compara c~ao entre seis diferentes m etodos de detec c~ao de mudan ca de conceito tamb em foi realizada, comparando a precis~ao, os tempos de avalia c~ao, manipula c~ao das mudan cas de conceito, incluindo as taxas de falsos positivos e negativos, bem como a m edia da dist^ancia ao ponto de mudan ca e o seu desvio padr~ao.
Gonçalves, Júnior Paulo Mauricio. "Multivariate non-parametric statistical tests to reuse classifiers in recurring concept drifting environments". Universidade Federal de Pernambuco, 2013. https://repositorio.ufpe.br/handle/123456789/12288.
Texto completoSubmitted by João Arthur Martins (joao.arthur@ufpe.br) on 2015-03-12T19:25:11Z No. of bitstreams: 2 license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5) tese Paulo Mauricio Gonçalves Jr..pdf: 2957463 bytes, checksum: de163caadf10cbd5442e145778865224 (MD5)
Made available in DSpace on 2015-03-12T19:25:11Z (GMT). No. of bitstreams: 2 license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5) tese Paulo Mauricio Gonçalves Jr..pdf: 2957463 bytes, checksum: de163caadf10cbd5442e145778865224 (MD5) Previous issue date: 2013-04-23
Fluxos de dados s~ao um modelo de processamento de dados recente, onde os dados chegam continuamente, em grandes quantidades, a altas velocidades, de modo que eles devem ser processados em tempo real. Al em disso, v arias institui c~oes p ublicas e privadas armazenam grandes quantidades de dados que tamb em devem ser processadas. Classi cadores tradicionais n~ao s~ao adequados para lidar com grandes quantidades de dados por basicamente duas raz~oes. Primeiro, eles costumam ler os dados dispon veis v arias vezes at e convergirem, o que e impratic avel neste cen ario. Em segundo lugar, eles assumem que o contexto representado por dados e est avel no tempo, o que pode n~ao ser verdadeiro. Na verdade, a mudan ca de contexto e uma situa c~ao comum em uxos de dados, e e chamado de mudan ca de conceito. Esta tese apresenta o rcd, uma estrutura que oferece uma abordagem alternativa para lidar com os uxos de dados que sofrem de mudan cas de conceito recorrentes. Ele cria um novo classi cador para cada contexto encontrado e armazena uma amostra dos dados usados para constru -lo. Quando uma nova mudan ca de conceito ocorre, rcd compara o novo contexto com os antigos, utilizando um teste estat stico n~ao param etrico multivariado para veri car se ambos os contextos prov^em da mesma distribui c~ao. Se assim for, o classi cador correspondente e reutilizado. Se n~ao, um novo classi cador e gerado e armazenado. Tr^es tipos de testes foram realizados. Um compara o rcd com v arios algoritmos adaptativos (entre as abordagens individuais e de agrupamento) em conjuntos de dados arti ciais e reais, entre os mais utilizados na area de pesquisa de mudan ca de conceito, com mudan cas bruscas e graduais. E observada a capacidade dos classi cadores em representar cada contexto, como eles lidam com as mudan cas de conceito e os tempos de treinamento e teste necess arios para avaliar os conjuntos de dados. Os resultados indicam que rcd teve resultados estat sticos semelhantes ou melhores, em compara c~ao com os outros classi cadores. Nos conjuntos de dados do mundo real, rcd apresentou precis~oes pr oximas do melhor classi cador em cada conjunto de dados. Outro teste compara dois testes estat sticos (knn e Cramer) em suas capacidades de representar e identi car contextos. Os testes foram realizados utilizando classi cadores tradicionais e adaptativos como base do rcd, em conjuntos de dados arti ciais e do mundo real, com v arias taxas de varia c~ao. Os resultados indicam que, em m edia, KNN obteve melhores resultados em compara c~ao com o teste de Cramer, al em de ser mais r apido. Independentemente do crit erio utilizado, rcd apresentou valores mais elevados de precis~ao em compara c~ao com seus respectivos classi cadores base. Tamb em e apresentada uma melhoria do rcd onde os testes estat sticos s~ao executadas em paralelo por meio do uso de um pool de threads. Os testes foram realizados em tr^es processadores com diferentes n umeros de n ucleos. Melhores resultados foram obtidos quando houve um elevado n umero de mudan cas de conceito detectadas, o tamanho das amostras utilizadas para representar cada distribui c~ao de dados era grande, e havia uma alta freq u^encia de testes. Mesmo que nenhuma destas condi c~oes se aplicam, a execu c~ao paralela e seq uencial ainda t^em performances muito semelhantes. Finalmente, uma compara c~ao entre seis diferentes m etodos de detec c~ao de mudan ca de conceito tamb em foi realizada, comparando a precis~ao, os tempos de avalia c~ao, manipula c~ao das mudan cas de conceito, incluindo as taxas de falsos positivos e negativos, bem como a m edia da dist^ancia ao ponto de mudan ca e o seu desvio padr~ao.
Di, Molfetta Sabino. "Studio del modello di vita e di affidabilità di condensatori "Brick" in film per applicazione Automotive per macchine elettriche o ibride". Master's thesis, Alma Mater Studiorum - Università di Bologna, 2019.
Buscar texto completoLeboullenger, Déborah. "Essais sur la transition énergétique : enjeux, valorisation, financement et risques". Thesis, Paris 10, 2017. http://www.theses.fr/2017PA100065/document.
Texto completoThis thesis deals with the main challenges that we need to address to foster the private financing of a low-carbon energy transition. A massive amount of investment in low carbon assets is needed and most of the effort must come from final energy consumers such as households. Their ability as well as the ability of the financial intermediation institutions (that is banks in Europe) to valorise low-carbon investments and risk profile is the key for a successful low-carbon energy transition in France and in every industrialized country. These researches focus more particularly on the housing sector which represents 44% of the final energy consumption and 21% of the total greenhouses gases emissions in France. The first chapter of this thesis takes the viewpoint that only a disaggregated approach can actually permit macroeconomic and nationwide objectives to reduce final energy consumption match microeconomic arbitrages regarding energy spending in the private residential housing sector. Using segmentation and decision tree growing econometric techniques, the chapter proposes a typology of energy spending and a segmentation analysis and of the energy transition “market” in the housing sector. The second chapter uses frontier functions estimation technique on a local French private housing market to determine if selling prices contain a “Green Property Value”. An empirical analysis is then conducted to determine if this value can offset the upfront cost of energy retrofit. The last chapter takes the prism of the financial institutions. It attempts a first evaluation of the impact and exposition to climate related risks, those are physical, transition, liability and systemic risks, on the banking system and its prudential regulation framework
Bourniquel, Bernard. "Evaluation des deformations mecaniques de surface par diffraction x. Optimisation de la mesure des contraintes residuelles. Application au controle qualite du grenaillage de precontrainte". Nantes, 1988. http://www.theses.fr/1988NANT2037.
Texto completoJakel, Roland. "Grundlagen der Elasto-Plastizität in Creo Simulate - Theorie und Anwendung". Universitätsbibliothek Chemnitz, 2012. http://nbn-resolving.de/urn:nbn:de:bsz:ch1-qucosa-87141.
Texto completoThis presentation describes the basics of elasto-plasticity and its application with the finite element software Creo Simulate (formerly Pro/MECHANICA) from PTC. The first part describes the characteristics of plastic behavior, different plastic material laws, yield criteria for multiaxial stress states and different hardening models. In the second part, the opportunities and limitations of analyzing elasto-plastic problems with the FEM-code are described and user information is provided. The last part finally presents different examples. Deeply treated is the behavior of a uniaxial tensile test specimen before and after elongation with necking appears
Abouelnagah, Younes. "Efficient Temporal Synopsis of Social Media Streams". Thesis, 2013. http://hdl.handle.net/10012/7689.
Texto completoChou, Cheng-Chieh y 周正杰. "Planning step-stress test plans based on censored data". Thesis, 2015. http://ndltd.ncl.edu.tw/handle/24428386624288427049.
Texto completo淡江大學
數學學系博士班
103
In this dissertation, we discuss a k-level step-stress accelerated life-testing (ALT) experiment with unequal duration steps. Under the Type-I and Type-I hybrid censoring schemes, the general log-location-scale and exponential lifetime distributions with mean lives which are a linear function of stress for the former and a log-linear function of stress for the latter, along with a cumulative exposure model, are considered as the working models. The determination of the optimal unequal duration steps for exponential, Weibull and lognormal distributions are addressed using the variance-optimality criterion. Numerical results show that for the general log-location-scale and exponential distributions, the optimal k-level step-stress ALT model with unequal duration steps reduces just to a 2-level step-stress ALT model when the available data is either Type-I or Type-I hybrid censored data. Moreover, using the induction argument, we are capable to give a theoretical proof for this result based on a Type-I exponential censored data.
Lin, Ying-Po y 林英博. "Optimal Step-Stress Test under Progressive Type I Censoring with Grouped Data". Thesis, 2003. http://ndltd.ncl.edu.tw/handle/74396472995776456245.
Texto completo淡江大學
統計學系
91
In the study of product reliability, a life test usually has to be conducted. There are several types of life testing experiments. Type I and Type II censoring schemes have been studied rather extensively by lots of researchers in order to obtain the lifetimes of products. These two schemes do not allow for units to be removed from the test at points other than the final termination point. However, this allowance will be desirable for some experimenters. Therefore, a progressive censoring scheme is proposed to handle this problem. With today's high technology, many products are designed to work without failure for years. Thus, some life tests result in few or no failures in a short life testing time. One approach to solve this problem is to accelerate the life of products by increasing the levels of stress in order to obtain failures quickly. Moreover, in practice, it is often impossible continuously to observe or inspect the testing process, even with censoring. We might only be able to inspect the test units intermittently. Hence, we observe only the number of failures within the time period, but not the ssociated failure times. Data of this type are called grouped data. In this thesis, we are going to combine progressive censoring, accelerated life test and grouped data to develop a step-stress accelerated life-testing scheme with type I progressive group-censoring. We will obtain the estimators of the parameter in the proposed model when the failure time distribution is exponential. The problem of choosing the optimal length of the inspection interval will also be addressed using the variance and the D-optimality criteria.
Chen, Hsin-Hao y 陳信豪. "Acceptance Sampling Plans under Step-stress Test and Type Ⅰ Interval Censoring Data". Thesis, 2006. http://ndltd.ncl.edu.tw/handle/92860141347330162235.
Texto completo國立政治大學
統計研究所
94
In life test experiment we use interval censoring to complete it when we can not inspect the experiment units continuously due to some accidents or for convenience. Furthermore, it is difficult to obtain enough units of breakdown products for many long life components and products. At this moment we can adopt step-stress life test to proceed the experiment. Using this method we can make the test units breakdown early for reducing the time test needed effectively and save prime cost. In this thesis, acceptance sampling plans are established for Rayleigh lifetime data under step-stress and type I interval censoring scheme. The minimum sample sizes and the corresponding critical values of lifetime needed for test plans are found. Some tables are provided for the use of the proposed test plans.
Wang, Ye. "Robust Text Mining in Online Social Network Context". Thesis, 2018. https://vuir.vu.edu.au/38645/.
Texto completoHaldenwang, Nils. "Reliable General Purpose Sentiment Analysis of the Public Twitter Stream". Doctoral thesis, 2017. https://repositorium.ub.uni-osnabrueck.de/handle/urn:nbn:de:gbv:700-2017092716282.
Texto completoFernandes, Sebastião Cardoso. "How to deal with extreme cases for credit risk monitoring: a case study in a credit risk data science company". Master's thesis, 2018. http://hdl.handle.net/10362/35455.
Texto completo