Dissertationen zum Thema „Data-based testing“
Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an
Machen Sie sich mit Top-50 Dissertationen für die Forschung zum Thema "Data-based testing" bekannt.
Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.
Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.
Sehen Sie die Dissertationen für verschiedene Spezialgebieten durch und erstellen Sie Ihre Bibliographie auf korrekte Weise.
Tsai, Bor-Yuan. „A hybrid object-oriented class testing method : based on state-based and data-flow testing“. Thesis, University of Sunderland, 2000. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.311294.
Der volle Inhalt der QuelleLi, Dongmei. „Resampling-based Multiple Testing with Applications to Microarray Data Analysis“. The Ohio State University, 2009. http://rave.ohiolink.edu/etdc/view?acc_num=osu1243993319.
Der volle Inhalt der QuelleNoor, Tanzeem Bin. „A Similarity-based Test Case Quality Metric using Historical Failure Data“. IEEE, 2015. http://hdl.handle.net/1993/31045.
Der volle Inhalt der QuelleFebruary 2016
Nordholm, Johan. „Model-Based Testing: An Evaluation“. Thesis, Karlstad University, Faculty of Economic Sciences, Communication and IT, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:kau:diva-5188.
Der volle Inhalt der QuelleTesting is a critical activity in the software development process in order to obtain systems of high quality. Tieto typically develops complex systems, which are currently tested through a large number of manually designed test cases. Recent development within software testing has resulted in methods and tools that can automate the test case design, the generation of test code and the test result evaluation based on a model of the system under test. This testing approach is called model-based testing (MBT).
This thesis is a feasibility study of the model-based testing concept and has been performed at the Tieto office in Karlstad. The feasibility study included the use and evaluation of the model-based testing tool Qtronic, developed by Conformiq, which automatically designs test cases given a model of the system under test as input. The experiments for the feasibility study were based on the incremental development of a test object, which was the client protocol module of a simplified model for an ATM (Automated Teller Machine) client-server system. The experiments were evaluated both individually and by comparison with the previous experiment since they were based on incremental development. For each experiment the different tasks in the process of testing using Qtronic were analyzed to document the experience gained as well as to identify strengths and weaknesses.
The project has shown the promise inherent in using a model-based testing approach. The application of model-based testing and the project results indicate that the approach should be further evaluated since experience will be crucial if the approach is to be adopted within Tieto’s organization.
Lima, Lucas Albertins de. „Test case prioritization based on data reuse for Black-box environments“. Universidade Federal de Pernambuco, 2009. https://repositorio.ufpe.br/handle/123456789/1922.
Der volle Inhalt der QuelleConselho Nacional de Desenvolvimento Científico e Tecnológico
Albertins de Lima, Lucas; Cezar Alves Sampaio, Augusto. Test case prioritization based on data reuse for Black-box environments. 2009. Dissertação (Mestrado). Programa de Pós-Graduação em Ciência da Computação, Universidade Federal de Pernambuco, Recife, 2009.
Tracey, Nigel James. „A search-based automated test-data generation framework for safety-critical software“. Thesis, University of York, 2000. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.325796.
Der volle Inhalt der QuelleWoldeselassie, Tilahun. „A simple microcomputer-based nuclear medicine data processing system design and performance testing“. Thesis, University of Aberdeen, 1990. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.316066.
Der volle Inhalt der QuelleMairhofer, Stefan. „Search-based software testing and complex test data generation in a dynamic programming language“. Thesis, Blekinge Tekniska Högskola, Avdelningen för programvarusystem, 2008. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-4340.
Der volle Inhalt der QuelleNdashimye, Maurice. „Accounting for proof test data in Reliability Based Design Optimization“. Thesis, Stellenbosch : Stellenbosch University, 2015. http://hdl.handle.net/10019.1/97108.
Der volle Inhalt der QuelleENGLISH ABSTRACT: Recent studies have shown that considering proof test data in a Reliability Based Design Optimization (RBDO) environment can result in design improvement. Proof testing involves the physical testing of each and every component before it enters into service. Considering the proof test data as part of the RBDO process allows for improvement of the original design, such as weight savings, while preserving high reliability levels. Composite Over-Wrapped Pressure Vessels (COPV) is used as an example application of achieving weight savings while maintaining high reliability levels. COPVs are light structures used to store pressurized fluids in space shuttles, the international space station and other applications where they are maintained at high pressure for extended periods of time. Given that each and every COPV used in spacecraft is proof tested before entering service and any weight savings on a spacecraft results in significant cost savings, this thesis put forward an application of RBDO that accounts for proof test data in the design of a COPV. The method developed in this thesis shows that, while maintaining high levels of reliability, significant weight savings can be achieved by including proof test data in the design process. Also, the method enables a designer to have control over the magnitude of the proof test, making it possible to also design the proof test itself depending on the desired level of reliability for passing the proof test. The implementation of the method is discussed in detail. The evaluation of the reliability was based on the First Order Reliability Method (FORM) supported by Monte Carlo Simulation. Also, the method is implemented in a versatile way that allows the use of analytical as well as numerical (in the form of finite element) models. Results show that additional weight savings can be achieved by the inclusion of proof test data in the design process.
AFRIKAANSE OPSOMMING: Onlangse studies het getoon dat die gebruik van ontwerp spesifieke proeftoets data in betroubaarheids gebaseerde optimering (BGO) kan lei tot 'n verbeterde ontwerp. BGO behels vele aspekte in die ontwerpsgebied. Die toevoeging van proeftoets data in ontwerpsoptimering bring te weë; die toetsing van 'n ontwerp en onderdele voor gebruik, die aangepaste en verbeterde ontwerp en gewig-besparing met handhawing van hoë betroubaarsheidsvlakke. 'n Praktiese toepassing van die BGO tegniek behels die ontwerp van drukvatte met saamgestelde materiaal bewapening. Die drukvatontwerp is 'n ligte struktuur wat gebruik word in die berging van hoë druk vloeistowwe in bv. in ruimtetuie, in die internasionale ruimtestasie en in ander toepassings waar hoë druk oor 'n tydperk verlang word. Elke drukvat met saamgestelde materiaal bewapening wat in ruimtevaartstelsels gebruik word, word geproeftoets voor gebruik. In ruimte stelselontwerp lei massa besparing tot 'n toename in loonvrag. Die tesis beskryf 'n optimeringsmetode soos ontwikkel en gebaseer op 'n BGO tegniek. Die metode word toegepas in die ontwerp van drukvatte met saamgestelde materiaal bewapening. Die resultate toon dat die gebruik van proeftoets data in massa besparing optimering onderhewig soos aan hoë betroubaarheidsvlakke moontlik is. Verdermeer, die metode laat ook ontwerpers toe om die proeftoetsvlak aan te pas om sodoende by ander betroubaarheidsvlakke te toets. In die tesis word die ontwikkeling en gebruik van die optimeringsmetode uiteengelê. Die evaluering van betroubaarheidsvlakke is gebaseer op 'n eerste orde betroubaarheids-tegniek wat geverifieer word met talle Monte Carlo simulasie resultate. Die metode is ook so geskep dat beide analitiese sowel as eindige element modelle gebruik kan word. Ten slotte, word 'n toepassing getoon waar resultate wys dat die gebruik van die optimeringsmetode met die insluiting van proeftoets data wel massa besparing kan oplewer.
Bayley, Gwain. „PC-based bit error rate analyser for a 2 Mbps data link“. Thesis, University of Cape Town, 1988. http://hdl.handle.net/11427/23153.
Der volle Inhalt der QuelleMoyers, Kevin Keith. „A microcomputer-based data acquisition system for diagnostic monitoring and control of high-speed electric motors“. Thesis, Virginia Tech, 1987. http://hdl.handle.net/10919/41576.
Der volle Inhalt der QuelleBearing vibration and outer race temperature were continuously monitored for each ball bearing in the motor. In addition, the stator winding and motor casing temperature were monitored.
The monitoring system was successful in detecting an unbalance in the rotor caused by the loss of a small piece of balancing putty. The motor was shut down before any further damage occurred. In a separate test, excessive clearance between a bearing outer race and the motor caused high vibration readings. The motor was monitored until the condition began to deteriorate and the bearing outer race began to spin significantly. Again, the monitoring system powered down the motor before any significant damage occurred.
The speed of the motor tested is controlled by a PWM (pulse width modulation) technique. The resulting voltage and current waveforms are asymmetrical and contain high frequency components. Special circuitry was designed and constructed to interface sensors for measuring the voltage and current inputs to a spectrum analyzer. Using frequency and order analysis techniques, the real and reactive power inputs to the three·phase motor were measured.
Master of Science
Jury, Owen T. „The Design of Telemetry Acquisition and Analysis Vans for Testing Construction and Mining Equipment“. International Foundation for Telemetering, 1997. http://hdl.handle.net/10150/607566.
Der volle Inhalt der QuelleCaterpillar Inc. has over 25 years of experience using instrument vans equipped with telemetry to support product testing. These vans provide the capability to instrument the product, to acquire telemetered data, and to analyze the data. They are being used in tests performed on construction and mining equipment at Caterpillar's proving grounds and at customer job sites throughout North America. This paper presents a design summary of the newest generation vans. It starts with an overview of the major subsystems and concentrates on the Caterpillar developed software that tightly integrates the various hardware and software components. This software greatly enhances the productivity of the system and makes it possible for the van to perform a large variety and quantity of tests required by our internal customers.
Meyer, Mark J. „Understanding the challenges in HEV 5-cycle fuel economy calculations based on dynamometer test data“. Thesis, Virginia Tech, 2011. http://hdl.handle.net/10919/35648.
Der volle Inhalt der QuelleMaster of Science
Moore, Albert W. „A computer-based training course for assessing material safety data sheet comprehension“. Thesis, This resource online, 1994. http://scholar.lib.vt.edu/theses/available/etd-06232009-063332/.
Der volle Inhalt der QuelleNandi, Shinjini. „Multiple Testing Procedures for One- and Two-Way Classified Hypotheses“. Diss., Temple University Libraries, 2019. http://cdm16002.contentdm.oclc.org/cdm/ref/collection/p245801coll10/id/580415.
Der volle Inhalt der QuellePh.D.
Multiple testing literature contains ample research on controlling false discoveries for hypotheses classified according to one criterion, which we refer to as `one-way classified hypotheses'. However, one often encounters the scenario of `two-way classified hypotheses' where hypotheses can be partitioned into two sets of groups via two different criteria. Associated multiple testing procedures that incorporate such structural information are potentially more effective than their one-way classified or non-classified counterparts. To the best of our knowledge, very little research has been pursued in this direction. This dissertation proposes two types of multiple testing procedures for two-way classified hypotheses. In the first part, we propose a general methodology for controlling the false discovery rate (FDR) using the Benjamini-Hochberg (BH) procedure based on weighted p-values. The weights can be appropriately chosen to reflect one- or two-way classified structure of hypotheses, producing novel multiple testing procedures for two-way classified hypotheses. Newer results for one-way classified hypotheses have been obtained in this process. Our proposed procedures control the false discovery rate (FDR) non-asymptotically in their oracle forms under positive regression dependence on subset of null p-values (PRDS) and in their data-adaptive forms for independent p-values. Simulation studies demonstrate that our proposed procedures can be considerably more powerful than some contemporary methods in many instances and that our data-adaptive procedures can non-asymptotically control the FDR under certain dependent scenarios. The proposed two-way adaptive procedure is applied to a data set from microbial abundance study, for which it makes more discoveries than an existing method. In the second part, we propose a Local false discovery rate (Lfdr) based multiple testing procedure for two-way classified hypotheses. The procedure has been developed in its oracle form under a model based framework that isolates the effects due to two-way grouping from the significance of an individual hypothesis. Simulation studies show that our proposed procedure successfully controls the average proportion of false discoveries, and is more powerful than existing methods.
Temple University--Theses
Stewart, Patrick. „Statistical Inferences on Inflated Data Based on Modified Empirical Likelihood“. Bowling Green State University / OhioLINK, 2020. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1590455262157706.
Der volle Inhalt der QuelleKalaji, Abdul Salam. „Search-based software engineering : a search-based approach for testing from extended finite state machine (EFSM) models“. Thesis, Brunel University, 2010. http://bura.brunel.ac.uk/handle/2438/4575.
Der volle Inhalt der QuelleKurin, Erik, und Adam Melin. „Data-driven test automation : augmenting GUI testing in a web application“. Thesis, Linköpings universitet, Programvara och system, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-96380.
Der volle Inhalt der QuelleHathaway, Drew Aaron. „The use of immersive technologies to improve consumer testing: the impact of multiple immersion levels on data quality and panelist engagement for the evaluation of cookies under a preparation-based scenario“. The Ohio State University, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=osu1448994162.
Der volle Inhalt der QuelleKindstedt, Jonas. „Antibiotic resistance among European strains of Pseudomonas aeruginosa : A study based on resistance data, published articles, and susceptibility testing methods“. Thesis, Umeå universitet, Kemiska institutionen, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-101862.
Der volle Inhalt der QuelleSalles, Lucio Salles de. „Short continuously reinforced concrete pavement design recommendations based on non-destructive ultrasonic data and stress simulation“. Universidade de São Paulo, 2017. http://www.teses.usp.br/teses/disponiveis/3/3138/tde-20102017-082704/.
Der volle Inhalt der QuelleQuatro seções de pavimento de concreto continuamente armado (PCCA) foram construídas no campus da Universidade de São Paulo, com o objetivo de introduzir esta estrutura, de reconhecido sucesso internacional, à comunidade técnica de engenharia de transportes brasileira. As seções foram projetadas com uma placa de concreto de 50 m de extensão, curta em comparação ao PCCA tradicional, com a finalidade de simular paradas e terminais de ônibus - locais de grande interesse para a infraestrutura pública. A tese aqui apresentada conclui este projeto de pesquisa iniciado em 2010. Como o objetivo inicial deste estudo foi o desenvolvimento de recomendações de projeto coerentes, confiáveis e intuitivas para a utilização do PCCA no Brasil, foi necessário um profundo entendimento de suas peculiaridades estruturais e de desempenho. Para isso, o processo de fissuração das secções experimentais foi acompanhado durante sete anos. Devido à curta extensão e falta de ancoragem das seções, o PCCA \"curto\" apresentou um padrão de fissuração diferente do PCCA tradicional com muito menos fissuras visíveis na superfície do que o esperado. Para abordar esta questão, uma nova tecnologia ultrassônica para ensaios não destrutivos de estruturas de concreto foi aplicada. Pela interpretação do sinal de ultrassom, foi possível descobrir várias fissuras incipientes (não visíveis) dentro das placas - muitas dessas foram observadas na superfície da placa em levantamentos de fissuras posteriores - e caracterizar fissuras visíveis e não-visíveis quanto à profundidade da fissura. O mapa de fissuração atualizado com fissuras não visíveis mostrou semelhanças com PCCA tradicional. Além disso, a análise dos dados de ultrassom forneceu informações importantes sobre a variação da espessura, localização da armadura longitudinal e condição do concreto, que foram aplicados em simulações teóricas (software de elementos finitos) do PCCA curto. Simulações foram propostas considerando diferentes geometrias, primeiramente com fissuras transversais como juntas com alta eficiência de transferência de carga (LTE) e posteriormente com uma placa contínua, sem fissuras ou juntas. Esta última simulação foi mais precisa alcançando um fator de conversão entre tensões de campo e simuladas na ordem de 0,7 a 1,0. Dados de deflexão e análise de LTE em fissuras e placas entre fissuras atestaram novamente o comportamento contínuo das placas, o que vai em contradição com os modelos atuais de dimensionamento e de previsão de desempenho para o PCCA. Ademais, o tráfego crítico e condições de carga ambiental correspondentes ao clima e tráfego de ônibus típicos brasileiros foram investigados e relacionados usando um modelo de fadiga resultando em recomendações de projeto para o PCCA de curta extensão sendo direcionado para projetos de longo prazo para mais de 20 anos de operação. O gráfico de projeto foi aplicado com sucesso para investigar três falhas apresentadas pelo PCCA curto experimental devido a deficiências de espessura apontadas pelo teste ultrassônico.
Naňo, Andrej. „Automatické generování testovacích dat informačních systémů“. Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2021. http://www.nusl.cz/ntk/nusl-445520.
Der volle Inhalt der QuellePersson, Jon. „Deterministisk Komprimering/Dekomprimering av Testvektorer med Hjälp av en Inbyggd Processor och Faxkodning“. Thesis, Linköping University, Department of Computer and Information Science, 2005. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-2855.
Der volle Inhalt der QuelleModern semiconductor design methods makes it possible to design increasingly complex system-on-a-chips (SOCs). Testing such SOCs becomes highly expensive due to the rapidly increasing test data volumes with longer test times as a result. Several approaches exist to compress the test stimuli and where hardware is added for decompression. This master’s thesis presents a test data compression method based on a modified facsimile code. An embedded processor on the SOC is used to decompress and apply the data to the cores of the SOC. The use of already existing hardware reduces the need of additional hardware.
Test data may be rearranged in some manners which will affect the compression ratio. Several modifications are discussed and tested. To be realistic a decompressing algorithm has to be able to run on a system with limited resources. With an assembler implementation it is shown that the proposed method can be effectively realized in such environments. Experimental results where the proposed method is applied to benchmark circuits show that the method compares well with similar methods.
A method of including the response vector is also presented. This approach makes it possible to abort a test as soon as an error is discovered, still compressing the data used. To correctly compare the test response with the expected one the data needs to include don’t care bits. The technique uses a mask vector to mark the don’t care bits. The test vector, response vector and mask vector is merged in four different ways to find the most optimal way.
Hensen, Bernadette. „Increasing men's uptake of HIV-testing in sub-Saharan Africa : a systematic review of interventions and analyses of population-based data from rural Zambia“. Thesis, London School of Hygiene and Tropical Medicine (University of London), 2016. http://researchonline.lshtm.ac.uk/2531234/.
Der volle Inhalt der QuelleDoungsa-ard, Chartchai. „Generation of Software Test Data from the Design Specification Using Heuristic Techniques. Exploring the UML State Machine Diagrams and GA Based Heuristic Techniques in the Automated Generation of Software Test Data and Test Code“. Thesis, University of Bradford, 2011. http://hdl.handle.net/10454/5380.
Der volle Inhalt der QuelleSouza, Francisco Carlos Monteiro. „Uma abordagem para geração de dados de teste para o teste de mutação utilizando técnicas baseadas em busca“. Universidade de São Paulo, 2017. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-28092017-162339/.
Der volle Inhalt der QuelleMutation Testing is a powerful test criterion to detect faults and measure the effectiveness of a test data set. However, it is a computationally expensive testing technique. The high cost comes mainly from the effort to generate adequate test data to kill the mutants and by the existence of equivalent mutants. In this thesis, an approach called Reach, Infect and Propagation to Mutation Testing (RIP-MuT) is presented to generate test data and to suggest equivalent mutants. The approach is composed of two modules: (i) an automated test data generation using hill climbing and a fitness scheme according to Reach, Infect, and Propagate (RIP) conditions; and (ii) a method to suggest equivalent mutants based on the analyses of RIP conditions during the process of test data generation. The experiments were conducted to evaluate the effectiveness of the RIP-MuT approach and a comparative study with a genetic algorithm and random testing. The RIP-MuT approach achieved a mean mutation score of 18.25% higher than the GA and 35.93% higher than random testing. The proposed method for detection of equivalent mutants demonstrate to be feasible for cost reduction in this activity since it obtained a precision of 75.05% on suggesting equivalent mutants. Therefore, the results indicate that the approach produces effective test data able to strongly kill the majority of mutants on C programs, and also it can assist in suggesting equivalent mutants correctly.
Ramasamy, Kandasamy Manimozhian. „Efficient state space exploration for parallel test generation“. Thesis, [Austin, Tex. : University of Texas, 2009. http://hdl.handle.net/2152/ETD-UT-2009-05-131.
Der volle Inhalt der QuelleTitle from PDF title page (University of Texas Digital Repository, viewed on August 10, 2009). Includes bibliographical references.
Lindahl, John, und Douglas Persson. „Data-driven test case design of automatic test cases using Markov chains and a Markov chain Monte Carlo method“. Thesis, Malmö universitet, Fakulteten för teknik och samhälle (TS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-43498.
Der volle Inhalt der QuelleNeves, Vânia de Oliveira. „Automatização do teste estrutural de software de veículos autônomos para apoio ao teste de campo“. Universidade de São Paulo, 2015. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-15092015-090805/.
Der volle Inhalt der QuelleIntelligent autonomous vehicle (or just autonomous vehicle - AV) is a type of embedded system that integrates physical (hardware) and computational (software) components. Its main feature is the ability to move and operate partially or fully autonomously. Autonomy grows with the ability to perceive and move within the environment, robustness and ability to solve and perform tasks dealing with different situations (intelligence). Autonomous vehicles represent an important research topic that has a direct impact on society. However, as this field progresses some secondary problems arise, such as how to know if these systems have been sufficiently tested. One of the testing phases of an AV is the field testing, where the vehicle is taken to a controlled environment and it should execute the mission for which it was programed freely. It is generally used to ensure that autonomous vehicles show the intended behavior, but it usually does not take into consideration the code structure. The vehicle (hardware and software) could pass the field testing, but important parts of the code may never have been executed. During the field testing, the input data are collected in logs that can be further analyzed to evaluate the test results and to perform other types of offline tests. This thesis presents a set of proposals to support the analysis of field testing from the point of view of the structural testing. The approach is composed of a class model in the context of the field testing, a tool that implements this model and a genetic algorithm to generate test data. It also shows heuristics to reduce the data set contained in a log without reducing substantially the coverage obtained and combination and mutation strategies that are used in the algorithm. Case studies have been conducted to evaluate the heuristics and strategies, and are also presented and discussed.
Enderlin, Ivan. „Génération automatique de tests unitaires avec Praspel, un langage de spécification pour PHP“. Thesis, Besançon, 2014. http://www.theses.fr/2014BESA2067/document.
Der volle Inhalt der QuelleThe works presented in this memoir are about the validation of PHPprograms through a new specification language, along with its tools. These works follow three axes: specification language, automatic test data generation and automatic unit test generation. The first contribution is Praspel, a new specification language for PHP, based on the Design by Contract. Praspel specifies data with realistic domains, which are new structures allowing to validate and generate data. Based on a contract, we are able to perform Contract-based Testing, i.e.using contracts to automatically generate unit tests. The second contribution isabout test data generation. For booleans, integers and floating point numbers, auniform random generation is used. For arrays, a dedicated constraint solver has been implemented and used. For strings, a grammar description language along with an LL(⋆) compiler compiler and several algorithms for data generation are used. Finally, the object generation is supported. The third contribution defines contract coverage criteria. These latters provide test objectives. All these contributions are implemented and experimented into tools distributed to the PHP community
Olsson, Jakob. „Measuring the Technical and Process Benefits of Test Automation based on Machine Learning in an Embedded Device“. Thesis, KTH, Programvaruteknik och datorsystem, SCS, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-231785.
Der volle Inhalt der QuelleInlärningsbaserad testning är en testningsparadigm som kombinerar model-baserad testning med maskininlärningsalgoritmer för att automatisera systemmodellering, testfallsgenering, exekvering av tester och utfallsbedömning. Ett verktyg som är byggt på LBT är LBTest, utvecklat på CSC skolan på KTH. LBTest nyttjar maskininlärningsalgoritmer med färdiga ekvivalent- och model-checkers, och modellerar användarkrav med linjär temporal logik. I denna studie undersöks det om det är lämpat att använda LBT för att testa en mikrobus arkitektur inom inbyggda telekommunikationsenheter. Utöver det undersöks även hur testprocessen skulle kunna ytterligare automatiseras med hjälp av en data modell för att automatisera generering av användarkrav.
Vitale, Raffaele. „Novel chemometric proposals for advanced multivariate data analysis, processing and interpretation“. Doctoral thesis, Universitat Politècnica de València, 2017. http://hdl.handle.net/10251/90442.
Der volle Inhalt der QuelleLa presente tesis doctoral, concebida principalmente para apoyar y reforzar la relación entre la academia y la industria, se desarrolló en colaboración con Shell Global Solutions (Amsterdam, Países Bajos) en el esfuerzo de aplicar y posiblemente extender los enfoques ya consolidados basados en variables latentes (es decir, Análisis de Componentes Principales - PCA - Regresión en Mínimos Cuadrados Parciales - PLS - o PLS discriminante - PLSDA) para la resolución de problemas complejos no sólo en los campos de mejora y optimización de procesos, sino también en el entorno más amplio del análisis de datos multivariados. Con este fin, en todos los capítulos proponemos nuevas soluciones algorítmicas eficientes para abordar tareas dispares, desde la transferencia de calibración en espectroscopia hasta el modelado en tiempo real de flujos de datos. El manuscrito se divide en las seis partes siguientes, centradas en diversos temas de interés: Parte I - Prefacio, donde presentamos un resumen de este trabajo de investigación, damos sus principales objetivos y justificaciones junto con una breve introducción sobre PCA, PLS y PLSDA; Parte II - Sobre las extensiones basadas en kernels de PCA, PLS y PLSDA, donde presentamos el potencial de las técnicas de kernel, eventualmente acopladas a variantes específicas de la recién redescubierta proyección de pseudo-muestras, formulada por el estadista inglés John C. Gower, y comparamos su rendimiento respecto a metodologías más clásicas en cuatro aplicaciones a escenarios diferentes: segmentación de imágenes Rojo-Verde-Azul (RGB), discriminación y monitorización de procesos por lotes y análisis de diseños de experimentos de mezclas; Parte III - Sobre la selección del número de factores en el PCA por pruebas de permutación, donde aportamos una guía extensa sobre cómo conseguir la selección de componentes de PCA mediante pruebas de permutación y una ilustración completa de un procedimiento algorítmico original implementado para tal fin; Parte IV - Sobre la modelización de fuentes de variabilidad común y distintiva en el análisis de datos multi-conjunto, donde discutimos varios aspectos prácticos del análisis de componentes comunes y distintivos de dos bloques de datos (realizado por métodos como el Análisis Simultáneo de Componentes - SCA - Análisis Simultáneo de Componentes Distintivos y Comunes - DISCO-SCA - Descomposición Adaptada Generalizada de Valores Singulares - Adapted GSVD - ECO-POWER, Análisis de Correlaciones Canónicas - CCA - y Proyecciones Ortogonales de 2 conjuntos a Estructuras Latentes - O2PLS). Presentamos a su vez una nueva estrategia computacional para determinar el número de factores comunes subyacentes a dos matrices de datos que comparten la misma dimensión de fila o columna y dos planteamientos novedosos para la transferencia de calibración entre espectrómetros de infrarrojo cercano; Parte V - Sobre el procesamiento y la modelización en tiempo real de flujos de datos de alta dimensión, donde diseñamos la herramienta de Procesamiento en Tiempo Real (OTFP), un nuevo sistema de manejo racional de mediciones multi-canal registradas en tiempo real; Parte VI - Epílogo, donde presentamos las conclusiones finales, delimitamos las perspectivas futuras, e incluimos los anexos.
La present tesi doctoral, concebuda principalment per a recolzar i reforçar la relació entre l'acadèmia i la indústria, es va desenvolupar en col·laboració amb Shell Global Solutions (Amsterdam, Països Baixos) amb l'esforç d'aplicar i possiblement estendre els enfocaments ja consolidats basats en variables latents (és a dir, Anàlisi de Components Principals - PCA - Regressió en Mínims Quadrats Parcials - PLS - o PLS discriminant - PLSDA) per a la resolució de problemes complexos no solament en els camps de la millora i optimització de processos, sinó també en l'entorn més ampli de l'anàlisi de dades multivariades. A aquest efecte, en tots els capítols proposem noves solucions algorítmiques eficients per a abordar tasques dispars, des de la transferència de calibratge en espectroscopia fins al modelatge en temps real de fluxos de dades. El manuscrit es divideix en les sis parts següents, centrades en diversos temes d'interès: Part I - Prefaci, on presentem un resum d'aquest treball de recerca, es donen els seus principals objectius i justificacions juntament amb una breu introducció sobre PCA, PLS i PLSDA; Part II - Sobre les extensions basades en kernels de PCA, PLS i PLSDA, on presentem el potencial de les tècniques de kernel, eventualment acoblades a variants específiques de la recentment redescoberta projecció de pseudo-mostres, formulada per l'estadista anglés John C. Gower, i comparem el seu rendiment respecte a metodologies més clàssiques en quatre aplicacions a escenaris diferents: segmentació d'imatges Roig-Verd-Blau (RGB), discriminació i monitorització de processos per lots i anàlisi de dissenys d'experiments de mescles; Part III - Sobre la selecció del nombre de factors en el PCA per proves de permutació, on aportem una guia extensa sobre com aconseguir la selecció de components de PCA a través de proves de permutació i una il·lustració completa d'un procediment algorítmic original implementat per a la finalitat esmentada; Part IV - Sobre la modelització de fonts de variabilitat comuna i distintiva en l'anàlisi de dades multi-conjunt, on discutim diversos aspectes pràctics de l'anàlisis de components comuns i distintius de dos blocs de dades (realitzat per mètodes com l'Anàlisi Simultània de Components - SCA - Anàlisi Simultània de Components Distintius i Comuns - DISCO-SCA - Descomposició Adaptada Generalitzada en Valors Singulars - Adapted GSVD - ECO-POWER, Anàlisi de Correlacions Canòniques - CCA - i Projeccions Ortogonals de 2 blocs a Estructures Latents - O2PLS). Presentem al mateix temps una nova estratègia computacional per a determinar el nombre de factors comuns subjacents a dues matrius de dades que comparteixen la mateixa dimensió de fila o columna, i dos plantejaments nous per a la transferència de calibratge entre espectròmetres d'infraroig proper; Part V - Sobre el processament i la modelització en temps real de fluxos de dades d'alta dimensió, on dissenyem l'eina de Processament en Temps Real (OTFP), un nou sistema de tractament racional de mesures multi-canal registrades en temps real; Part VI - Epíleg, on presentem les conclusions finals, delimitem les perspectives futures, i incloem annexos.
Vitale, R. (2017). Novel chemometric proposals for advanced multivariate data analysis, processing and interpretation [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/90442
TESIS
Johann, Matthew A. „Fire-Robust Structural Engineering: A Framework Approach to Structural Design for Fire Conditions“. Link to electronic thesis, 2002. http://www.wpi.edu/Pubs/ETD/Available/etd-1219102-155849.
Der volle Inhalt der QuelleKeywords: structural engineering; fire safety; framework approach; performance-based design; information management; finite element; lumped-parameter; laboratory tests; steel; beam; restrained; plastic analysis. Includes bibliographical references (p. 180-182).
Su, Weizhe. „Bayesian Hidden Markov Model in Multiple Testing on Dependent Count Data“. University of Cincinnati / OhioLINK, 2020. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1613751403094066.
Der volle Inhalt der QuelleOlofsson, Louise. „Sustainable and scalable testing strategy in a multilayered service-based architecture“. Thesis, Luleå tekniska universitet, Institutionen för system- och rymdteknik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:ltu:diva-80134.
Der volle Inhalt der QuelleEmer, Maria Claudia Figueiredo Pereira. „Abordagem de teste baseada em defeitos para esquemas de dados“. [s.n.], 2007. http://repositorio.unicamp.br/jspui/handle/REPOSIP/261002.
Der volle Inhalt der QuelleTese (doutorado) - Universidade Estadual de Campinas, Faculdade de Engenharia Eletrica e de Computação
Made available in DSpace on 2018-08-09T21:13:12Z (GMT). No. of bitstreams: 1 Emer_MariaClaudiaFigueiredoPereira_D.pdf: 934024 bytes, checksum: dbb2079115b56358ff3dc9e747df6386 (MD5) Previous issue date: 2007
Resumo: Dados são manipulados em várias aplicações de software envolvendo operações críticas. Em tais aplicações assegurar a qualidade dos dados manipulados é fundamental. Esquemas de dados definem a estrutura lógica e os relacionamentos entre os dados. O teste de esquemas por meio de abordagens, critérios e ferramentas de teste específicos é uma forma pouco explorada de assegurar a qualidade de dados definidos por esquemas. Este trabalho propõe uma abordagem de teste baseada em classes de defeitos comumente identificados em esquemas de dados. Um metamodelo de dados é definido para especificar os esquemas que podem ser testados e as restrições aos dados nos esquemas. Defeitos possíveis de serem revelados são os relacionados à definição incorreta ou ausente de restrições aos dados no esquema. A abordagem inclui a geração automática de um conjunto de teste que contém instâncias de dados e consultas a essas instâncias; as instâncias de dados e as consultas são geradas de acordo com padrões definidos em cada classe de defeito. Experimentos nos contextos de aplicações Web e de base de dados foram realizados para ilustrar a aplicação da abordagem
Abstract: Data are used in several software applications involving critical operations. In such applications to ensure the quality of the manipulated data is fundamental. Data schemas define the logical structure and the relationships among data. Testing schemas by means of specific testing approaches, criteria and tools has not been explored adequately as a way to ensure the quality of data defined by schemas. This work proposes a testing approach based on fault classes usually identified in data schemas. A data metamodel is defined to specify the schemas that can be tested and the constraints to the data in schemas. This testing approach provides means for revealing faults related to incorrect or absent definition of constraints for the data in the schema. The approach includes the automatic generation of a test set which contains data instances and queries to these instances; the data instances and queries are generated according to patterns defined in each fault class. Experiments in the contexts of Web and database applications were carried out to illustrate the testing approach application
Doutorado
Engenharia de Computação
Doutor em Engenharia Elétrica
CARVALHO, Gustavo Henrique Porto de. „NAT2TEST: generating test cases from natural language requirements based on CSP“. Universidade Federal de Pernambuco, 2016. https://repositorio.ufpe.br/handle/123456789/17929.
Der volle Inhalt der QuelleMade available in DSpace on 2016-09-28T12:33:15Z (GMT). No. of bitstreams: 2 license_rdf: 1232 bytes, checksum: 66e71c371cc565284e70f40736c94386 (MD5) GustavoHPCarvalho_Doutorado_CInUFPE_2016.pdf: 1763137 bytes, checksum: aed7b3ab2f6235757818003678633c9b (MD5) Previous issue date: 2016-02-26
High trustworthiness levels are usually required when developing critical systems, and model based testing (MBT) techniques play an important role generating test cases from specification models. Concerning critical systems, these models are usually created using formal or semi-formal notations. Moreover, it is also desired to clearly and formally state the conditions necessary to guarantee that an implementation is correct with respect to its specification by means of a conformance relation, which can be used to prove that the test generation strategy is sound. Despite the benefits of MBT, those who are not familiar with the models syntax and semantics may be reluctant to adopt these formalisms. Furthermore, most of these models are not available in the very beginning of the project, when usually natural-language requirements are available. Therefore, the use of MBT is postponed. Here, we propose an MBT strategy for generating test cases from controlled naturallanguage (CNL) requirements: NAT2TEST, which refrains the user from knowing the syntax and semantics of the underlying notations, besides allowing early use of MBT via naturallanguage processing techniques; the formal and semi-formal models internally used by our strategy are automatically generated from the natural-language requirements. Our approach is tailored to data-flow reactive systems: a class of embedded systems whose inputs and outputs are always available as signals. These systems can also have timed-based behaviour, which may be discrete or continuous. The NAT2TEST strategy comprises a number of phases. Initially, the requirements are syntactically analysed according to a CNL we proposed to describe data-flow reactive systems. Then, the requirements informal semantics are characterised based on the case grammar theory. Afterwards, we derive a formal representation of the requirements considering a model of dataflow reactive systems we defined. Finally, this formal model is translated into communicating sequential processes (CSP) to provide means for generating test cases. We prove that our test generation strategy is sound with respect to our timed input-output conformance relation based on CSP: csptio. Besides CSP, we explore the generation of other target notations (SCR and IMR) from which we can generate test cases using commercial tools (T-VEC and RT-Tester, respectively). The whole process is fully automated by the NAT2TEST tool. Our strategy was evaluated considering examples from the literature, the aerospace (Embraer) and the automotive (Mercedes) industry. We analysed performance and the ability to detect defects generated via mutation. In general, our strategy outperformed the considered baseline: random testing. We also compared our strategy with relevant commercial tools.
Testes baseados em modelos (MBT) consiste em criar modelos para especificar o comportamento esperado de sistemas e, a partir destes, gerar testes que verificam se implementações possuem o nível de confiabilidade esperado. No contexto de sistemas críticos, estes modelos são normalmente (semi)formais e deseja-se uma definição precisa das condições necessárias para garantir que uma implementação é correta em relação ao modelo da especificação. Esta definição caracteriza uma relação de conformidade, que pode ser usada para provar que uma estratégia de MBT é consistente (sound). Apesar dos benefícios, aqueles sem familiaridade com a sintaxe e a semântica dos modelos empregados podem relutar em adotar estes formalismos. Aqui, propõe-se uma estratégia de MBT para gerar casos de teste a partir de linguagem natural controlada (CNL). Esta estratégia (NAT2TEST) dispensa a necessidade de conhecer a sintaxe e a semântica das notações formais utilizadas internamente, uma vez que os modelos intermediários são gerados automaticamente a partir de requisitos em linguagem natural. Esta estratégia é apropriada para sistemas reativos baseados em fluxos de dados: uma classe de sistemas embarcados cujas entradas e saídas estão sempre disponíveis como sinais. Estes sistemas também podem ter comportamento dependente do tempo (discreto ou contínuo). Na estratégia NAT2TEST, inicialmente, os requisitos são analisados sintaticamente de acordo com a CNL proposta neste trabalho para descrever sistemas reativos. Em seguida, a semântica informal dos requisitos é caracterizada utilizando a teoria de gramática de casos. Posteriormente, deriva-se uma representação formal dos requisitos considerando um modelo definido neste trabalho para sistemas reativos. Finalmente, este modelo é traduzido em uma especificação em communicating sequential processes (CSP) para permitir a geração de testes. Este trabalho prova que a estratégia de testes proposta é consistente considerando a relação de conformidade temporal baseada em entradas e saídas também definida aqui: csptio. Além de CSP, foi explorada a geração de outras notações formais (SCR e IMR), a partir das quais é possível gerar casos de teste usando ferramentas comerciais (T-VEC e RT-Tester, respectivamente). Todo o processo é automatizado pela ferramenta NAT2TEST. A estratégia NAT2TEST foi avaliada considerando exemplos da literatura, da indústria aeroespacial (Embraer) e da automotiva (Mercedes). Foram analisados o desempenho e a capacidade de detectar defeitos gerados através de operadores de mutação. Em geral, a estratégia NAT2TEST apresentou melhores resultados do que a referência adotada: testes aleatórios. A estratégia NAT2TEST também foi comparada com ferramentas comerciais relevantes.
ITO, Hideo, und Gang ZENG. „Low-Cost IP Core Test Using Tri-Template-Based Codes“. Institute of Electronics, Information and Communication Engineers, 2007. http://hdl.handle.net/2237/15029.
Der volle Inhalt der QuelleSadeghzadeh, Seyedehsaloumeh. „Optimal Data-driven Methods for Subject Classification in Public Health Screening“. Diss., Virginia Tech, 2019. http://hdl.handle.net/10919/101611.
Der volle Inhalt der QuelleDoctor of Philosophy
馮可達 und Ho-tat Fung. „Soil property determination through a knowledge-based system with emphasis on undrained shear strength“. Thesis, The University of Hong Kong (Pokfulam, Hong Kong), 1997. http://hub.hku.hk/bib/B31236868.
Der volle Inhalt der QuelleStrålfors, Annika. „Making test automation sharable: The design of a generic test automation framework for web based applications“. Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-240981.
Der volle Inhalt der QuelleZhang, Zhidong 1957. „Cognitive assessment in a computer-based coaching environment in higher education : diagnostic assessment of development of knowledge and problem-solving skill in statistics“. Thesis, McGill University, 2007. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=102853.
Der volle Inhalt der QuelleThe objective of this study was to develop a Bayesian assessment model that implements DCA in a specific domain of statistics, and evaluate it in relation to its potential to achieve the objectives of DCA. This study applied a method for model development to the ANOVA score model domain to attain the objectives of the study. The results documented: (a) the process of model development in a specific domain; (b) the properties of the Bayesian assessment model; (c) the performance of the network in tracing students' progress towards mastery by using the model to successfully update the posterior probabilities; (d) the use of estimates of log odds ratios of likelihood of mastery as a measure of "progress toward mastery;" (e) the robustness of diagnostic inferences based on the network; and (f) the use of the Bayesian assessment model for diagnostic assessment with a sample of 20 students who completed the assessment tasks. The results indicated that the Bayesian assessment network provided valid diagnostic information about specific cognitive components, and was able to track development towards achieving mastery of learning goals.
Trondman, Anna-Kari. „Pollen-based quantitative reconstruction of land-cover change in Europe from 11,500 years ago until present - A dataset suitable for climate modelling“. Doctoral thesis, Linnéuniversitetet, Institutionen för biologi och miljö (BOM), 2014. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-40775.
Der volle Inhalt der QuelleMoussa, Ahmed S. „On learning and visualizing lexicographic preference trees“. UNF Digital Commons, 2019. https://digitalcommons.unf.edu/etd/882.
Der volle Inhalt der QuelleRubbestad, Gustav, und William Söderqvist. „Hacking a Wi-Fi based drone“. Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-299887.
Der volle Inhalt der QuelleDrönare, eller UAV från engelskans Unmanned Aerial Vehicle, har ökat i popularitet bland privatpersoner sedan tidigt 2000tal. En majoritet av drönare för nybörjare är baserade på WiFi och styrs med en kontroll som vanligtvis är en smart phone. Detta innebär att dessa drönare kan vara sårbara för olika typer av attacker på nätverket, vilket utvärderas och testas i denna rapport på drönaren Ryze Tello. Flera hot identifierades med hotmodellering och ett flertal valdes ut för penetrationtest. Detta genomförs med syftet att svara på forskningsfrågan: Hur sårbar är Ryze Tello mot WiFi baserade attacker? Svaret på forskningsfrågan är att drönaren Ryze Tello är relativt säker, med undantaget att den inte har ett standardlösenord. Ett lösenord sattes på nätverket, men lösenordet knäcktes ändå med en ordboksattack. Detta möjliggjorde attacker så som instruktionsinjicering och förmågan att se videoströmmen från drönaren samtidigt som den kan kontrolleras via kommandon i terminalen.
Lindberg, Tobias. „A/B-testing for web design: A comparative study of response times between MySQL and PostgreSQL : Implementation of a web based tool for design comparisons with stored images“. Thesis, Högskolan i Skövde, Institutionen för informationsteknologi, 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:his:diva-15409.
Der volle Inhalt der QuelleDe, Voir Christopher S. „Wavelet Based Feature Extraction and Dimension Reduction for the Classification of Human Cardiac Electrogram Depolarization Waveforms“. PDXScholar, 2005. https://pdxscholar.library.pdx.edu/open_access_etds/1740.
Der volle Inhalt der QuelleLuo, Dan, und Yajing Ran. „Micro Drivers behind the Changes of CET1 Capital Ratio : An empirical analysis based on the results of EU-wide stress test“. Thesis, Internationella Handelshögskolan, Högskolan i Jönköping, IHH, Företagsekonomi, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:hj:diva-44140.
Der volle Inhalt der QuelleJara-Almonte, J. „Extraction of eigen-pairs from beam structures using an exact element based on a continuum formulation and the finite element method“. Diss., Virginia Polytechnic Institute and State University, 1985. http://hdl.handle.net/10919/54300.
Der volle Inhalt der QuellePh. D.
Potnuru, Srinath. „Fuzzing Radio Resource Control messages in 5G and LTE systems : To test telecommunication systems with ASN.1 grammar rules based adaptive fuzzer“. Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2021. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-294140.
Der volle Inhalt der Quelle5G-telekommunikationssystem måste vara extremt tillförlitliga för att möta behoven för den kommande utvecklingen inom kommunikation. Systemen som används måste testas noggrant och måste överensstämma med deras standarder. Programvara och nätverksprotokoll testas ofta med tekniker som fuzzing, penetrationstest, kodgranskning, testning av överensstämmelse. Med fuzzing kan testare skicka utformade input för att övervaka System Under Test (SUT) för ett svar. 3GPP, standardiseringsorganet för telekomsystemet, producerar ofta nya versioner av specifikationer för att möta kraven och bristerna från tidigare utgåvor. Detta leder till många versioner av specifikationer för ett nätverksprotokoll som Radio Resource Control (RRC) och testare behöver ständigt uppdatera testverktygen och testmiljön. I detta arbete visar vi att genom att använda den generiska karaktären av RRC-specifikationer, som ges i beskrivningsspråket Abstract Syntax Notation One (ASN.1), kan man designa ett testverktyg för att anpassa sig till alla versioner av 3GPP-specifikationer. Detta uppsatsarbete introducerar en ASN.1-baserad adaptiv fuzzer som kan användas för att testa RRC och andra nätverksprotokoll baserat på ASN.1- beskrivningsspråk. Fuzzer extraherar kunskap om pågående RRC meddelanden med användning av protokollbeskrivningsfiler för RRC, dvs RRC ASN.1 schema från 3GPP, och använder kunskapen för att fuzz RRC meddelanden. Den adaptiva fuzzer identifierar enskilda fält, delmeddelanden och anpassade datatyper enligt specifikationer när innehållet i befintliga meddelanden muteras. Dessutom har den adaptiva fuzzer identifierat en tidigare oidentifierad sårbarhet i Evolved Packet Core (EPC) för srsLTE och openLTE, två opensource LTE-implementeringar, vilket bekräftar tillämpligheten för robusthetsprovning av RRC och andra nätverksprotokoll.