Tesis sobre el tema "Formal and symbolic calculation"
Crea una cita precisa en los estilos APA, MLA, Chicago, Harvard y otros
Consulte los 50 mejores tesis para su investigación sobre el tema "Formal and symbolic calculation".
Junto a cada fuente en la lista de referencias hay un botón "Agregar a la bibliografía". Pulsa este botón, y generaremos automáticamente la referencia bibliográfica para la obra elegida en el estilo de cita que necesites: APA, MLA, Harvard, Vancouver, Chicago, etc.
También puede descargar el texto completo de la publicación académica en formato pdf y leer en línea su resumen siempre que esté disponible en los metadatos.
Explore tesis sobre una amplia variedad de disciplinas y organice su bibliografía correctamente.
Vu, Thi Xuan. "Homotopy algorithms for solving structured determinantal systems". Electronic Thesis or Diss., Sorbonne université, 2020. http://www.theses.fr/2020SORUS478.
Texto completoMultivariate polynomial systems arising in numerous applications have special structures. In particular, determinantal structures and invariant systems appear in a wide range of applications such as in polynomial optimization and related questions in real algebraic geometry. The goal of this thesis is to provide efficient algorithms to solve such structured systems. In order to solve the first kind of systems, we design efficient algorithms by using the symbolic homotopy continuation techniques. While the homotopy methods, in both numeric and symbolic, are well-understood and widely used in polynomial system solving for square systems, the use of these methods to solve over-detemined systems is not so clear. Meanwhile, determinantal systems are over-determined with more equations than unknowns. We provide probabilistic homotopy algorithms which take advantage of the determinantal structure to compute isolated points in the zero-sets of determinantal systems. The runtimes of our algorithms are polynomial in the sum of the multiplicities of isolated points and the degree of the homotopy curve. We also give the bounds on the number of isolated points that we have to compute in three contexts: all entries of the input are in classical polynomial rings, all these polynomials are sparse, and they are weighted polynomials. In the second half of the thesis, we deal with the problem of finding critical points of a symmetric polynomial map on an invariant algebraic set. We exploit the invariance properties of the input to split the solution space according to the orbits of the symmetric group. This allows us to design an algorithm which gives a triangular description of the solution space and which runs in time polynomial in the number of points that we have to compute. Our results are illustrated by applications in studying real algebraic sets defined by invariant polynomial systems by the means of the critical point method
Krandick, Werner. "Symbolic methods for polynomial complex root calculation /". The Ohio State University, 1992. http://rave.ohiolink.edu/etdc/view?acc_num=osu1487776210796097.
Texto completoQian, Kairong Computer Science & Engineering Faculty of Engineering UNSW. "Formal symbolic verification using heuristic search and abstraction techniques". Awarded by:University of New South Wales. School of Computer Science and Engineering, 2006. http://handle.unsw.edu.au/1959.4/25703.
Texto completoRitter, Gerd. "Formal sequential equivalence checking of digital systems by symbolic simulation". Phd thesis, [S.l.] : [s.n.], 2001. http://elib.tu-darmstadt.de/diss/000113/thesis.pdf.
Texto completoKavish, Daniel Ryan. "Interactionist Labeling: Formal and Informal Labeling's Effects on Juvenile Delinquency". OpenSIUC, 2012. https://opensiuc.lib.siu.edu/theses/883.
Texto completoKavish, Daniel Ryan. "Interactionist Labeling: A Structural Equation Model of Formal Labeling, Juvenile Delinquency, and Adult Criminality". OpenSIUC, 2016. https://opensiuc.lib.siu.edu/dissertations/1311.
Texto completoMorrison, George Campbell. "Automated coverage calculation and test case generation". Thesis, Stellenbosch : Stellenbosch University, 2012. http://hdl.handle.net/10019.1/20041.
Texto completoENGLISH ABSTRACT: This research combines symbolic execution, a formal method of static analysis, with various test adequacy criteria, to explore the e ectiveness of using symbolic execution for calculating code coverage on a program's existing JUnit test suites. Code coverage is measured with a number of test adequacy criteria, including statement coverage, branch coverage, condition coverage, method coverage, class coverage, and loop coverage. The results of the code coverage calculation is then used to automatically generate JUnit test cases for areas of a program that are not su ciently covered. The level of redundancy of each test case is also calculated during coverage calculation, thereby identifying fully redundant, and partially redundant, test cases. The combination of symbolic execution and code coverage calculation is extended to perform coverage calculation during a manual execution of a program, allowing testers to measure the e ectiveness of manual testing. This is implemented as an Eclipse plug-in, named ATCO, which attempts to take advantage of the Eclipse workspace and extensible user interface environment to improve usability of the tool by minimizing the user interaction required to use the tool. The code coverage calculation process uses constraint solving to determine method parameter values to reach speci c areas in the program. Constraint solving is an expensive computation, so the tool was parallellised using Java's Concurrency package, to reduce the overall execution time of the tool.
AFRIKAANSE OPSOMMING: Hierdie navorsing kombineer simboliese uitvoering, 'n formele metode van statiese analise, met verskeie toets genoegsaamheid kriteria, om die e ektiwiteit van die gebruik van simboliese uitvoer te ondersoek vir die berekening van kode dekking op 'n program se bestaande JUnit toets stelle. Kode dekking word gemeet deur verskeie toets genoegsaamheid kriteria, insluited stelling dekking, tak dekking, kondisie dekking, metode dekking, klas dekking, en lus dekking. Die resultate van die kode dekking berekeninge word dan gebruik om outomaties JUnit toets voorbeelde te genereer vir areas van 'n program wat nie doeltre end ondersoek word nie. Die vlak van oortolligheid van elke toets voorbeeld word ook bereken gedurende die dekkingsberekening, en daardeur word volledig oortollige, en gedeeltelik oortollige, toets voorbeelde identi seer. Die kombinasie van simboliese uitvoer en kode dekking berekening is uitgebrei deur die uitvoer van dekking berekeninge van 'n gebruiker-beheerde uitvoer, om sodoende kode dekking van 'n gebruiker-beheerde uitvoer van 'n program te meet. Dit laat toetsers toe om die e ektiwiteit van hulle beheerde uitvoer te meet. Bogenoemde word ge mplimenteer as 'n Eclipse aanvoegsel, genaamd ATCO, wat poog om voordeel te trek vanuit die Eclipse werkspasie, en die uitbreibare gebruiker oordrag omgewing, om die bruikbaarheid van ATCO te verbeter, deur die vermindering van die gebruiker interaksie wat benodig word om ATCO te gebruik. Die kode dekking berekeningsproses gebruik beperking oplossing om metode invoer waardes te bereken, om spesi eke areas in die program te bereik. Beperking oplossing is 'n duur berekening, so ATCO is geparalleliseer, met behulp van Java se Concurrency pakket, om die algehele uitvoer tyd van die program te verminder.
Klein, Joachim, Christel Baier, Philipp Chrszon, Marcus Daum, Clemens Dubslaff, Sascha Klüppelholz, Steffen Märcker y David Müller. "Advances in Symbolic Probabilistic Model Checking with PRISM". Springer, 2016. https://tud.qucosa.de/id/qucosa%3A74267.
Texto completoZhao, Hong. "Automatic generation and reduction of the semi-fuzzy knowledge base in symbolic processing and numerical calculation". Thesis, National Library of Canada = Bibliothèque nationale du Canada, 1995. http://www.collectionscanada.ca/obj/s4/f2/dsk3/ftp04/NQ27811.pdf.
Texto completoHansen, Sonja Maria [Verfasser], Hilde [Gutachter] Haider y Robert [Gutachter] Gaschler. "The potential of symbolic approximation. Disentangling the effects of approximation vs. calculation demands in nonsymbolic and symbolic representations. / Sonja Maria Hansen ; Gutachter: Hilde Haider, Robert Gaschler". Köln : Universitäts- und Stadtbibliothek Köln, 2016. http://d-nb.info/1121745261/34.
Texto completoBrown, Douglas Graeme. "Formal network behaviour analysis using model checking". Thesis, Queensland University of Technology, 2016. https://eprints.qut.edu.au/93693/1/Douglas_Brown_Thesis.pdf.
Texto completoCampo, Anna Laetitia a. "Anthropomorphic representations in prehistoric Cyprus : a formal and symbolic analysis of figurines, c. 3500 - 1800 BC". Thesis, University of Cambridge, 1987. https://www.repository.cam.ac.uk/handle/1810/272955.
Texto completoSarkar, Somwrita. "Acquiring symbolic design optimization problem reformulation knowledge". Connect to full text, 2009. http://hdl.handle.net/2123/5683.
Texto completoTitle from title screen (viewed November 13, 2009). Submitted in fulfilment of the requirements for the degree of Doctor of Philosophy to the Faculty of Architecture, Design and Planning in the Faculty of Science. Includes graphs and tables. Includes bibliographical references. Also available in print form.
Adams, Sara Elisabeth. "Abstraction discovery and refinement for model checking by symbolic trajectory evaluation". Thesis, University of Oxford, 2014. http://ora.ox.ac.uk/objects/uuid:27276f9c-eba5-42a9-985d-1812097773f8.
Texto completoKephart, David E. "Topology, morphisms, and randomness in the space of formal languages". [Tampa, Fla.] : University of South Florida, 2005. http://purl.fcla.edu/fcla/etd/SFE0001250.
Texto completoAïssat, Romain. "Infeasible Path Detection : a Formal Model and an Algorithm". Thesis, Université Paris-Saclay (ComUE), 2017. http://www.theses.fr/2017SACLS036/document.
Texto completoWhite-box, path-based, testing is largely used for the validation of programs. Given the control-flow graph (CFG) of the program under test, a test suit is generated by selecting a collection of paths of interest, then trying to provide, for each path, some concrete input values that will make the program follow that path during a run.For the first step, there are various ways to define paths of interest: structural testing methods select some set of paths that fulfills coverage criteria related to elements of the graph; in random-based techniques, paths are selected according to a given distribution of probability over these elements (for instance, uniform probability over all paths of length less than a given bound). Both approaches can be combined as in structural statistical testing. The random-based methods above have the advantage of providing a way to assess the quality of a test set as the minimal probability of covering an element of a criterion.The second step requires to produce for each path its path predicate, i.e. the conjunction of the constraints over the input parameters that must hold for the system to run along that path. This is done using symbolic execution. Then, constraint-solving is used to compute test data. If there is no input values such that the path predicate evaluates to true, the path is infeasible. It is very common for a program to have infeasible paths and such paths can largely outnumber feasible paths. Infeasible paths selected during the first step will not contribute to the final test suite, and there is no better choice than to select another path, hoping for its feasibility. Handling infeasible paths is the serious limitation of structural methods since most of the time is spent selecting useless paths. It is also a major challenge for all techniques in static analysis of programs, since the quality of the approximations they provide is lowered by data computed for paths that do not correspond to actual program runs.To overcome this problem, different methods have been proposed, like concolic testing or random testing based on the input domain. In path-biased random testing, paths are drawn according to a given distribution and their feasibility is checked in a second step. We present an algorithm that builds better approximations of the behavior of a program than its CFG, providing a transformed CFG, which still over-approximates the set of feasible paths but with fewer infeasible paths. This transformed graph is used for drawing paths at random.We modeled our graph transformations and formally proved, using the interactive theorem proving environment Isabelle/HOL, the key properties that establish the correctness of our approach.Our algorithm uses symbolic execution and constraint solving, which allows to detect whether some paths are infeasible. Since programs can contain loops, their graphs can contain cycles. In order to avoid to follow infinitely a cyclic path, we enrich symbolic execution with the detection of subsumptions. A subsumption can be interpreted as the fact that some node met during the analysis is a particular case of another node met previously: there is no need to explore the successors of the subsumed node: they are subsumed by the successors of the subsumer. Our algorithm has been implemented by a prototype, whose design closely follows said formalization, giving a good level of confidence in its correctness.In this thesis, we introduce the theoretical concepts on which our approach relies, its formalization in Isabelle/HOL, the algorithms our prototype implements and the various experiments done and results obtained using it
Konecny, Jan. "Isotone fuzzy Galois connections and their applications in formal concept analysis". Diss., Online access via UMI:, 2009.
Buscar texto completoIncludes bibliographical references.
Nyström, Jan Henry. "Analysing Fault Tolerance for Erlang Applications". Doctoral thesis, Uppsala universitet, Avdelningen för datorteknik, 2009. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-101975.
Texto completoASTEC
Ivanova, Elena. "Efficient Synthesis of Safety Controllers using Symbolic Models and Lazy Algorithms". Electronic Thesis or Diss., université Paris-Saclay, 2021. http://www.theses.fr/2021UPASG088.
Texto completoThis thesis focuses on the development of efficient abstraction-based controller synthesis approaches for cyber-physical systems (CPS). While abstraction-based methods for CPS design have been the subject of intensive research over the last decades, the scalability of these techniques remains an issue. This thesis focus on developing lazy synthesis algorithms for safety specifications. Safety specifications consist in maintaining the trajectory of the system inside a given safe set. This specification is of the utmost importance in many engineering problems, often prioritized over other performance requirements. Lazy approaches outperform the classical synthesis algorithm [Tabuada, 2009] by avoiding computations, which are non-essential for synthesis goals. Chapter 1 motivates the thesis and discusses the state of the art. Chapter 2 structures the existing lazy synthesis approaches and emphasizes three sources of efficiency: information about a priori controllable states, priorities on inputs, and non-reachable from initial set states. Chapter 3 proposes an algorithm, which iteratively explores states on the boundary of controllable domain while avoiding exploration of internal states, supposing that they are safely controllable a priory. A closed-loop safety controller for the original problem is then defined as follows: we use the abstract controller to push the system from a boundary state back towards the interior, while for inner states, any admissible input is valid. Chapter 4 presents an algorithm that restricts the controller synthesis computations to reachable states only while prioritizing longer-duration transitions. The original system is abstracted by a symbolic model with an adaptive grid. Moreover, a novel type of time sampling is also considered. Instead of using transitions of predetermined duration, the duration of the transitions is constrained by state intervals that must contain the reachable set. Chapter 5 is dedicated to monotone transition systems. The introduced lazy synthesis approach benefits from a monotone property of transition systems and the ordered structure of the state (input) space, and the fact that directed safety specifications are considered. The considered class of specifications is then enriched by intersections of upper and lower-closed safety requirements. Chapter 6 concludes the discussion and raises new issues for future research
David, Robin. "Formal Approaches for Automatic Deobfuscation and Reverse-engineering of Protected Codes". Thesis, Université de Lorraine, 2017. http://www.theses.fr/2017LORR0013/document.
Texto completoMalware analysis is a growing research field due to the criticity and variety of assets targeted as well as the increasing implied costs. These softwares frequently use evasion tricks aiming at hindering detection and analysis techniques. Among these, obfuscation intent to hide the program behavior. This thesis studies the potential of Dynamic Symbolic Execution (DSE) for reverse-engineering. First, we propose two variants of DSE algorithms adapted and designed to fit on protected codes. The first is a flexible definition of the DSE path predicate computation based on concretization and symbolization. The second is based on the definition of a backward-bounded symbolic execution algorithm. Then, we show how to combine these techniques with static analysis in order to get the best of them. Finally, these algorithms have been implemented in different tools Binsec/se, Pinsec and Idasec interacting alltogether and tested on several malicious codes and commercial packers. Especially, they have been successfully used to circumvent and remove the obfuscation targeted in real-world malwares like X-Tunnel from the famous APT28/Sednit group
Debant, Alexandre. "Symbolic verification of distance-bounding protocols : application to payment protocols". Thesis, Rennes 1, 2020. http://www.theses.fr/2020REN1S057.
Texto completoThe rise of new technologies, and in particular Near Field Communication (NFC) tags, offers new applications such as contactless payments, key-less entry systems, transport ticketing... Due to their security concerns, new security protocols, called distance-bounding protocols, have been developed to ensure the physical proximity of the de- vices during a session. In order to prevent flaws and attacks, these protocols require formal verification. In this manuscript, we present several techniques that allow for an automatic verification of such protocols. To this aim, we first present a symbolic model which faithfully models time and locations. Then we develop two approaches : either ba- sed on a new verification procedure, or leveraging existing tools like Proverif. Along this manuscript, we pay a particular attention to apply our results to contactless payment protocols
Sarkar, Somwrita. "Acquiring symbolic design optimization problem reformulation knowledge: On computable relationships between design syntax and semantics". Thesis, The University of Sydney, 2009. http://hdl.handle.net/2123/5683.
Texto completoDallon, Antoine. "Vérification de propriétés d'indistinguabilité pour les protocoles cryptographiques". Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLN044/document.
Texto completoThis thesis presents methods to verify cryptographic protocolsin the symbolic model: formal methods allowto verify that small distributed programssatisfy equivalence properties.Those properties state that an attackercannot decide what scenario is beeing played.Strong secrecy, and privacy type properties, like anonymityand unlinkeability, can be modelled through this formalism.Moreover, protocols are executed simultaneouslyby an unbounded number of agents, for an unbounded numberof sessions,which leads to indecidability results.So, we have either to consider an arbitrary number of sessions,and search for semi-decision proceduresand decidable classes;or to establish decision procedures for a finite numberof sessions.When we started the work presented in this thesis,the existing equivalence checkers in the bounded modelwere highly limited. They could only handlea~very small number of sessions (sometimes no more than three).This thesis presents efficient decision proceduresfor bounded verification of equivalence properties.Our first step is to provide small attack results.First, for deterministic processes, there existsan attack if, and ony if, there is a well-typed attack,assuming that there is no confusion between variable types.Second, when there exists a flaw,the attacker needs at most three constants to find it.Then, our second step is to translatethe indistinguishability problem as a reachability problemin a planning system. We solve this second problemthrough planning graph algorithm and SAT encoding.In a final step, we present the implementation ofthe SAT-Equiv tool, which allows us to evaluate our approach.In particular, a benchmark with comparable tools provesthe efficiency of SAT-Equiv
Smart, Angela. "Undergraduate Students’ Connections Between the Embodied, Symbolic, and Formal Mathematical Worlds of Limits and Derivatives: A Qualitative Study Using Tall’s Three Worlds of Mathematics". Thèse, Université d'Ottawa / University of Ottawa, 2013. http://hdl.handle.net/10393/24247.
Texto completoClochard, Martin. "Méthodes et outils pour la spécification et la preuve de propriétés difficiles de programmes séquentiels". Thesis, Université Paris-Saclay (ComUE), 2018. http://www.theses.fr/2018SACLS071/document.
Texto completoThis thesis is set in the domain of deductive verification of programs, which consists of transforming a property to be verified about a program into a logical statement, and then proving this statement. Effective verification of a program can pose many practical difficulties. In fact, the concepts behind the program may be sufficient to impede verification. Indeed, some programs can be quite short and use only simple constructions, and yet prove very difficult to verify. This leads us to the following question: in the context of a deductive program verification environment based on automatic provers, what methods can be applied to reduce the effort required both to specify complex behaviors, as well as to prove that a program respects these expected behaviors? To carry out our study, we placed ourselves in the context of the deductive verification environment of programs Why3. The verification of programs in Why3 is based on the generation of verification conditions, and the use of external provers to prove them, whether these provers are automatic or interactive. We have developed several methods, some general and others specific to some program classes, to reduce manual effort. Our contributions are as follows. First, we add features to Why3 to assist the verification process, including a lightweight declarative proof mechanism based on the notion of cut indicators. Then we present a method for checking the absence of arithmetic overflow, for use cases which are difficult to process by standard methods. Finally, we are interested in the development of a generic library for the specification and proof of code generating programs
Cheval, Vincent. "Automatic verification of cryptographic protocols : privacy-type properties". Phd thesis, École normale supérieure de Cachan - ENS Cachan, 2012. http://tel.archives-ouvertes.fr/tel-00861389.
Texto completoGamard, Guilhem. "Couverture d'un mot bidimensionnel par un motif chevauchant". Thesis, Montpellier, 2017. http://www.theses.fr/2017MONTS027/document.
Texto completoWe study the notion of quasiperiodicity, introduced by Apostolico and Ehrenfeucht at the beginning of the 1990's, then extended to infinite words by Solomon Marcus at the beginning of the 2000's. A (finite or infinite) word w is quasiperiodic if it can be covered by occurrences, possibly overlapping, of another finite word, call its quasiperiod. In 2006, Monteil and Marcus introduced a stronger notion: multi-scale quasiperiodicity, the property of having infinitely many quasiperiods.First we study quasiperiodicity of two-dimensional infinite words. We show that, by contrast with the one-dimensional case where quasiperiodicity do not force any property on infinite words, there exist quasiperiods q which force 2D q-quasiperiodic words to have zero entropy. We also show that multi-scale quasiperiodicity in two dimension force the existence of uniform frequencies for factors.Then we give results on infinite words in one dimension. Most notably we give a method to determine the quasiperiods of an infinite words from its square and special factors. We show that the family of periodic words and standard Sturmian words are characterizable in terms of multi-scale quasiperiodicity
Reis, Teofilo de Souza. "Conectivos flexíveis : uma abordagem categorial às semânticas de traduções possíveis". [s.n.], 2008. http://repositorio.unicamp.br/jspui/handle/REPOSIP/278896.
Texto completoDissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Filosofia e Ciencias Humanas
Made available in DSpace on 2018-08-11T21:55:18Z (GMT). No. of bitstreams: 1 Reis_TeofilodeSouza_M.pdf: 733611 bytes, checksum: 0e64d330d9e71079eddd94de91f141c2 (MD5) Previous issue date: 2008
Resumo: Neste trabalho apresentamos um novo formalismo de decomposição de Lógicas, as Coberturas por Traduções Possíveis, ou simplesmente CTPs. As CTPs constituem uma versão formal das Semânticas de Traduções Possíveis, introduzidas por W. Carnielli em 1990. Mostramos como a adoção de um conceito mais geral de morfismo de assinaturas proposicionais (usando multifunções no lugar de funções) nos permite definir uma categoria Sig?, na qual os conectivos, ao serem traduzidos de uma assinatura para outra, gozam de grande flexibilidade. A partir de Sig?, contruímos a categoria Log? de lógicas tarskianas e morfismos (os quais são funções obtidas a partir de um morfismo de assinaturas, isto é, de uma multifunção). Estudamos algumas características de Sig? e Log?, afim de verificar que estas categorias podem de fato acomodar as construções que pretendemos apresentar. Mostramos como definir em Log? o conjunto de traduções possíveis de uma fórmula, e a partir disto definimos a noção de CTP para uma lógica L. Por fim, exibimos um exemplo concreto de utilização desta nova ferramenta, e discutimos brevemente as possíveis abordagens para uma continuação deste trabalho.
Abstract: We present a general study of a new formalism of decomposition of logics, the Possible- Translations Coverings (in short PTC 's) which constitute a formal version of Possible-Translations Semantics, introduced by W. Carnielli in 1990. We show how the adoption of a more general notion of propositional signatures morphism allows us to define a category Sig?, in which the connectives, when translated from a signature to another one, enjoy of great flexibility. Essentially, Sig? -morphisms will be multifunctions instead of functions. From Sig? we construct the category Log? of tarskian logics and morphisms between them (these .are functions obtained from signature morphisms, that is, from multifunctions) . We show how to define in Log? the set of possible translations of a given formula, and we define the notion of a PTC for a logic L. We analyze some properties of PTC 's and give concrete examples of the above mentioned constructions. We conclude with a discussion of the approaches to be used in a possible continuation of these investigations.
Mestrado
Mestre em Filosofia
Bréhard, Florent. "Certified numerics in function spaces : polynomial approximations meet computer algebra and formal proof". Thesis, Lyon, 2019. http://www.theses.fr/2019LYSEN032/document.
Texto completoRigorous numerics aims at providing certified representations for solutions of various problems, notably in functional analysis, e.g., differential equations or optimal control. Indeed, specific domains like safety-critical engineering or computer-assisted proofs in mathematics have stronger reliability requirements than what can be achieved by resorting to standard numerical analysis algorithms. Our goal consists in developing efficient algorithms, which are also validated / certified in the sense that all numerical errors (method or rounding) are taken into account. Specifically, a central contribution is to combine polynomial approximations with a posteriori fixed-point validation techniques. A C code library for rigorous polynomial approximations (RPAs) is provided, together with a Coq formal proof development, offering the highest confidence at the implementation level.After providing basic operations on RPAs, we focus on a new validation algorithm for Chebyshev basis solutions of D-finite functions, i.e., solutions of linear ordinary differential equations (LODEs) with polynomial coefficients. We give an in-depth complexity analysis, as well as an extension to general LODEs, and even coupled systems of them. These symbolic-numeric methods are finally used in several related problems: a new lower bound on the Hilbert number for quartic systems; a validation of trajectories arising in the linearized spacecraft rendezvous problem; the design of evaluation error efficient polynomial approximations; and the support and density reconstruction of particular measures using algebraic techniques
Wang, Xiaotian. "Mission-aware Vulnerability Assessment for Cyber-Physical System". Wright State University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=wright1440809206.
Texto completoBueno-Soler, Juliana 1976. "Multimodalidades anodicas e catodicas : a negação controlada em logicas multimodais e seu poder expressivo". [s.n.], 2009. http://repositorio.unicamp.br/jspui/handle/REPOSIP/280387.
Texto completoTese (doutorado) - Universidade Estadual de Campinas, Instituto de Filosofia e Ciencias Humanas
Made available in DSpace on 2018-09-11T21:14:41Z (GMT). No. of bitstreams: 1 Bueno-Soler_Juliana_D.pdf: 1230879 bytes, checksum: c04ce9e8061c154854f6283749f9c12b (MD5) Previous issue date: 2009
Resumo: O presente trabalho tem por objetivo investigar o papel da negação no âmbito das modalidades, de forma a poder esclarecer até que ponto a negação pode ser atenuada, controlada ou mesmo totalmente eliminada em favor da melhor expressabilidade lógica de certas teorias, asserções ou raciocínios que sofrem os efeitos da negação. Contudo, atenuar ou eliminar a negação tem um alto preço: métodos tradicionais em lógica podem deixar de ser válidos e certos resultados, como teoremas de completude para sistemas lógicos, podem ser derrogados. Do ponto de vista formal, a questão central que investigamos aqui e até que ponto tais métodos podem ser restabelecidos. Com tal finalidade, iniciamos nosso estudo a partir do que denominamos sistemas anódicos" (sem negação) e, a posteriori, introduzimos gradativamente o elemento catódico" (negações, com diversas gradações e diferentes características) nos sistemas modais por meio de combinações com certas lógicas paraconsistentes, as chamadas lógicas da inconsistência formal (LFIs). Todos os sistemas tratados são semanticamente caracterizados por semânticas de mundos possíveis; resultados de incompletude são também obtidos e discutidos. Obtemos ainda semânticas modais de traduções possíveis para diversos desses sistemas. Avançamos na direção das multimodalidades, investigando os assim chamados sistemas multimodais anódicos e catódicos. Finalmente, procuramos avaliar criticamente o alcance e o interesse dos resultados obtidos na direção da racionalidade sensível à negação.
Abstract: The present work aims to investigate the role of negations in the scope of modalities and in the reasoning expressed by modalities. The investigation starts from what we call anodic" systems (without any form of negation) and gradually reaches the cathodic" elements, where negations are introduced by means of combining modal logics with certain paraconsistent logics known as logics of formal inconsistency (LFIs). We obtain completeness results for all treated systems, and also show that certain incompleteness results can be obtained. The class of the investigated systems includes all normal modal logics that are extended by means of the schema Gk;l;m;n due to E. J. Lemmon and D. Scott combined with LFIs. We also tackle the question of obtaining modal possible-translations semantics for these systems. Analogous results are analyzed in the scope of multimodalities, where anodic as much as cathodic logics are studied. Finally, we advance a critical evaluation of the reach and scope of all the results obtained to what concerns expressibility of reasoning considered to be sensible to negation. We also critically assess the obtained results in contrast with problems of rationality that are sensible to negation.
Doutorado
Doutor em Filosofia
Chalk, Matěj. "Nástroj pro abstraktní regulární model checking". Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2018. http://www.nusl.cz/ntk/nusl-385924.
Texto completoRodrigues, Tarcísio Genaro. "Sobre os fundamentos de programação lógica paraconsistente". [s.n.], 2010. http://repositorio.unicamp.br/jspui/handle/REPOSIP/278897.
Texto completoDissertação (mestrado) - Universidade Estadual de Campinas, Instituto de Filosofia e Ciencias Humanas
Made available in DSpace on 2018-08-17T03:29:03Z (GMT). No. of bitstreams: 1 Rodrigues_TarcisioGenaro_M.pdf: 1141020 bytes, checksum: 59bb8a3ae7377c05cf6a8d8e6f7e45a5 (MD5) Previous issue date: 2010
Resumo: A Programação Lógica nasce da interação entre a Lógica e os fundamentos da Ciência da Computação: teorias de primeira ordem podem ser interpretadas como programas de computador. A Programação Lógica tem sido extensamente utilizada em ramos da Inteligência Artificial tais como Representação do Conhecimento e Raciocínio de Senso Comum. Esta aproximação deu origem a uma extensa pesquisa com a intenção de definir sistemas de Programação Lógica paraconsistentes, isto é, sistemas nos quais seja possível manipular informação contraditória. Porém, todas as abordagens existentes carecem de uma fundamentação lógica claramente definida, como a encontrada na programação lógica clássica. A questão básica é saber quais são as lógicas paraconsistentes subjacentes a estas abordagens. A presente dissertação tem como objetivo estabelecer uma fundamentação lógica e conceitual clara e sólida para o desenvolvimento de sistemas bem fundados de Programação Lógica Paraconsistente. Nesse sentido, este trabalho pode ser considerado como a primeira (e bem sucedida) etapa de um ambicioso programa de pesquisa. Uma das teses principais da presente dissertação é que as Lógicas da Inconsistência Formal (LFI's), que abrangem uma enorme família de lógicas paraconsistentes, proporcionam tal base lógica. Como primeiro passo rumo à definição de uma programação lógica genuinamente paraconsistente, demonstramos nesta dissertação uma versão simplificada do Teorema de Herbrand para uma LFI de primeira ordem. Tal teorema garante a existência, em princípio, de métodos de dedução automática para as lógicas (quantificadas) em que o teorema vale. Um pré-requisito fundamental para a definição da programação lógica é justamente a existência de métodos de dedução automática. Adicionalmente, para a demonstração do Teorema de Herbrand, são formuladas aqui duas LFI's quantificadas através de sequentes, e para uma delas demonstramos o teorema da eliminação do corte. Apresentamos também, como requisito indispensável para os resultados acima mencionados, uma nova prova de correção e completude para LFI's quantificadas na qual mostramos a necessidade de exigir o Lema da Substituição para a sua semântica
Abstract: Logic Programming arises from the interaction between Logic and the Foundations of Computer Science: first-order theories can be seen as computer programs. Logic Programming have been broadly used in some branches of Artificial Intelligence such as Knowledge Representation and Commonsense Reasoning. From this, a wide research activity has been developed in order to define paraconsistent Logic Programming systems, that is, systems in which it is possible to deal with contradictory information. However, no such existing approaches has a clear logical basis. The basic question is to know what are the paraconsistent logics underlying such approaches. The present dissertation aims to establish a clear and solid conceptual and logical basis for developing well-founded systems of Paraconsistent Logic Programming. In that sense, this text can be considered as the first (and successful) stage of an ambitious research programme. One of the main thesis of the present dissertation is that the Logics of Formal Inconsistency (LFI's), which encompasses a broad family of paraconsistent logics, provide such a logical basis. As a first step towards the definition of genuine paraconsistent logic programming we shown, in this dissertation, a simplified version of the Herbrand Theorem for a first-order LFI. Such theorem guarantees the existence, in principle, of automated deduction methods for the (quantified) logics in which the theorem holds, a fundamental prerequisite for the definition of logic programming over such logics. Additionally, in order to prove the Herbrand Theorem we introduce sequent calculi for two quantified LFI's, and cut-elimination is proved for one of the systems. We also present, as an indispensable requisite for the above mentioned results, a new proof of soundness and completeness for first-order LFI's in which we show the necessity of requiring the Substitution Lemma for the respective semantics
Mestrado
Filosofia
Mestre em Filosofia
Jacomme, Charlie. "Preuves de protocoles cryptographiques : méthodes symboliques et attaquants puissants". Electronic Thesis or Diss., université Paris-Saclay, 2020. http://www.theses.fr/2020UPASG005.
Texto completoThe use of communication protocols has become pervasive at all levels of our society. Yet, their uses come with risks, either about the security of the system or the privacy of the user. To mitigate those risks, we must provide the protocols with strong security guarantees: we need formal, extensive, modular and machine-checked proofs. However, such proofs are very difficult to obtain in practice. In this Thesis, we strive to ease this process in the case of cryptographic protocols and powerful attackers. The four main contributions of this Thesis, all based on symbolic methods, are 1) a methodology for extensive analyses via a case study of multi-factor authentication; 2) composition results to allow modular proofs of complex protocols in the computational model; 3) symbolic methods for deciding basic proof steps in computational proofs, formulated as problems on probabilistic programs; 4) a prototype of a mechanized prover in the Computationally Complete Symbolic Attacker model
Kotoun, Michal. "Symbolické automaty v analýze programů s řetězci". Master's thesis, Vysoké učení technické v Brně. Fakulta informačních technologií, 2020. http://www.nusl.cz/ntk/nusl-433553.
Texto completoHarrath, Nesrine. "A stepwise compositional approach to model and analyze system C designs at the transactional level and the delta cycle level". Thesis, Paris, CNAM, 2014. http://www.theses.fr/2014CNAM0957/document.
Texto completoEmbedded systems are increasingly integrated into existing real-time applications. They are usually composed of deeply integrated but heterogeneous hardware and software components. These components are developed under strict constraints. Accordingly, the work of design engineers became more tricky and challenging. To meet the high quality standards in nowadays embedded systems and to satisfy the rising industrial demands, the automatization of the developing process of those systems is gaining more and more importance. A major challenge is to develop an automated approach that can be used for the integrated verification and validation of complex and heterogeneous HW/SW systems.In this thesis, we propose a new compositional approach to model and verify hardware and software written in SystemC language. This approach is based on the SystemC Waiting State Automata (WSA). The SystemC Waiting State Automata are used to model the abstract behavior of hardware or software systems described in SystemC. They preserve the semantics of the SystemC scheduler at the temporal and the delta-cycle level. This model allows to reduce the complexity of the modeling process of complex systems due to the problem of state explosion during modeling while remaining faithful to the original system. The SystemC waiting state automaton is also compositional and supports refinement. In addition, this model is extended with parameters such as time and counters in order to take into account further aspects like temporality and other extra-functional properties such as QoS.In this thesis, we propose a stepwise approach on how to automatically extract the SystemC WSAs from SystemC descriptions. This construction is based on symbolic execution together with predicate abstraction. We propose a set of algorithms to symbolically compose and reduce the SystemC WSAs in order to study, analyze and verify concurrent behavior of systems as well as the data exchange between various components. We then propose to use the SystemC WSA to model and simulate hardware and software systems, and to compute the worst cas execution time (WCET) using the Timed SystemC WSA. Finally, we define how to apply model checking techniques to prove the correctness of the abstract analysis
Kanso, Bilal. "Modélisation et validation des systèmes informatiques complexes". Phd thesis, Ecole Centrale Paris, 2011. http://tel.archives-ouvertes.fr/tel-00650258.
Texto completoHarrath, Nesrine. "A stepwise compositional approach to model and analyze system C designs at the transactional level and the delta cycle level". Electronic Thesis or Diss., Paris, CNAM, 2014. http://www.theses.fr/2014CNAM0957.
Texto completoEmbedded systems are increasingly integrated into existing real-time applications. They are usually composed of deeply integrated but heterogeneous hardware and software components. These components are developed under strict constraints. Accordingly, the work of design engineers became more tricky and challenging. To meet the high quality standards in nowadays embedded systems and to satisfy the rising industrial demands, the automatization of the developing process of those systems is gaining more and more importance. A major challenge is to develop an automated approach that can be used for the integrated verification and validation of complex and heterogeneous HW/SW systems.In this thesis, we propose a new compositional approach to model and verify hardware and software written in SystemC language. This approach is based on the SystemC Waiting State Automata (WSA). The SystemC Waiting State Automata are used to model the abstract behavior of hardware or software systems described in SystemC. They preserve the semantics of the SystemC scheduler at the temporal and the delta-cycle level. This model allows to reduce the complexity of the modeling process of complex systems due to the problem of state explosion during modeling while remaining faithful to the original system. The SystemC waiting state automaton is also compositional and supports refinement. In addition, this model is extended with parameters such as time and counters in order to take into account further aspects like temporality and other extra-functional properties such as QoS.In this thesis, we propose a stepwise approach on how to automatically extract the SystemC WSAs from SystemC descriptions. This construction is based on symbolic execution together with predicate abstraction. We propose a set of algorithms to symbolically compose and reduce the SystemC WSAs in order to study, analyze and verify concurrent behavior of systems as well as the data exchange between various components. We then propose to use the SystemC WSA to model and simulate hardware and software systems, and to compute the worst cas execution time (WCET) using the Timed SystemC WSA. Finally, we define how to apply model checking techniques to prove the correctness of the abstract analysis
Laroque, Octavie. "Les lois symboliques. Une étude à partir du droit de la propriété littéraire et artistique". Thesis, Paris 2, 2017. http://www.theses.fr/2017PA020040.
Texto completoSymbolic laws are a recent manifestation of a contemporary legislative evil. They are not only incantatory declarations on the model of non-normative or "memorial" laws, since they can also be technical rulings, as intellectual property law is. To grasp this phenomenon, we must first identify what symbolic laws are. Characterized by the disharmony between their discourse and their normative qualities, these laws show a phenomenon of ineffective implementation. Unclear, unrealistic, sometimes lying, but endowed with a virtuous message, symbolic laws are the result of the instrumentalization of legislative action, an exercise where expressing values is more a concern than the concrete effects of the text. Secondly, it is important to determine how symbolic laws should be dealt with. As a sign of a change in legislative production and in intellectual property law, these laws are the figure of disorder: they mark the withdrawal of the true symbolism and its vain compensation by a false and flashy symbolism. This discovery calls for a restoration of order and the respect of legistic rules, where those who write the norms should be animated by the love of laws and guided by moral conscience. As intellectual property law is under attack by commercial and social demands, this study invites to think about its future and see how we could prevent the appearance of symbolic laws
Py, Frédéric. "Contrôle d'exécution dans une architecture hiérarchisée pour systèmes autonomes". Toulouse 3, 2005. http://www.theses.fr/2005TOU30199.
Texto completoThere is an increasing need for advanced autonomy in complex embedded real-time systems such as robots or satellites. Still, this raises a major problem : on one side we have complex sys-tems - therefore, hard to validate - with little human intervention, on the other side these systems are used in domains where safety is critical. How can we guaranty that an autonomous system, with high level decisional capabilities, will exhibit a proper behavior and will not jeopardize the mission? The work we present here integrate an on-line execution control component for hierar-chical architectures. We first describe the role of this program. Then we introduce the R2C, our controller based on synchronous hypothesis, and the tool used to generate it. We then discuss why it is important to take into account the decisional components in our controller. We eventu-ally illustrate our contribution with some experimental results. We then conclude and give some possible future work in this area
Bentakouk, Lina. "Test symbolique de services web composite". Phd thesis, Université Paris Sud - Paris XI, 2011. http://tel.archives-ouvertes.fr/tel-00675918.
Texto completoBenhamamouch, Bilel. "Calcul du pire temps d'exécution : méthode formelle s'adaptant à la sophistication croissante des architectures matérielles". Phd thesis, Université de Grenoble, 2011. http://tel.archives-ouvertes.fr/tel-00685866.
Texto completoLafourcade, Pascal. "Sécurité assistée par ordinateur pour les primitives cryptgraphiques, les protocoles de vote électronique et les réseaux de capteurs sans fil". Habilitation à diriger des recherches, Université de Grenoble, 2012. http://tel.archives-ouvertes.fr/tel-00807568.
Texto completoJournault, Matthieu. "Analyse statique modulaire précise par interprétation abstraite pour la preuve automatique de correction de programmes et pour l’inférence de contrats". Electronic Thesis or Diss., Sorbonne université, 2019. http://www.theses.fr/2019SORUS152.
Texto completoEnsuring the scalability of static analyzers defined by abstract interpretation poses difficulties. A classical technique known to speed up analyses is the discovery and reuse of summaries for some of the sequences of statements of the source code. In this thesis we focus on a subset of C that does not allow recursion and define a modular analyzer, able to infer, prove and use (to improve the efficiency) such summaries. Our modular analyzer is built on top of an existing C analyzer and is therefore able to handle unions, structures, arrays, memory allocations (static and dynamic), pointers, pointer arithmetics, pointer casts, function calls, string manipulations, ... . String handling is provided by a new abstract domain defined in this thesis. In this thesis we provide a lifting of classical numerical abstract domains to the representation of heterogeneous sets. This lifting can be used for relational domains and maintains only one numerical abstract state, by opposition to partitioning. The last point of interest of this thesis is the definition of an abstract domain able to represent sets of trees with numerically labeled leaves. This abstraction is based on regular and tree regular languages and delegates the handling of numerical constraints to an underlying domain able to represent heterogeneous sets of environments. As the thesis took place in the mopsa project, we provide an overview of some of the results obtained by the mopsa team during the thesis
Butin, Frédéric. "Structures de Poisson sur les Algèbres de Polynômes, Cohomologie et Déformations". Thesis, Lyon 1, 2009. http://www.theses.fr/2009LYO10192/document.
Texto completoDeformation quantization and McKay correspondence form the main themes of the study which deals with singular algebraic varieties, quotients of polynomial algebras, and polynomial algebras invariant under the action of a finite group. Our main tools are Poisson and Hochschild cohomologies and representation theory. Certain calculations are made with Maple and GAP. We calculate Hochschild homology and cohomology spaces of Klein surfaces by developing a generalization of HKR theorem in the case of non-smooth varieties and use the multivariate division and the Groebner bases. The closure of the minimal nilpotent orbit of a simple Lie algebra is a singular algebraic variety : on this one we construct invariant star-products, with the help of the BGS decomposition of Hochschild homology and cohomology, and of results on the invariants of the classical groups. We give the generators of the Joseph ideal associated to this orbit and calculate the infinitesimal characters. For simple Lie algebras of type B, C, D, we establish general results on the Poisson homology space in degree 0 of the invariant algebra, which support Alev's conjecture, then we are interested in the ranks 2 and 3. We compute Poincaré series of 2 variables for the finite subgroups of the special linear group in dimension 3, show that they are rational fractions, and associate to the subgroups a generalized Cartan matrix in order to obtain a McKay correspondence in dimension 3. All the study comes from 4 papers
Jacquemard, Florent. "Modèles d'automates d'arbres étendus pour la vérification de systèmes infinis". Habilitation à diriger des recherches, École normale supérieure de Cachan - ENS Cachan, 2011. http://tel.archives-ouvertes.fr/tel-00643595.
Texto completoCousin, Bernard. "Méthodologie de validation des systèmes structurés en couches par réseaux de Petri : application au protocole Transport". Phd thesis, Université Pierre et Marie Curie - Paris VI, 1987. http://tel.archives-ouvertes.fr/tel-00864063.
Texto completoBourreau, Pierre. "Jeux de typage et analyse de lambda-grammaires non-contextuelles". Phd thesis, Université Sciences et Technologies - Bordeaux I, 2012. http://tel.archives-ouvertes.fr/tel-00733964.
Texto completoUang, Chia-Yuan y 汪加元. "Automatic Assertion Checking Using Formal Symbolic Model Verifier". Thesis, 2005. http://ndltd.ncl.edu.tw/handle/15654269991971750495.
Texto completo國立交通大學
電機資訊學院碩士在職專班
93
Assertion based verification (ABV) methodology has emerged as a paradigm of high-level design verification. An assertion is used to specify what is to be exercised and verified against the intended functionality. However assertions which may contain conflicts among themselves are not inspected until later simulation stage.In this thesis, we present an automatic assertion checking which utilizes an existing symbolic model verifier as a model checker to check if there is any conflict among input assertions. We propose an approach to convert the assertions into structural Deterministic Finite Automata (DFA) and their corresponding properties. Those converted DFA and properties are then checked by using formal model verifier. This approach may facilitate assertion checking to find out potential conflict in the early stage of design activities without simulation.
LIN, HONG-JUN y 林弘峻. "An one-to-many guided learning environment for symbolic-calculation". Thesis, 1993. http://ndltd.ncl.edu.tw/handle/62405460806848997570.
Texto completo