Academic literature on the topic 'Register Automata'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Register Automata.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Register Automata"

1

FIGUEIRA, DIEGO, PIOTR HOFMAN, and SŁAWOMIR LASOTA. "Relating timed and register automata." Mathematical Structures in Computer Science 26, no. 6 (December 5, 2014): 993–1021. http://dx.doi.org/10.1017/s0960129514000322.

Full text
Abstract:
Timed and register automata are well-known models of computation over timed and data words, respectively. The former has clocks that allow to test the lapse of time between two events, whilst the latter includes registers that can store data values for later comparison. Although these two models behave differently in appearance, several decision problems have the same (un)decidability and complexity results for both models. As a prominent example, emptiness is decidable for alternating automata with one clock or register, both with non-primitive recursive complexity. This is not by chance.This work confirms that there is indeed a tight relationship between the two models. We show that a run of a timed automaton can be simulated by a register automaton over ordered data domain, and conversely that a run of a register automaton can be simulated by a timed automaton. These are exponential time reductions hold both in the finite and infinite words settings. Our results allow to transfer decidability results back and forth between these two kinds of models, as well complexity results modulo an exponential time reduction. We justify the usefulness of these reductions by obtaining new results on register automata.
APA, Harvard, Vancouver, ISO, and other styles
2

Tzevelekos, Nikos. "Fresh-register automata." ACM SIGPLAN Notices 46, no. 1 (January 26, 2011): 295–306. http://dx.doi.org/10.1145/1925844.1926420.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Figueira, Diego, Piotr Hofman, and Sławomir Lasota. "Relating timed and register automata." Electronic Proceedings in Theoretical Computer Science 41 (November 28, 2010): 61–75. http://dx.doi.org/10.4204/eptcs.41.5.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Murawski, A. S., S. J. Ramsay, and N. Tzevelekos. "Reachability in pushdown register automata." Journal of Computer and System Sciences 87 (August 2017): 58–83. http://dx.doi.org/10.1016/j.jcss.2017.02.008.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Quaas, Karin, and Mahsa Shirmohammadi. "Synchronizing Data Words for Register Automata." ACM Transactions on Computational Logic 20, no. 2 (April 4, 2019): 1–27. http://dx.doi.org/10.1145/3309760.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Almagor, Shaull, Michaël Cadilhac, Filip Mazowiecki, and Guillermo A. Pérez. "Weak Cost Register Automata are Still Powerful." International Journal of Foundations of Computer Science 31, no. 06 (September 2020): 689–709. http://dx.doi.org/10.1142/s0129054120410026.

Full text
Abstract:
We consider one of the weakest variants of cost register automata over a tropical semiring, namely copyless cost register automata over [Formula: see text] with updates using [Formula: see text] and increments. We show that this model can simulate, in some sense, the runs of counter machines with zero-tests. We deduce that a number of problems pertaining to that model are undecidable, namely equivalence, upperboundedness, and semilinearity. In particular, the undecidability of equivalence disproves a conjecture of Alur et al. from 2012. To emphasize how weak these machines are, we also show that they can be expressed as a restricted form of linearly-ambiguous weighted automata.
APA, Harvard, Vancouver, ISO, and other styles
7

Orlov, Alexei O., Ravi Kummamuru, R. Ramasubramaniam, Craig S. Lent, Gary H. Bernstein, and Gregory L. Snider. "Clocked quantum-dot cellular automata shift register." Surface Science 532-535 (June 2003): 1193–98. http://dx.doi.org/10.1016/s0039-6028(03)00214-0.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Seki, Hiroyuki, Reo Yoshimura, and Yoshiaki Takata. "Optimal run problem for weighted register automata." Theoretical Computer Science 850 (January 2021): 185–201. http://dx.doi.org/10.1016/j.tcs.2020.11.003.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Allender, Eric, Andreas Krebs, and Pierre McKenzie. "Better Complexity Bounds for Cost Register Automata." Theory of Computing Systems 63, no. 3 (June 13, 2018): 367–85. http://dx.doi.org/10.1007/s00224-018-9871-4.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Kadiev, P. A., K. K. Nazarov, and Z. G. Kardashova. "Formation of streams of binary sequences with controlled structure of "cellular" automata in homogeneous register environments." Herald of Dagestan State Technical University. Technical Sciences 48, no. 4 (February 11, 2022): 90–99. http://dx.doi.org/10.21822/2073-6185-2021-48-4-90-99.

Full text
Abstract:
Objective. Development of a method for organizing the process of forming flows register structure, patented by the author at the Department of DSTU, which is a cellular automaton of binary sequences with a controlled structure of "cellular" automata in homogeneous register environments. Method. To solve the set system problem, a process model was built in order to determine the factors that allow changing the sequence of flow elements. Result. When studying the most common generators of pseudo-random streams based on linear register media with modulo two adders in feedback circuits, which are "cellular" automata, it was established from the tables of environment states that the factors determining the structure of the generated streams of binary sequences are the sequence of states , are the initial state of the register of a homogeneous medium and the feedback structure determined by the transition function of the cellular automaton. Conclusion. It is shown that by varying the initial state of the cellular automaton and the structure of feedbacks as control tools, it seems possible to form binary pseudo-random streams of states with different structures, different order of flow elements, bringing their characteristics closer to random ones. Examples of the implementation of the flow structure control process are given, confirming this assumption. A typical structure of a stream shaper with a managed structure based on a homogeneous one is given.
APA, Harvard, Vancouver, ISO, and other styles

Dissertations / Theses on the topic "Register Automata"

1

Rueda, Cebollero Guillem. "Learning Cache Replacement Policies using Register Automata." Thesis, Uppsala universitet, Institutionen för informationsteknologi, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-212677.

Full text
Abstract:
Processors are a basic unit of the computer which accomplish the mission of processing data stored in the memory. Large memories are required to process a big amount of data. Not all data is required at the same time, few data is required faster than other. For this reason, the memory is structured  in a hierarchy, from smaller and faster to bigger and slower. The cache memory is one of the fastest elements and closest to the processor in the memory hierarchy. The processor design companies hides its characteristics, usually under a confidential documentation that can not be accessed by the software developers. One of the most important characteristics kept in secret in this documentation is the replacement policy. The most famous replacement policies are known but the hardware designers can apply modifications for performance, cost or design reasons. The obfuscation of a part of the processor implies many developers to avoid problems with, for example, the runtime. If a task must be executed always in a certain time, the developer will take always the case requiring more time to execute (also called "Worst Case Execution Time") implying an underutilisation of the processor. This project will be focused on a new method to represent  and infer the replacement policy: modelling the replacement policies with automaton and using a learning process framework called LearnLib to guess them. This is not the first project trying to match the cache memory characteristics, actually a previous project is the basis to find a more general model to define the replacement policies. The results of LearnLib are modelled as an automaton. In order to test the effectiveness of this framework, different replacement policies will be simulated and verified. To provide a interface with a real cache memories is developed a program called hwquery. This program will interface a real cache request for using it in Learnlib.
APA, Harvard, Vancouver, ISO, and other styles
2

Exibard, Léo. "Automatic synthesis of systems with data." Electronic Thesis or Diss., Aix-Marseille, 2021. http://www.theses.fr/2021AIXM0312.

Full text
Abstract:
Nous interagissons régulièrement avec des machines qui réagissent en temps réel à nos actions (robots, sites web etc). Celles-ci sont modélisées par des systèmes réactifs, caractérisés par une interaction constante avec leur environnement. L'objectif de la synthèse réactive est de générer automatiquement un tel système à partir de la description de son comportement afin de remplacer la phase de développement bas-niveau, sujette aux erreurs, par l'élaboration d'une spécification haut-niveau.Classiquement, on suppose que les signaux d'entrée de la machine sont en nombre fini. Un tel cadre échoue à modéliser les systèmes qui traitent des données issues d'un ensemble infini (un identifiant unique, la valeur d'un capteur, etc). Cette thèse se propose d'étendre la synthèse réactive au cas des mots de données. Nous étudions un modèle adapté à ce cadre plus général, et examinons la faisabilité des problèmes de synthèse associés. Nous explorons également les systèmes non réactifs, où l'on n'impose pas à la machine de réagir en temps réel
We often interact with machines that react in real time to our actions (robots, websites etc). They are modelled as reactive systems, that continuously interact with their environment. The goal of reactive synthesis is to automatically generate a system from the specification of its behaviour so as to replace the error-prone low-level development phase by a high-level specification design.In the classical setting, the set of signals available to the machine is assumed to be finite. However, this assumption is not realistic to model systems which process data from a possibly infinite set (e.g. a client id, a sensor value, etc.). The goal of this thesis is to extend reactive synthesis to the case of data words. We study a model that is well-suited for this more general setting, and examine the feasibility of its synthesis problem(s). We also explore the case of non-reactive systems, where the machine does not have to react immediately to its inputs
APA, Harvard, Vancouver, ISO, and other styles
3

Kuriakose, R. B., and F. Aghdasi. "Automatic student attendance register using RFID." Interim : Interdisciplinary Journal, Vol 6, Issue 2: Central University of Technology Free State Bloemfontein, 2007. http://hdl.handle.net/11462/406.

Full text
Abstract:
Published Article
The purpose of this project is to investigate the application of Radio Frequency Identification, RFID, to automatic student attendance register. The aim is that the students in any class can be recorded when they carry their student cards with them without having to individually swipe the card or allocate special interaction time. The successful implementation of this proposal will facilitate such record keeping in a non-intrusive and efficient manner and will provide the platform for further research on the correlation between attendance and performance of the students. The opportunity for related research is identified regarding the range of the parameters involved, ensuring that individual identifications do not clash and interfacing challenges with the central record keeping are overcome.
APA, Harvard, Vancouver, ISO, and other styles
4

Hauck, Shahram. "Automated CtP Calibration for Offset Printing : Dot gain compensation, register variation and trapping evaluation." Doctoral thesis, Linköpings universitet, Institutionen för teknik och naturvetenskap, 2015. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-119366.

Full text
Abstract:
Although offset printing has been and still is the most common printing technology for color print productions, its print productions are subject to variations due to environmental and process parameters. Therefore, it is very important to frequently control the print production quality criteria in order to make the process predictable, reproducible and stable. One of the most important parts in a modern industrial offset printing is Computer to Plate (CtP), which exposes the printing plate. One of the most important quality criteria for printing is to control the dot gain level. Dot gain refers to an important phenomenon that causes the printed elements to appear larger than their reference size sent to the CtP. It is crucial to have the dot gain level within an acceptable range, defined by ISO 12647-2 for offset printing. This is done by dot gain compensation methods in the Raster Image Processor (RIP). Dot gain compensation is however a complicated task in offset printing because of the huge number of parameters affecting dot gain. Another important quality criterion affecting the print quality in offset is the register variation caused by the misplacement of printing sheet in the printing unit. Register variation causes tone value variations, gray balance variation and blurred image details. Trapping is another important print quality criterion that should be measured in an offset printing process. Trapping occurs when the inks in different printing units are printed wet-on-wet in a multi-color offset printing machine. Trapping affects the gray balance and makes the resulting colors of overlapped inks pale. In this dissertation three different dot gain compensation methods are discussed. The most accurate and efficient dot gain compensation method, which is noniterative, has been tested, evaluated and applied using many offset printing workflows. To further increase the accuracy of this method, an approach to effectively select the correction points of a RIP with limited number of correction points, has also been proposed. Correction points are the tone values needed to be set in the RIP to define a dot gain compensation curve. To fulfill the requirement of having the register variation within the allowed range, it has to be measured and quantified. There have been two novel models proposed in this dissertation that determine the register variation value. One of the models is based on spectrophotometry and the other one on densitometry. The proposed methods have been evaluated by comparison to the industrial image processing based register variation model, which is expensive and not available in most printing companies. The results of all models were comparable, verifying that the proposed models are good  alternatives to the image processing based model. The existing models determining the trapping values are based on densitometric measurements and quantify the trapping effect by a percentage value. In this dissertation, a novel trapping model is proposed that quantifies the trapping effect by a color difference metric, i.e. , which is more useful and understandable for print machine operators. The comparison between the proposed trapping model and the existing models has shown very good correlations and verified that the proposed model has a bigger dynamic range. The proposed trapping model has also been extended to take into account the effect of ink penetration and gloss. The extended model has been tested using a  high glossy coated paper and the results have shown that the gloss and ink penetration can be neglected for this type of paper. An automated CtP calibration system for offset printing workflow has been introduced and described in this dissertation. This method is a good solution to generate the needed huge numbers of dot gain compensation curves to have an accurate CtP calibration.
APA, Harvard, Vancouver, ISO, and other styles
5

Jouhet, Vianney. "Automated adaptation of Electronic Heath Record for secondary use in oncology." Thesis, Bordeaux, 2016. http://www.theses.fr/2016BORD0373/document.

Full text
Abstract:
Avec la montée en charge de l’informatisation des systèmes d’information hospitaliers, une quantité croissante de données est produite tout au long de la prise en charge des patients. L’utilisation secondaire de ces données constitue un enjeu essentiel pour la recherche ou l’évaluation en santé. Dans le cadre de cette thèse, nous discutons les verrous liés à la représentation et à la sémantique des données, qui limitent leur utilisation secondaire en cancérologie. Nous proposons des méthodes basées sur des ontologies pour l’intégration sémantique des données de diagnostics. En effet, ces données sont représentées par des terminologies hétérogènes. Nous étendons les modèles obtenus pour la représentation de la maladie tumorale, et les liens qui existent avec les diagnostics. Enfin, nous proposons une architecture combinant entrepôts de données, registres de métadonnées et web sémantique. L’architecture proposée permet l’intégration syntaxique et sémantique d’un grand nombre d’observations. Par ailleurs, l’intégration de données et de connaissances (sous la forme d’ontologies) a été utilisée pour construire un algorithme d’identification de la maladie tumorale en fonction des diagnostics présents dans les données de prise en charge. Cet algorithme basé sur les classes de l’ontologie est indépendant des données effectivement enregistrées. Ainsi, il fait abstraction du caractère hétérogène des données diagnostiques initialement disponibles. L’approche basée sur une ontologie pour l’identification de la maladie tumorale, permet une adaptation rapide des règles d’agrégation en fonction des besoins spécifiques d’identification. Ainsi, plusieurs versions du modèle d’identification peuvent être utilisées avec des granularités différentes
With the increasing adoption of Electronic Health Records (EHR), the amount of data produced at the patient bedside is rapidly increasing. Secondary use is there by an important field to investigate in order facilitate research and evaluation. In these work we discussed issues related to data representation and semantics within EHR that need to be address in order to facilitate secondary of structured data in oncology. We propose and evaluate ontology based methods for heterogeneous diagnosis terminologies integration in oncology. We then extend obtained model to enable tumoral disease representation and links with diagnosis as recorded in EHR. We then propose and implement a complete architecture combining a clinical data warehouse, a metadata registry and web semantic technologies and standards. This architecture enables syntactic and semantic integration of a broad range of hospital information System observation. Our approach links data with external knowledge (ontology), in order to provide a knowledge resource for an algorithm for tumoral disease identification based on diagnosis recorded within EHRs. As it based on the ontology classes, the identification algorithm is uses an integrated view of diagnosis (avoiding semantic heterogeneity). The proposed architecture leading to algorithm on the top of an ontology offers a flexible solution. Adapting the ontology, modifying for instance the granularity provide a way for adapting aggregation depending on specific needs
APA, Harvard, Vancouver, ISO, and other styles
6

MANSOURI, NAZANIN. "AUTOMATED CORRECTNESS CONDITION GENERATION FOR FORMAL VERIFICATION OF SYNTHESIZED RTL DESIGNS." University of Cincinnati / OhioLINK, 2001. http://rave.ohiolink.edu/etdc/view?acc_num=ucin982064542.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Tabani, Hamid. "Low-power architectures for automatic speech recognition." Doctoral thesis, Universitat Politècnica de Catalunya, 2018. http://hdl.handle.net/10803/462249.

Full text
Abstract:
Automatic Speech Recognition (ASR) is one of the most important applications in the area of cognitive computing. Fast and accurate ASR is emerging as a key application for mobile and wearable devices. These devices, such as smartphones, have incorporated speech recognition as one of the main interfaces for user interaction. This trend towards voice-based user interfaces is likely to continue in the next years which is changing the way of human-machine interaction. Effective speech recognition systems require real-time recognition, which is challenging for mobile devices due to the compute-intensive nature of the problem and the power constraints of such systems and involves a huge effort for CPU architectures to reach it. GPU architectures offer parallelization capabilities which can be exploited to increase the performance of speech recognition systems. However, efficiently utilizing the GPU resources for speech recognition is also challenging, as the software implementations exhibit irregular and unpredictable memory accesses and poor temporal locality. The purpose of this thesis is to study the characteristics of ASR systems running on low-power mobile devices in order to propose different techniques to improve performance and energy consumption. We propose several software-level optimizations driven by the power/performance analysis. Unlike previous proposals that trade accuracy for performance by reducing the number of Gaussians evaluated, we maintain accuracy and improve performance by effectively using the underlying CPU microarchitecture. We use a refactored implementation of the GMM evaluation code to ameliorate the impact of branches. Then, we exploit the vector unit available on most modern CPUs to boost GMM computation, introducing a novel memory layout for storing the means and variances of the Gaussians in order to maximize the effectiveness of vectorization. In addition, we compute the Gaussians for multiple frames in parallel, significantly reducing memory bandwidth usage. Our experimental results show that the proposed optimizations provide 2.68x speedup over the baseline Pocketsphinx decoder on a high-end Intel Skylake CPU, while achieving 61% energy savings. On a modern ARM Cortex-A57 mobile processor our techniques improve performance by 1.85x, while providing 59% energy savings without any loss in the accuracy of the ASR system. Secondly, we propose a register renaming technique that exploits register reuse to reduce the pressure on the register file. Our technique leverages physical register sharing by introducing minor changes in the register map table and the issue queue. We evaluated our renaming technique on top of a modern out-of-order processor. The proposed scheme supports precise exceptions and we show that it results in 9.5% performance improvements for GMM evaluation. Our experimental results show that the proposed register renaming scheme provides 6% speedup on average for the SPEC2006 benchmarks. Alternatively, our renaming scheme achieves the same performance while reducing the number of physical registers by 10.5%. Finally, we propose a hardware accelerator for GMM evaluation that reduces the energy consumption by three orders of magnitude compared to solutions based on CPUs and GPUs. The proposed accelerator implements a lazy evaluation scheme where Gaussians are computed on demand, avoiding 50% of the computations. Furthermore, it employs a novel clustering scheme to reduce the size of the GMM parameters, which results in 8x memory bandwidth savings with a negligible impact on accuracy. Finally, it includes a novel memoization scheme that avoids 74.88% of floating-point operations. The end design provides a 164x speedup and 3532x energy reduction when compared with a highly-tuned implementation running on a modern mobile CPU. Compared to a state-of-the-art mobile GPU, the GMM accelerator achieves 5.89x speedup over a highly optimized CUDA implementation, while reducing energy by 241x.
El reconocimiento automático de voz (ASR) es una de las aplicaciones más importantes en el área de la computación cognitiva. ASR rápido y preciso se está convirtiendo en una aplicación clave para dispositivos móviles y portátiles. Estos dispositivos, como los Smartphones, han incorporado el reconocimiento de voz como una de las principales interfaces de usuario. Es probable que esta tendencia hacia las interfaces de usuario basadas en voz continúe en los próximos años, lo que está cambiando la forma de interacción humano-máquina. Los sistemas de reconocimiento de voz efectivos requieren un reconocimiento en tiempo real, que es un desafío para los dispositivos móviles debido a la naturaleza de cálculo intensivo del problema y las limitaciones de potencia de dichos sistemas y supone un gran esfuerzo para las arquitecturas de CPU. Las arquitecturas GPU ofrecen capacidades de paralelización que pueden aprovecharse para aumentar el rendimiento de los sistemas de reconocimiento de voz. Sin embargo, la utilización eficiente de los recursos de la GPU para el reconocimiento de voz también es un desafío, ya que las implementaciones de software presentan accesos de memoria irregulares e impredecibles y una localidad temporal deficiente. El propósito de esta tesis es estudiar las características de los sistemas ASR que se ejecutan en dispositivos móviles de baja potencia para proponer diferentes técnicas para mejorar el rendimiento y el consumo de energía. Proponemos varias optimizaciones a nivel de software impulsadas por el análisis de potencia y rendimiento. A diferencia de las propuestas anteriores que intercambian precisión por el rendimiento al reducir el número de gaussianas evaluadas, mantenemos la precisión y mejoramos el rendimiento mediante el uso efectivo de la microarquitectura subyacente de la CPU. Usamos una implementación refactorizada del código de evaluación de GMM para reducir el impacto de las instrucciones de salto. Explotamos la unidad vectorial disponible en la mayoría de las CPU modernas para impulsar el cálculo de GMM. Además, calculamos las gaussianas para múltiples frames en paralelo, lo que reduce significativamente el uso de ancho de banda de memoria. Nuestros resultados experimentales muestran que las optimizaciones propuestas proporcionan un speedup de 2.68x sobre el decodificador Pocketsphinx en una CPU Intel Skylake de alta gama, mientras que logra un ahorro de energía del 61%. En segundo lugar, proponemos una técnica de renombrado de registros que explota la reutilización de registros físicos para reducir la presión sobre el banco de registros. Nuestra técnica aprovecha el uso compartido de registros físicos mediante la introducción de cambios en la tabla de renombrado de registros y la issue queue. Evaluamos nuestra técnica de renombrado sobre un procesador moderno. El esquema propuesto admite excepciones precisas y da como resultado mejoras de rendimiento del 9.5% para la evaluación GMM. Nuestros resultados experimentales muestran que el esquema de renombrado de registros propuesto proporciona un 6% de aceleración en promedio para SPEC2006. Finalmente, proponemos un acelerador para la evaluación de GMM que reduce el consumo de energía en tres órdenes de magnitud en comparación con soluciones basadas en CPU y GPU. El acelerador propuesto implementa un esquema de evaluación perezosa donde las GMMs se calculan bajo demanda, evitando el 50% de los cálculos. Finalmente, incluye un esquema de memorización que evita el 74.88% de las operaciones de coma flotante. El diseño final proporciona una aceleración de 164x y una reducción de energía de 3532x en comparación con una implementación altamente optimizada que se ejecuta en una CPU móvil moderna. Comparado con una GPU móvil de última generación, el acelerador de GMM logra un speedup de 5.89x sobre una implementación CUDA optimizada, mientras que reduce la energía en 241x.
APA, Harvard, Vancouver, ISO, and other styles
8

Elrod, JoAnn Broeckel, Raina Merchant, Mohamud Daya, Scott Youngquist, David Salcido, Terence Valenzuela, and Graham Nichol. "Public health surveillance of automated external defibrillators in the USA: protocol for the dynamic automated external defibrillator registry study." BMJ PUBLISHING GROUP, 2017. http://hdl.handle.net/10150/623946.

Full text
Abstract:
Introduction: Lay use of automated external defibrillators (AEDs) before the arrival of emergency medical services (EMS) providers on scene increases survival after out-of-hospital cardiac arrest (OHCA). AEDs have been placed in public locations may be not ready for use when needed. We describe a protocol for AED surveillance that tracks these devices through time and space to improve public health, and survival as well as facilitate research. Methods and analysis: Included AEDs are installed in public locations for use by laypersons to treat patients with OHCA before the arrival of EMS providers on scene. Included cases of OHCA are patients evaluated by organised EMS personnel and treated for OHCA. Enrolment of 10 000 AEDs annually will yield precision of 0.4% in the estimate of readiness for use. Enrolment of 2500 patients annually will yield precision of 1.9% in the estimate of survival to hospital discharge. Recruitment began on 21 Mar 2014 and is ongoing. AEDs are found by using multiple methods. Each AED is then tagged with a label which is a unique two-dimensional (2D) matrix code; the 2D matrix code is recorded and the location and status of the AED tracked using a smartphone; these elements are automatically passed via the internet to a secure and confidential database in real time. Whenever the 2D matrix code is rescanned for any non-clinical or clinical use of an AED, the user is queried to answer a finite set of questions about the device status. The primary outcome of any clinical use of an AED is survival to hospital discharge. Results are summarised descriptively. Ethics and dissemination: These activities are conducted under a grant of authority for public health surveillance from the Food and Drug Administration. Results are provided periodically to participating sites and sponsors to improve public health and quality of care.
APA, Harvard, Vancouver, ISO, and other styles
9

Giancoli, Ana Paula Müller. "Proposta de sistema para registro eletrônico de ponto com gerenciamento remoto." Universidade de Taubaté, 2011. http://www.bdtd.unitau.br/tedesimplificado/tde_busca/arquivo.php?codArquivo=243.

Full text
Abstract:
O presente trabalho propõe uma arquitetura de sistema para registro eletrônico de ponto baseado em software livre e capaz de atender os principais requisitos extraídos da portaria 1510 do Ministério do Trabalho. Essa arquitetura utiliza o sistema operacional Linux, a linguagem de programação Python, o framework Web Plone e o servidor de aplicações Zope, a fim de proporcionar, entre outros benefícios, a segurança, o acesso ao código fonte da aplicação e a independência de fornecedores. A validação é obtida por meio de testes práticos realizados em protótipo que adota os elementos do sistema. Os resultados satisfatórios obtidos nesses testes indicam que a aludida arquitetura é adequada para a aplicação em questão.
This study aims at proposing a system architecture for electronic clocking in and out based on free software to fulfill the main requirements extracted from the 1510 regulation from the Ministry of Labor. This architecture uses the Linux operating system; the Python programming language; the Web Plone framework and the Zope application server to provide, among other benefits, the security, the access to application source code, and the independence from suppliers. The validation is obtained though practical tests on prototypes that adopt the elements of the system. The satisfactory results obtained in these tests indicate that the aforementioned architecture is suitable for the application in question.
APA, Harvard, Vancouver, ISO, and other styles
10

Sánchez, Belenguer Carlos. "Surface Registration Techniques Applied to Archaeological Fragment Reconstruction." Doctoral thesis, Universitat Politècnica de València, 2015. http://hdl.handle.net/10251/56152.

Full text
Abstract:
[EN] Reconstruction of broken archaeological artifacts from fragments is a very time-consuming task that requires a big effort if performed manually. In fact, due to budgetary limitations, this is not even attempted in countless sites around the world, leaving vast quantities of material unstudied and stored indefinitely. This Thesis dissertation faces the application of surface registration techniques to the automatic re-assembly of broken archaeological artifacts from fragments. To efficiently do so, the reconstruction problem has been divided into two groups: 3 degrees of freedom and 6 degrees of freedom problems. This distinction is motivated for two major reasons: archaeological interest of the application and computational complexity of the solution. First kind of problems (3 degrees of freedom) deal with 2D objects or with flat 3D objects, like ripped-up documents or frescoes, respectively. In both cases, the mural paintings and engravings on the fragments' surface are of huge importance in the field of Cultural Heritage Recovery. In this sense, archaeologically speaking, the value of the reconstruction is not the final model itself, but the information stored in the upper surface. In terms of computation complexity, the reduced solution space allows using exhaustive techniques to ensure the quality of the results, while keeping execution times low. A fast hierarchical technique is introduced to face this kind of problems. Starting from an exhaustive search strategy, the technique progressively incorporates new features that lead to a hierarchical search strategy. Convergence and correction of the resulting technique are ensured using an optimistic cost function. Internal search calculations are optimized so the only operations performed are additions, subtractions and comparisons over aligned data. All heavy geometric operations are carried out by the GPU on a pre-processing stage that only happens once per fragment. Second kind of problems (6 degrees of freedom) deal with more general situations, where no special constraints are considered. Typical examples are broken sculptures, friezes, columns... In this case, computational complexity increases considerably with the extra 3 degrees of freedom, making exhaustive approaches prohibitive. To face this problems, an efficient sparse technique is introduced that uses a pre-processing stage to reduce the size of the problem: singular key-points in the original point cloud are selected based on a multi-scale feature extraction process driven by the saliency of each point. By computing a modified version of the PFH descriptor, the local neighborhood of each key- point is described in a compact histogram. Using exclusively the selected key-points and their associated descriptors, a very fast one-to-one search algorithm is executed for each possible pair of fragments. This process uses a three-level hierarchical search strategy driven by the local similarity between key-points, and applying a set of geometric consistence tests for intermediate results. Finally, a graph-based global registration algorithm uses all the individual matches to provide the final reconstruction of the artifact by creating clusters of matching fragments, appending new potential matches and joining individual clusters into bigger structures.
[ES] La reconstrucción de objetos arqueológicos fracturados a partir de fragmentos es una actividad que, si se realiza manualmente, supone un gran coste temporal. De hecho, debido a restricciones presupuestarias, esta tarea no llega a abordarse en incontables yacimientos arqueológicos, dejando grandes cantidades de material sin ser estudiado y almacenado indefinidamente. La presente propuesta de tesis aborda la aplicación de técnicas de registro de superficies a el re-ensamblado automático de objetos arqueológicos fracturados a partir de fragmentos. Por motivos de eficiencia, el problema de la reconstrucción se ha dividido en dos grupos: problemas de 3 grados de libertad y problemas de 6 grados de libertad. Esta distinción está motivada por dos razones: (1) el interés arqueológico de la aplicación final de las técnicas desarrolladas y (2) la complejidad computacional de la solución propuesta. El primer tipo de problemas (3 grados de libertad) se enfrenta a objetos bidimensionales o tridimensionales planos como documentos fragmentados y frescos, respectivamente. En ambos casos, los murales y grabados sobre la superficie de los fragmentos son de gran importancia en el ámbito de la conservación del patrimonio cultural. En este sentido, desde el punto de vista arqueológico, el valor de la reconstrucción final no radica en el modelo en sí, sino en la información almacenada sobre su superficie. En términos de complejidad computacional, el reducido espacio de soluciones permite emplear técnicas de búsqueda exhaustivas que garantizan la corrección de los resultados obtenidos con tiempos de ejecución acotados. La técnica propuesta para abordar este tipo de problemas parte de una estrategia exhaustiva y, progresivamente, incorpora nuevas optimizaciones que culminan con una técnica íntegramente jerárquica. La convergencia y corrección de la solución propuesta están garantizadas gracias a una función de coste optimista. Los cálculos internos durante las búsquedas han sido optimizados de modo que sólo son necesarias operaciones de adición/substracción y comparaciones sobre datos alineados en memoria. Todas las operaciones complejas asociadas a la manipulación de datos geométricos son realizadas por la GPU durante una etapa de pre-procesamiento que se ejecuta una sola vez por fragmento. El segundo tipo de problemas (6 grados de libertad) se enfrenta a situaciones más generales, en las que ninguna restricción especifica puede ser asumida. Ejemplos típicos son esculturas fragmentadas, frisos, columnas... En este caso, la complejidad computacional incrementa considerablemente debido a los 3 grados de libertad adicionales por lo que el coste temporal de las estrategias exhaustivas resulta prohibitivo. Para abordar este tipo de problemas, se propone una técnica dispersa eficiente apoyada en una fase de pre-procesamiento cuyo objetivo consiste en reducir la talla de los datos de entrada: a partir de las nubes de puntos originales, puntos clave singulares son identificados gracias a un proceso de extracción de características multi-escala apoyado en el valor de saliencia de cada punto. Mediante el cálculo de una versión modificada del descriptor PFH (Persistent Feature Histograms), el vecindario local de cada punto clave es descrito en un histograma compacto. Empleando únicamente estos puntos y sus descriptores asociados, un algoritmo de búsqueda uno-a-uno muy rápido se ejecuta sobre cada par de fragmentos. Dicho proceso emplea una estrategia de búsqueda jerárquica de tres niveles, dirigida por la similitud entre puntos clave y que aplica un conjunto de tests de consistencia geométrica sobre los resultados intermedios. Finalmente, un algoritmo de registro global toma como datos de entrada todas las correspondencias individuales para generar la reconstrucción final del objeto.
[CAT] La reconstrucció d'objectes arqueològics fracturats a partir de fragments és una activitat que, si es realitza manualment, suposa un gran cost temporal. De fet, a causa de restriccions pressupostàries, esta tasca no arriba a abordar-se en incomptables jaciments arqueològics, deixant grans quantitats de material sense ser estudiat i emmagatzemat indefinidament. La present proposta de tesi aborda l'aplicació de tècniques de registre de superfícies a l're-enssamblatge automàtic d'objectes arqueològics fracturats a partir de fragments. Per motius d'eficiència, el problema de la reconstrucció s'ha dividit en dos grups: problemes de 3 graus de llibertat i problemes de 6 graus de llibertat. Esta distinció està motivada per dues raons: (1) l'interès arqueològic de l'aplicació final de les tècniques desenvolupades i (2) la complexitat computacional de la solució proposada. El primer tipus de problemes (3 graus de llibertat) s'enfronta a objectes bidimensionals o tridimensionals plans com documents fragmentats i frescos, respectivament. En tots dos casos, els murals i gravats sobre la superfície dels fragments són de gran importància en l'àmbit de la conservació del patrimoni cultural. En este sentit, des del punt de vista arqueològic, el valor de la reconstrucció final no es basa en el model en si, sinó en la informació emmagatzemada sobre la seva superfície. En termes de complexitat computacional, el reduït espai de solucions permet emprar tècniques de recerca exhaustives que garanteixen la correcció dels resultats obtinguts amb temps d'execució acotats. La tècnica proposada per abordar aquest tipus de problemes part d'una estratègia exhaustiva i, progressivament, incorpora noves optimitzacions que culminen amb una tècnica íntegrament jeràrquica. La convergència i correcció de la solució proposada estan garantides gràcies a una funció de cost optimista. Els càlculs interns durant les recerques s'han optimitzat de manera que només són necessàries operacions d'addició / substracció i comparacions sobre dades alineats en memòria. Totes les operacions complexes associades a la manipulació de dades geomètriques són realitzades per la GPU durant una etapa de pre-processament que s'executa una única vegada per fragment. El segon tipus de problemes (6 graus de llibertat) s'enfronta a situacions més generals, en què cap restricció especifica pot ser assumida. Exemples típics són escultures fragmentades, frisos, columnes ... En este cas, la complexitat computacional s'incrementa considerablement a causa dels 3 graus de llibertat addicionals pel que el cost temporal de les estratègies exhaustives resulta prohibitiu. Per abordar este tipus de problemes, es proposa una tècnica dispersa eficient recolzada en una fase de pre-processament l'objectiu del qual consisteix a reduir la talla de les dades d'entrada: a partir dels núvols de punts originals, s'identifiquen punts clau singulars gràcies a un procés d'extracció de característiques multi-escala recolzat en el valor de saliència de cada punt. Mitjançant el càlcul d'una versió modificada del descriptor PFH (Persistent Feature Histograms), els veins locals de cada punt clau és descriuen en un histograma compacte. Emprant únicament estos punts i els seus descriptors associats, un algoritme de cerca un-a-un molt ràpid s'executa sobre cada parell de fragments. Aquest procés fa servir una estratègia de cerca jeràrquica de tres nivells, dirigida per la similitud entre punts clau i que aplica un conjunt de tests de consistència geomètrica sobre els resultats intermedis. Finalment, un algoritme de registre global pren com a dades d'entrada totes les correspondències individuals per generar la reconstrucció final de l'objecte.
Sánchez Belenguer, C. (2015). Surface Registration Techniques Applied to Archaeological Fragment Reconstruction [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/56152
TESIS
APA, Harvard, Vancouver, ISO, and other styles

Books on the topic "Register Automata"

1

United States. Congress. Office of Technology Assessment., ed. Automated record checks of firearm purchasers: Issues and options. Washington, D.C: Congress of the United States, Office of Technology Assessment, 1991.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
2

National Register of Foreign Collaborations (India) and India. Dept. of Scientific & Industrial Research., eds. Technology in Indian process control instruments and data acquisition systems industry: A status report prepared under the National Register of Foreign Collaborations. New Delhi: Govt. of India, Dept. of Scientific & Industrial Research, Ministry of Science and Technology, 1991.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
3

Scotland. Land Register of Scotland (Automated Registration) etc. Regulations 2014. Stationery Office, The, 2014.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
4

Scotland. Land Register of Scotland (Automated Registration) etc. Regulations 2014. Stationery Office, The, 2010.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
5

Ruiz, Antonio Lloris, Luis Parrilla Roure, Encarnación Castillo Morales, and Antonio García Ríos. Algebraic Circuits. Springer, 2014.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
6

Roure, Luis Parrilla, Antonio Lloris Lloris Ruiz, Encarnación Castillo Morales, and Antonio García Ríos. Algebraic Circuits. Springer, 2016.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
7

Ruiz, Antonio Lloris, Luis Parrilla Roure, Encarnación Castillo Morales, and Antonio García Ríos. Algebraic Circuits. Springer Berlin / Heidelberg, 2014.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
8

Ślusarski, Marek. Metody i modele oceny jakości danych przestrzennych. Publishing House of the University of Agriculture in Krakow, 2017. http://dx.doi.org/10.15576/978-83-66602-30-4.

Full text
Abstract:
The quality of data collected in official spatial databases is crucial in making strategic decisions as well as in the implementation of planning and design works. Awareness of the level of the quality of these data is also important for individual users of official spatial data. The author presents methods and models of description and evaluation of the quality of spatial data collected in public registers. Data describing the space in the highest degree of detail, which are collected in three databases: land and buildings registry (EGiB), geodetic registry of the land infrastructure network (GESUT) and in database of topographic objects (BDOT500) were analyzed. The results of the research concerned selected aspects of activities in terms of the spatial data quality. These activities include: the assessment of the accuracy of data collected in official spatial databases; determination of the uncertainty of the area of registry parcels, analysis of the risk of damage to the underground infrastructure network due to the quality of spatial data, construction of the quality model of data collected in official databases and visualization of the phenomenon of uncertainty in spatial data. The evaluation of the accuracy of data collected in official, large-scale spatial databases was based on a representative sample of data. The test sample was a set of deviations of coordinates with three variables dX, dY and Dl – deviations from the X and Y coordinates and the length of the point offset vector of the test sample in relation to its position recognized as a faultless. The compatibility of empirical data accuracy distributions with models (theoretical distributions of random variables) was investigated and also the accuracy of the spatial data has been assessed by means of the methods resistant to the outliers. In the process of determination of the accuracy of spatial data collected in public registers, the author’s solution was used – resistant method of the relative frequency. Weight functions, which modify (to varying degree) the sizes of the vectors Dl – the lengths of the points offset vector of the test sample in relation to their position recognized as a faultless were proposed. From the scope of the uncertainty of estimation of the area of registry parcels the impact of the errors of the geodetic network points was determined (points of reference and of the higher class networks) and the effect of the correlation between the coordinates of the same point on the accuracy of the determined plot area. The scope of the correction was determined (in EGiB database) of the plots area, calculated on the basis of re-measurements, performed using equivalent techniques (in terms of accuracy). The analysis of the risk of damage to the underground infrastructure network due to the low quality of spatial data is another research topic presented in the paper. Three main factors have been identified that influence the value of this risk: incompleteness of spatial data sets and insufficient accuracy of determination of the horizontal and vertical position of underground infrastructure. A method for estimation of the project risk has been developed (quantitative and qualitative) and the author’s risk estimation technique, based on the idea of fuzzy logic was proposed. Maps (2D and 3D) of the risk of damage to the underground infrastructure network were developed in the form of large-scale thematic maps, presenting the design risk in qualitative and quantitative form. The data quality model is a set of rules used to describe the quality of these data sets. The model that has been proposed defines a standardized approach for assessing and reporting the quality of EGiB, GESUT and BDOT500 spatial data bases. Quantitative and qualitative rules (automatic, office and field) of data sets control were defined. The minimum sample size and the number of eligible nonconformities in random samples were determined. The data quality elements were described using the following descriptors: range, measure, result, and type and unit of value. Data quality studies were performed according to the users needs. The values of impact weights were determined by the hierarchical analytical process method (AHP). The harmonization of conceptual models of EGiB, GESUT and BDOT500 databases with BDOT10k database was analysed too. It was found that the downloading and supplying of the information in BDOT10k creation and update processes from the analyzed registers are limited. An effective approach to providing spatial data sets users with information concerning data uncertainty are cartographic visualization techniques. Based on the author’s own experience and research works on the quality of official spatial database data examination, the set of methods for visualization of the uncertainty of data bases EGiB, GESUT and BDOT500 was defined. This set includes visualization techniques designed to present three types of uncertainty: location, attribute values and time. Uncertainty of the position was defined (for surface, line, and point objects) using several (three to five) visual variables. Uncertainty of attribute values and time uncertainty, describing (for example) completeness or timeliness of sets, are presented by means of three graphical variables. The research problems presented in the paper are of cognitive and application importance. They indicate on the possibility of effective evaluation of the quality of spatial data collected in public registers and may be an important element of the expert system.
APA, Harvard, Vancouver, ISO, and other styles
9

Building an Electronic Disease Register (Primary Care Health Informatics). Radcliffe Publishing Ltd, 2001.

Find full text
APA, Harvard, Vancouver, ISO, and other styles
10

Lowe, Nick, Bev Ellis, and Alan Gillies. Building an Electronic Disease Register: Getting the Computer to Work for You. Taylor & Francis Group, 2018.

Find full text
APA, Harvard, Vancouver, ISO, and other styles

Book chapters on the topic "Register Automata"

1

Tzevelekos, Nikos, and Radu Grigore. "History-Register Automata." In Lecture Notes in Computer Science, 17–33. Berlin, Heidelberg: Springer Berlin Heidelberg, 2013. http://dx.doi.org/10.1007/978-3-642-37075-5_2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

D’Antoni, Loris, Tiago Ferreira, Matteo Sammartino, and Alexandra Silva. "Symbolic Register Automata." In Computer Aided Verification, 3–21. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-25540-4_1.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Howar, Falk, Bernhard Steffen, Bengt Jonsson, and Sofia Cassel. "Inferring Canonical Register Automata." In Lecture Notes in Computer Science, 251–66. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-27940-9_17.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Aarts, Fides, Falk Howar, Harco Kuppens, and Frits Vaandrager. "Algorithms for Inferring Register Automata." In Leveraging Applications of Formal Methods, Verification and Validation. Technologies for Mastering Change, 202–19. Berlin, Heidelberg: Springer Berlin Heidelberg, 2014. http://dx.doi.org/10.1007/978-3-662-45234-9_15.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Merten, Maik, Falk Howar, Bernhard Steffen, Sofia Cassel, and Bengt Jonsson. "Demonstrating Learning of Register Automata." In Tools and Algorithms for the Construction and Analysis of Systems, 466–71. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-28756-5_32.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Murawski, Andrzej S., Steven J. Ramsay, and Nikos Tzevelekos. "Reachability in Pushdown Register Automata." In Mathematical Foundations of Computer Science 2014, 464–73. Berlin, Heidelberg: Springer Berlin Heidelberg, 2014. http://dx.doi.org/10.1007/978-3-662-44522-8_39.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Senda, Ryoma, Yoshiaki Takata, and Hiroyuki Seki. "Generalized Register Context-Free Grammars." In Language and Automata Theory and Applications, 259–71. Cham: Springer International Publishing, 2019. http://dx.doi.org/10.1007/978-3-030-13435-8_19.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Garhewal, Bharat, Frits Vaandrager, Falk Howar, Timo Schrijvers, Toon Lenaerts, and Rob Smits. "Grey-Box Learning of Register Automata." In Lecture Notes in Computer Science, 22–40. Cham: Springer International Publishing, 2020. http://dx.doi.org/10.1007/978-3-030-63461-2_2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Abdulla, Parosh Aziz, Mohamed Faouzi Atig, Ahmet Kara, and Othmane Rezine. "Verification of Buffered Dynamic Register Automata." In Networked Systems, 15–31. Cham: Springer International Publishing, 2015. http://dx.doi.org/10.1007/978-3-319-26850-7_2.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Grigore, Radu, Dino Distefano, Rasmus Lerchedahl Petersen, and Nikos Tzevelekos. "Runtime Verification Based on Register Automata." In Tools and Algorithms for the Construction and Analysis of Systems, 260–76. Berlin, Heidelberg: Springer Berlin Heidelberg, 2013. http://dx.doi.org/10.1007/978-3-642-36742-7_19.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Conference papers on the topic "Register Automata"

1

Tzevelekos, Nikos. "Fresh-register automata." In the 38th annual ACM SIGPLAN-SIGACT symposium. New York, New York, USA: ACM Press, 2011. http://dx.doi.org/10.1145/1926385.1926420.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Jie Fu and H. G. Tanner. "Optimal planning on register automata." In 2012 American Control Conference - ACC 2012. IEEE, 2012. http://dx.doi.org/10.1109/acc.2012.6315508.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Murawski, Andrzej S., Steven J. Ramsay, and Nikos Tzevelekos. "Bisimilarity in Fresh-Register Automata." In 2015 30th Annual ACM/IEEE Symposium on Logic in Computer Science (LICS). IEEE, 2015. http://dx.doi.org/10.1109/lics.2015.24.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Chen, Yu-Fang, Ondrej Lengal, Tony Tan, and Zhilin Wu. "Register automata with linear arithmetic." In 2017 32nd Annual ACM/IEEE Symposium on Logic in Computer Science (LICS). IEEE, 2017. http://dx.doi.org/10.1109/lics.2017.8005111.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Segoufin, Luc, and Victor Vianu. "Projection Views of Register Automata." In SIGMOD/PODS '20: International Conference on Management of Data. New York, NY, USA: ACM, 2020. http://dx.doi.org/10.1145/3375395.3387651.

Full text
APA, Harvard, Vancouver, ISO, and other styles
6

Touili, Tayssir. "Register Automata for Malware Specification." In ARES 2022: The 17th International Conference on Availability, Reliability and Security. New York, NY, USA: ACM, 2022. http://dx.doi.org/10.1145/3538969.3544442.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Alur, Rajeev, Loris DAntoni, Jyotirmoy Deshmukh, Mukund Raghothaman, and Yifei Yuan. "Regular Functions and Cost Register Automata." In 2013 Twenty-Eighth Annual IEEE/ACM Symposium on Logic in Computer Science (LICS 2013). IEEE, 2013. http://dx.doi.org/10.1109/lics.2013.65.

Full text
APA, Harvard, Vancouver, ISO, and other styles
8

Bojanczyk, Mikolaj, Bartek Klin, and Joshua Moerman. "Orbit-Finite-Dimensional Vector Spaces and Weighted Register Automata." In 2021 36th Annual ACM/IEEE Symposium on Logic in Computer Science (LICS). IEEE, 2021. http://dx.doi.org/10.1109/lics52264.2021.9470634.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Figueira, Diego. "Forward-XPath and extended register automata on data-trees." In the 13th International Conference. New York, New York, USA: ACM Press, 2010. http://dx.doi.org/10.1145/1804669.1804699.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Daviaud, Laure, Pierre-Alain Reynier, and Jean-Marc Talbot. "A Generalised Twinning Property for Minimisation of Cost Register Automata." In LICS '16: 31st Annual ACM/IEEE Symposium on Logic in Computer Science. New York, NY, USA: ACM, 2016. http://dx.doi.org/10.1145/2933575.2934549.

Full text
APA, Harvard, Vancouver, ISO, and other styles

Reports on the topic "Register Automata"

1

Perez, Jorge E., and Vijay K. Madisetti. Integrated Automatic Target Detection from Pixel-Registered Visual-Thermal-Range Images. Fort Belvoir, VA: Defense Technical Information Center, January 1998. http://dx.doi.org/10.21236/ada358524.

Full text
APA, Harvard, Vancouver, ISO, and other styles
2

Neeley, Aimee, Stace E. Beaulieu, Chris Proctor, Ivona Cetinić, Joe Futrelle, Inia Soto Ramos, Heidi M. Sosik, et al. Standards and practices for reporting plankton and other particle observations from images. Woods Hole Oceanographic Institution, July 2021. http://dx.doi.org/10.1575/1912/27377.

Full text
Abstract:
This technical manual guides the user through the process of creating a data table for the submission of taxonomic and morphological information for plankton and other particles from images to a repository. Guidance is provided to produce documentation that should accompany the submission of plankton and other particle data to a repository, describes data collection and processing techniques, and outlines the creation of a data file. Field names include scientificName that represents the lowest level taxonomic classification (e.g., genus if not certain of species, family if not certain of genus) and scientificNameID, the unique identifier from a reference database such as the World Register of Marine Species or AlgaeBase. The data table described here includes the field names associatedMedia, scientificName/ scientificNameID for both automated and manual identification, biovolume, area_cross_section, length_representation and width_representation. Additional steps that instruct the user on how to format their data for a submission to the Ocean Biodiversity Information System (OBIS) are also included. Examples of documentation and data files are provided for the user to follow. The documentation requirements and data table format are approved by both NASA’s SeaWiFS Bio-optical Archive and Storage System (SeaBASS) and the National Science Foundation’s Biological and Chemical Oceanography Data Management Office (BCO-DMO).
APA, Harvard, Vancouver, ISO, and other styles
3

McCarthy, Sean T., Aneesa Motala, Emily Lawson, and Paul G. Shekelle. Prevention in Adults of Transmission of Infection With Multidrug-Resistant Organisms. Rapid Review. Agency for Healthcare Research and Quality (AHRQ), April 2024. http://dx.doi.org/10.23970/ahrqepc_mhs4mdro.

Full text
Abstract:
Objectives. This rapid review summarizes literature for patient safety practices intended to prevent and control the transmission of multidrug-resistant organisms (MDROs). Methods. We followed rapid review processes of the Agency for Healthcare Research and Quality Evidence-based Practice Center Program. We searched PubMed to identify eligible systematic reviews from 2011 to May 2023 and primary studies published from 2011 to May 2023, supplemented by targeted gray literature searches. We included literature that addressed patient safety practices intending to prevent or control transmission of MDROs which were implemented in hospitals and nursing homes and that included clinical outcomes of infection or colonization with MDROs as well as unintended consequences such as mental health effects and noninfectious adverse healthcare-associated outcomes. The protocol for the review has been registered in PROSPERO (CRD42023444973). Findings. Our search retrieved 714 citations, of which 42 articles were eligible for review. Systematic reviews, which were primarily of observational studies, included a wide variety of infection prevention and control (IPC) practices, including universal gloving, contact isolation precautions, adverse effects of patient isolation, patient and/or staff cohorting, room decontamination, patient decolonization, IPC practices specifically in nursing homes, features of organizational culture to facilitate implementation of IPC practices and the role of dedicated IPC staff. While systematic reviews were of good or fair quality, strength of evidence for the conclusions was always low or very low, due to reliance on observational studies. Decolonization strategies showed some benefit in certain populations, such as nursing home patients and patients discharging from acute care hospitalization. Universal gloving showed a small benefit in the intensive care unit. Contact isolation targeting patients colonized or infected with MDROs showed mixed effects in the literature and may be associated with mental health and noninfectious (e.g., falls and pressure ulcers) adverse effects when compared with standard precautions, though based on before/after studies in which such precautions were ceased. There was no significant evidence of benefit for patient cohorting (except possibly in outbreak settings), automated room decontamination or cleaning feedback protocols, and IPC practices in long-term settings. Infection rates may be improved when IPC practices are implemented in the context of certain logistical and staffing characteristics including a supportive organizational culture, though again strength of evidence was low. Dedicated infection prevention staff likely improve compliance with other patient safety practices, though there is little evidence of their downstream impact on rates of infection. Conclusions. Selected infection prevention and control interventions had mixed evidence for reducing healthcare-associated infection and colonization by multidrug resistant organisms. Where these practices did show benefit, they often had evidence that applied only to certain subpopulations (such as intensive care unit patients), though overall strength of evidence was low.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography