Dissertations / Theses on the topic 'Batches of task graphs'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 15 dissertations / theses for your research on the topic 'Batches of task graphs.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.
Toch, Lamiel. "Contributions aux techniques d’ordonnancement sur plates-formes parallèles ou distribuées." Electronic Thesis or Diss., Besançon, 2012. http://www.theses.fr/2012BESA2045.
Full textWorks presented in this document tackle scheduling of parallel applications in either parallel (cluster) or distributed (computing grid) platforms. In our researches we were concentrated on either scheduling of applications modeled by a DAG, directed acyclic graph, for computing grid or scheduling of parallel programs (parallel jobs) represented by a rectangular shape whose the two dimensions are the number of requested processors and the execution time. The researches follow three main topics. The first topic concerns the scheduling of a set of instances of an application for computing grid. The second topic deals with the scheduling of parallel jobs inclusters. The third one tackles the scheduling of parallel jobs in multiprocessor machines. We brought contributions on these three topics. The first contribution under the first topic consists of the advanced experimental study of three algorithms for scheduling a set of instances of an application on a heterogeneous platform without communication costs : a list-based algorithm, a steady-state algorithm and genetic algorithm. Moreover we integrate communications in this genetic algorithm. The second contribution under the second topic is the design of a new technique for scheduling parallel jobs in clusters : job folding which uses virtualization of processors. The third contribution deals with a new technique which comes from statistics and signal cessing applied to scheduling of parallel jobs in a multiprocessor machine. Eventually we givesome works that we carried out but which did not give significant results for scheduling
Dechu, Satish. "Task graphs mapping on to network processors using simulated annealing /." Available to subscribers only, 2007. http://proquest.umi.com/pqdweb?did=1453188941&sid=15&Fmt=2&clientId=1509&RQT=309&VName=PQD.
Full textNegelspach, Greg L. "Grain size management in repetitive task graphs for multiprocessor computer scheduling." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 1994. http://handle.dtic.mil/100.2/ADA288575.
Full textChieregato, Federico. "Modelling task execution time in Directed Acyclic Graphs for efficient distributed management." Master's thesis, Alma Mater Studiorum - Università di Bologna, 2022.
Find full textWitt, Carl Philipp. "Predictive Resource Management for Scientific Workflows." Doctoral thesis, Humboldt-Universität zu Berlin, 2020. http://dx.doi.org/10.18452/21608.
Full textScientific experiments produce data at unprecedented volumes and resolutions. For the extraction of insights from large sets of raw data, complex analysis workflows are necessary. Scientific workflows enable such data analyses at scale. To achieve scalability, most workflow management systems are designed as an additional layer on top of distributed resource managers, such as batch schedulers or distributed data processing frameworks. However, like distributed resource managers, they do not automatically determine the amount of resources required for executing individual tasks in a workflow. The status quo is that workflow management systems delegate the challenge of estimating resource usage to the user. This limits the performance and ease-of-use of scientific workflow management systems, as users often lack the time, expertise, or incentives to estimate resource usage accurately. This thesis is an investigation of how to learn and predict resource usage during workflow execution. In contrast to prior work, an integrated perspective on prediction and scheduling is taken, which introduces various challenges, such as quantifying the effects of prediction errors on system performance. The main contributions are: 1. A survey of peak memory usage prediction in batch processing environments. It provides an overview of prior machine learning approaches, commonly used features, evaluation metrics, and data sets. 2. A static workflow scheduling method that uses statistical methods to predict which scheduling decisions can be improved. 3. A feedback-based approach to scheduling and predictive resource allocation, which is extensively evaluated using simulation. The results provide insights into the desirable characteristics of scheduling heuristics and prediction models. 4. A prediction model that reduces memory wastage. The design takes into account the asymmetric costs of overestimation and underestimation, as well as follow up costs of prediction errors.
Kasinger, Charles D. "A periodic scheduling heuristic for mapping iterative task graphs onto distributed memory multiprocessors." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 1994. http://handle.dtic.mil/100.2/ADA286047.
Full textKoman, Charles Brian. "A tool for efficient execution and development of repetitive task graphs on a distributed memory multiprocessor." Thesis, Monterey, Calif. : Springfield, Va. : Naval Postgraduate School ; Available from National Technical Information Service, 1995. http://handle.dtic.mil/100.2/ADA305995.
Full textPop, Ruxandra. "Mapping Concurrent Applications to Multiprocessor Systems with Multithreaded Processors and Network on Chip-Based Interconnections." Licentiate thesis, Linköpings universitet, Institutionen för datavetenskap, 2011. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-64256.
Full textGurhem, Jérôme. "Paradigmes de programmation répartie et parallèle utilisant des graphes de tâches pour supercalculateurs post-pétascale." Thesis, Lille, 2021. http://www.theses.fr/2021LILUI005.
Full textSince the middle of the 1990s, message passing libraries are the most used technology to implement parallel and distributed applications. However, they may not be a solution efficient enough on exascale machines since scalability issues will appear due to the increase in computing resources. Task-based programming models can be used, for example, to avoid collective communications along all the resources like reductions, broadcast or gather by transforming them into multiple operations on tasks. Then, these operations can be scheduled by the scheduler to place the data and computations in a way that optimize and reduce the data communications. The main objective of this thesis is to study what must be task-based programming for scientific applications and to propose a specification of such distributed and parallel programming, by experimenting for several simplified representations of important scientific applications for TOTAL, and classical dense and sparse linear methods.During the dissertation, several programming languages and paradigms are studied. Dense linear methods to solve linear systems, sequences of sparse matrix vector product and the Kirchhoff seismic pre-stack depth migration are studied and implemented as task-based applications. A taxonomy, based on several of these languages and paradigms is proposed.Software were developed using these programming models for each simplified application. As a result of these researches, a methodology for parallel task programming is proposed, optimizing data movements, in general, and for targeted scientific applications, in particular
Bouguelia, Sara. "Modèles de dialogue et reconnaissance d'intentions composites dans les conversations Utilisateur-Chatbot orientées tâches." Electronic Thesis or Diss., Lyon 1, 2023. http://www.theses.fr/2023LYO10106.
Full textDialogue Systems (or simply chatbots) are in very high demand these days. They enable the understanding of user needs (or user intents), expressed in natural language, and on fulfilling such intents by invoking the appropriate back-end APIs (Application Programming Interfaces). Chatbots are famed for their easy-to-use interface and gentle learning curve (it only requires one of humans' most innate ability, the use of natural language). The continuous improvement in Artificial Intelligence (AI), Natural Language Processing (NLP), and the countless number of devices allow performing real-world tasks (e.g., making a reservation) by using natural language-based interactions between users and a large number of software enabled services.Nonetheless, chatbot development is still in its preliminary stage, and there are several theoretical and technical challenges that need to be addressed. One of the challenges stems from the wide range of utterance variations in open-end human-chatbot interactions. Additionally, there is a vast space of software services that may be unknown at development time. Natural human conversations can be rich, potentially ambiguous, and express complex and context-dependent intents. Traditional business process and service composition modeling and orchestration techniques are limited to support such conversations because they usually assume a priori expectation of what information and applications will be accessed and how users will explore these sources and services. Limiting conversations to a process model means that we can only support a small fraction of possible conversations. While existing advances in NLP and Machine Learning (ML) techniques automate various tasks such as intent recognition, the synthesis of API calls to support a broad range of potentially complex user intents is still largely a manual, ad-hoc and costly process.This thesis project aims at advancing the fundamental understanding of cognitive services engineering. In this thesis we contribute novel abstractions and techniques focusing on the synthesis of API calls to support a broad range of potentially complex user intents. We propose reusable and extensible techniques to recognize and realize complex intents during humans-chatbots-services interactions. These abstractions and techniques seek to unlock the seamless and scalable integration of natural language-based conversations with software-enabled services
Lai, Lein-Fu, and 賴聯福. "Task-Based Specifications through Conceptual Graphs." Thesis, 1995. http://ndltd.ncl.edu.tw/handle/78938169182813236389.
Full text國立中央大學
資訊及電子工程研究所
83
In this paper, we have proposed the use of conceptual graphs to express task-based specifications in which the specification is driven by the task structure of problem solving knowledge, pieces of the specification can be refined iteratively and verification is performed for a single layer or between layers. Issues in mapping task-based specifications into conceptual graphs are identified, for example, the representation of constraints, the relationship between a task, its constraints and state model, rigid and soft postconditions, the distinction between follow and immediately follow operators, and the composition operator in task state expressions. To alleviate the problems, the notion of demons has been adopted to represent task state expressions as well as state models, constraint overlays have been used for describing the state model, and canonical formation rules have been applied to compose task state expressions. These are illustrated using the problem domain of R1/SOAR. In addition, we proposed the verification for both the model specification and the process specification. Constraint networks, constraint satisfaction, task progression and comparision for specificity are used to verify task-based specifications.
Chen, Lu-Ann, and 陳履安. "A Study on the Parallel Task Scheduling for Directed Acyclic Graphs." Thesis, 1997. http://ndltd.ncl.edu.tw/handle/59v564.
Full text國立清華大學
資訊科學研究所
85
The DAGs (Directed Acyclic Graphs) are directed graphs with no cycle on it. The nodes on such a graph are partial ordered. Each node and each edge have their weights as computation or communication overheads. Some properties of nodes can be used to help scheduler decide the importance of the nodes. The scheduler used the properties to schedule the entire porojects off-line or during compile time. This kind of scheduling is so called static scheduling or DAG scheduling. The problem of non-preemptively scheduling a set of n partially ordered tasks on p identical processors is a rich land of research and lots of literature has been produced. However, the higher quality the scheduling results are, the higher time complexity the algorithm has. The Earliest Time First (ETF) algorithm is the first algorithm which takes communication delays between processes into considerations. Its result is pretty good. Its the complexity is O(n2p), where the n and p are the number of tasks and processors. Lately a method called as Fast Assignment using Search Technique (FAST) algorithm with takes O(e) complexity was proposed, where e is the number of edges in the input graph. The FAST algorithm uses static scheduling technique as its initial scheduling technique, and then modify its initial scheduling with the local search technique. In this thesis, a new algorithm with O(e) time complexity for DAG scheduling is proposed. The proposed one uses a new method called Update Parents (UP) and a different priority assignment method for each task from the FAST. A improvenment for the ETF algorithm is also proposed. The improved method reduces the time complexity of ETF to O(n2). Let the running time denotes the time required for a program to complete a algorithm, and scheduled time or makespan denotes the scheduled result generated by a algorithm. In the testbed that processor number is 40 and task number is about 432, the proposed algorithm runs 6% slower than the FAST but the makespan is better by 1.3%. In the same environment, the ETF is slower than the FAST by 677% and the result is better by 1.88%. The improved ETF runs faster than the original one by 35.7%.
Fu, Lai Lien, and 賴聯福. "Task-Based Conceptual Graphs as a Basis for Automated Software Engineering." Thesis, 1999. http://ndltd.ncl.edu.tw/handle/75973147143239865136.
Full text國立中央大學
資訊工程研究所
87
Traditional software paradigm contains two fundamental flaws which exacerbate the maintenance of software systems: (1) there is no technology for managing the knowledge-intensive activities which constitute the software development processes, and (2) maintenance is performed on source code. Therefore, automated software engineering is used to directly obviate both these fundamental flaws through full automation of the compilation process. Automated software engineering provides the automation of software development processes including specification acquisition, verification, automatic programming, and maintenance. The automation avoids people producing and maintaining software, which reduce cost and development times while increasing reliability. As was pointed out by Rich and Waters, there are three technical issues in automatic software engineering. First, the specification language used to communicate users and the system can be a natural language, special-purpose languages, logical formalisms, and very high level languages. Logic is the most powerful formal description language known. As a result, it is reasonable to suppose that it might make a good communication medium between users and the system. Unfortunately, there are two fundamental barriers to the use of logical formalisms: it is computationally intractable and is difficult for users to write and understand. Second, the approach to mapping from the specification into the implementation can be procedural methods, deductive methods, transformation methods, and inspection methods. A major strength of transformational methods is that they provide a very clear representation for certain kinds of programming knowledge. For this reason, transformational methods in some form dominate current research in automatic programming. Finally, the system must have domain knowledge to interpret the terms used by users and programming knowledge to produce programs. Representing and using such knowledge is a major challenge. The knowledge representation should have several properties: (1) expressiveness: the representation must be able to express as many different kinds of knowledge as possible, (2) convenient combination: the combination of different kinds knowledge must be easy to implement, (3) semantic soundness: the representation must be based on a mathematical foundation, and (4) manipulability: the representation must be easy to be manipulated for the purpose of improvement. Conceptual graphs as a general knowledge representation formalism has been widely adopted in knowledge and software engineering. As a first-order logic vehicle, conceptual graphs are powerful in describing static facts, constraints, and relationships in the real world. With its direct mapping to natural languages, the conceptual graph can serve as an intermediate language for translating computer-oriented formalisms to and from natural languages. With its graphic representation and mapping to first order logic, it can serve as a readable and formal specification language. Though conceptual graphs has the potential to be an excellent formalism for automated software engineering, the limitation in performing computation poses significant barriers to the automatic execution of software systems through conceptual graphs. Hence, a method for operationalizing conceptual graphs needs to be worked out. In this dissertation, task-based conceptual graphs (TBCG) is used as a basis for automating the software development processes. We use task-based conceptual graphs to capture and represent a conceptual model for the problem domain. To construct a conceptual model, task-based specification methodology is used to serve as the mechanism to structure the knowledge captured in conceptual models; whereas, conceptual graphs (CGs) are adopted as the formalism to express task-based specifications and to provide a reasoning capability for the purpose of automation. Contextual retrieval mechanism proposes a multiple-viewed approach to help to acquire specifications and elicit information in contexts. To automate the reuse of existing task-based conceptual graphs, contextual retrieval mechanism uses fuzzy similarity to automatically match two surrounding contexts of specifications and utilizes purpose-directed analogy for retrieving reusable specifications. The verification of task-based conceptual graphs is performed on: (1) model specifications of a task by checking the consistency of constraints networks through the constraint satisfaction algorithm, and by repairing constraint violation using the constraint relaxation method; and (2) process specifications of a task by checking the interlevel and intralevel consistency based on operators and rules of inference inherited in conceptual graphs. Once task-based conceptual graphs have been constructed and verified, a blackboard system will automatically transform TBCG specifications into an executable program in CLIPS. We can then perform maintenance directly on the specification rather than on the source code. Small logical changes may have large and complex effects while performing maintenance at the implementation level; whereas, maintenance at the specification level is fairly simple and explainable. Hence, the problems in traditional software paradigm are avoided by automatic reimplementation from the revised specifications. Our approach offers several benefits that are useful for automating software development processes. First, the use of conceptual graphs together with task-based specifications in specifying software requirements helps in capturing richer semantics than that of task-based specifications or conceptual graphs alone. The expressive power of conceptual graphs facilitates the capturing of semantics that task-based specifications is difficult to express. Second, requirements specifications for different views are represented in their conceptual graphical specifications, and are tightly integrated under the general notion of tasks. In addition, artifacts constructed in each model are sharable. Third, contextual retrieval mechanism provides an incremental context acquisition approach to help elicit detailed information of contexts. The efficiency of reusing requirements specification is also increased by excluding out irrelevant contexts to reduce the number of possible candidates for selection. Fourth, the computation of fuzzy similarity not only facilitates the matching between two sets of conceptual graphs but also deals with the uncertainty inherent to contexts that cannot be completely described. Finally, conceptual graphs is operationalized by providing automatic translation into CLIPS. Conceptual graphs can then have the capability to perform computation, solve problems, and simulating processes.
Ojha, Prakhar. "Utilizing Worker Groups And Task Dependencies in Crowdsourcing." Thesis, 2017. http://etd.iisc.ac.in/handle/2005/4265.
Full textRavindran, Rajeswaran Chockalingapuram. "Scheduling Heuristics for Maximizing the Output Quality of Iris Task Graphs in Multiprocessor Environment with Time and Energy Bounds." 2012. https://scholarworks.umass.edu/theses/826.
Full text