Auswahl der wissenschaftlichen Literatur zum Thema „Distributed Parallel Application“

Geben Sie eine Quelle nach APA, MLA, Chicago, Harvard und anderen Zitierweisen an

Wählen Sie eine Art der Quelle aus:

Machen Sie sich mit den Listen der aktuellen Artikel, Bücher, Dissertationen, Berichten und anderer wissenschaftlichen Quellen zum Thema "Distributed Parallel Application" bekannt.

Neben jedem Werk im Literaturverzeichnis ist die Option "Zur Bibliographie hinzufügen" verfügbar. Nutzen Sie sie, wird Ihre bibliographische Angabe des gewählten Werkes nach der nötigen Zitierweise (APA, MLA, Harvard, Chicago, Vancouver usw.) automatisch gestaltet.

Sie können auch den vollen Text der wissenschaftlichen Publikation im PDF-Format herunterladen und eine Online-Annotation der Arbeit lesen, wenn die relevanten Parameter in den Metadaten verfügbar sind.

Zeitschriftenartikel zum Thema "Distributed Parallel Application"

1

Spahi, Enis, und D. Altilar. „ITU-PRP: Parallel and Distributed Computing Middleware for Java Developers“. International Journal of Business & Technology 3, Nr. 1 (November 2014): 2–13. http://dx.doi.org/10.33107/ijbte.2014.3.1.01.

Der volle Inhalt der Quelle
Annotation:
ITU-PRP provides a Parallel Programming Framework for Java Developers on which they can adapt their sequential application code to operate on a distributed multi-host parallel environment. Developers would implement parallel models, such as Loop Parallelism, Divide and Conquer, Master-Slave and Fork-Join by the help of an API Library provided under framework. Produced parallel applications would be submitted to a middleware called Parallel Running Platform (PRP), on which parallel resources for parallel processing are being organized and performed. The middleware creates Task Plans (TP) according to application’s parallel model, assigns best available resource Hosts, in order to perform fast parallel processing. Task Plans will be created dynamically in real time according to resources actual utilization status or availability, instead of predefined/preconfigured task plans. ITU-PRP achieves better efficiency on parallel processing over big data sets and distributes divided base data to multiple hosts to be operated by Coarse-Grained parallelism. According to this model distributed parallel tasks would operate independently with minimal interaction until processing ends.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
2

Sikora, Andrzej, und Ewa Niewiadomska-Szynkiewicz. „Parallel and Distributed Simulation of Ad Hoc Networks“. Journal of Telecommunications and Information Technology, Nr. 3 (26.06.2023): 76–84. http://dx.doi.org/10.26636/jtit.2009.3.943.

Der volle Inhalt der Quelle
Annotation:
Modeling and simulation are traditional methods used to evaluate wireless network design. This paper addresses issues associated with the application of parallel discrete event simulation to mobile ad hoc networks design and analysis. The basic characteristics and major issues pertaining to ad hoc networks modeling and simulation are introduced. The focus is on wireless transmission and mobility models. Particular attention is paid to the MobASim system, a Javabased software environment for parallel and distributed simulation of mobile ad hoc networks. We describe the design, performance and possible applications of presented simulation software.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
3

César, Eduardo, Anna Morajko, Tomàs Margalef, Joan Sorribes, Antonio Espinosa und Emilio Luque. „Dynamic Performance Tuning Supported by Program Specification“. Scientific Programming 10, Nr. 1 (2002): 35–44. http://dx.doi.org/10.1155/2002/549617.

Der volle Inhalt der Quelle
Annotation:
Performance analysis and tuning of parallel/distributed applications are very difficult tasks for non-expert programmers. It is necessary to provide tools that automatically carry out these tasks. These can be static tools that carry out the analysis on a post-mortem phase or can tune the application on the fly. Both kind of tools have their target applications. Static automatic analysis tools are suitable for stable application while dynamic tuning tools are more appropriate to applications with dynamic behaviour. In this paper, we describe KappaPi as an example of a static automatic performance analysis tool, and also a general environment based on parallel patterns for developing and dynamically tuning parallel/distributed applications.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
4

VADHIYAR, SATHISH S., und JACK J. DONGARRA. „SRS: A FRAMEWORK FOR DEVELOPING MALLEABLE AND MIGRATABLE PARALLEL APPLICATIONS FOR DISTRIBUTED SYSTEMS“. Parallel Processing Letters 13, Nr. 02 (Juni 2003): 291–312. http://dx.doi.org/10.1142/s0129626403001288.

Der volle Inhalt der Quelle
Annotation:
The ability to produce malleable parallel applications that can be stopped and reconfigured during the execution can offer attractive benefits for both the system and the applications. The reconfiguration can be in terms of varying the parallelism for the applications, changing the data distributions during the executions or dynamically changing the software components involved in the application execution. In distributed and Grid computing systems, migration and reconfiguration of such malleable applications across distributed heterogeneous sites which do not share common file systems provides flexibility for scheduling and resource management in such distributed environments. The present reconfiguration systems do not support migration of parallel applications to distributed locations. In this paper, we discuss a framework for developing malleable and migratable MPI message-passing parallel applications for distributed systems. The framework includes a user-level checkpointing library called SRS and a runtime support system that manages the checkpointed data for distribution to distributed locations. Our experiments and results indicate that the parallel applications, with instrumentation to SRS library, were able to achieve reconfigurability incurring about 15-35% overhead.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
5

Ogle, D. M., K. Schwan und R. Snodgrass. „Application-dependent dynamic monitoring of distributed and parallel systems“. IEEE Transactions on Parallel and Distributed Systems 4, Nr. 7 (Juli 1993): 762–78. http://dx.doi.org/10.1109/71.238299.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
6

Čiegis, Raimondas, Vadimas Starikovičius, Natalija Tumanova und Minvydas Ragulskis. „Application of distributed parallel computing for dynamic visual cryptography“. Journal of Supercomputing 72, Nr. 11 (04.05.2016): 4204–20. http://dx.doi.org/10.1007/s11227-016-1733-8.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
7

Chi, Song. „Application of MATLAB Parallel Programming Technology“. Applied Mechanics and Materials 602-605 (August 2014): 3787–90. http://dx.doi.org/10.4028/www.scientific.net/amm.602-605.3787.

Der volle Inhalt der Quelle
Annotation:
The parallel application program development technology and process is analyzed based on the MATLAB parallel and distributed computing toolbox. Hereby, the comparison of the serial computing and the parallel computing is done by computer simulations, and the parallel computing program design and develop method is proposed. The simulations results show that, the parallel computing technology has many advantages in the calculation of high intensive and it is convenience of the parallel application development using MATLAB.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
8

Prylli, L., und B. Tourancheau. „Execution-Driven Simulation of Parallel Applications“. Parallel Processing Letters 08, Nr. 01 (März 1998): 95–109. http://dx.doi.org/10.1142/s0129626498000122.

Der volle Inhalt der Quelle
Annotation:
This paper presents our work on the simulation of distributed memory parallel computers. We design a distributed simulator that takes as input an application written for a MIMD computer and run it on a workstations cluster with just a recompilation of the code. The hardware of the target machine is simulated so that the behavior of the application is identical to a native run on the simulated computer with virtual timings and trace file. Moreover, our analysis sets up the conditions required to achieve a good speedup as a function of the number of simulation hosts, the network latency and the granularity of the application.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
9

MORAJKO, ANNA, OLEG MORAJKO, JOSEP JORBA, TOMÀS MARGALEF und EMILIO LUQUE. „AUTOMATIC PERFORMANCE ANALYSIS AND DYNAMIC TUNING OF DISTRIBUTED APPLICATIONS“. Parallel Processing Letters 13, Nr. 02 (Juni 2003): 169–87. http://dx.doi.org/10.1142/s0129626403001227.

Der volle Inhalt der Quelle
Annotation:
The classical way of tuning parallel/distributed applications is based on the analysis of the monitoring information obtained from an execution of the application. However, this "measure and modify" approach is not feasible when the applications have a dynamic behavior. In this case, another approach is required to accomplish performance expectations. This paper presents a solution based on the dynamic tuning approach that addresses these issues. In this approach, an application is monitored, its performance bottlenecks are detected and the application is modified automatically during the execution, without stopping, recompiling or re-running it. The introduced modifications adapt the behavior of the application to dynamic variations.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
10

Sakariya, Harsh Bipinbhai, und Ganesh D. „Taxonomy of Load Balancing Strategies in Distributed Systems“. International Journal of Innovative Research in Computer and Communication Engineering 12, Nr. 03 (25.03.2024): 1796–802. http://dx.doi.org/10.15680/ijircce.2024.1203070.

Der volle Inhalt der Quelle
Annotation:
Large-scale parallel and distributed computing systems are becoming more popular as a result of falling hardware prices and improvements in computer networking technologies. Improved performance and resource sharing are potential benefits of distributed computing systems. We have provided a summary of distributed computing in this essay. The differences between parallel and distributed computing, terms related to distributed computing, task distribution in distributed computing, performance metrics in distributed computing systems, parallel distributed algorithm models, benefits of distributed computing, and distributed computing's application domain were all covered in this paper.
APA, Harvard, Vancouver, ISO und andere Zitierweisen

Dissertationen zum Thema "Distributed Parallel Application"

1

Moraes, Sergio A. S. „A distributed processing framework with application to graphics“. Thesis, University of Sussex, 1994. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.387338.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
2

Hamza, Golyeri. „Safran: A Distributed And Parallel Application Development Framework For Networks Of Heterogeneous Workstations“. Master's thesis, METU, 2005. http://etd.lib.metu.edu.tr/upload/12606023/index.pdf.

Der volle Inhalt der Quelle
Annotation:
With the rapid advances in high-speed network technologies and steady decrease in the cost of hardware involved, network of workstation (NOW) environments began to attract attention as competitors against special purpose, high performance parallel processing environments. NOWs attract attention as parallel and distributed computing environments because they provide high scalability in terms of computing capacity and they have much smaller cost/performance ratios with high availability. However, they are harder to program for parallel and distributed applications because of the issues involved due to their loosely coupled nature. Some of the issues to be considered are the heterogeneity in the software and hardware architectures, uncontrolled external loads, network overheads, frequently changing system characteristics like workload on processors and network links, and security of applications and hosts. The general objective of this work is to provide the design and implementation of a JavaTM-based, high performance and flexible platform i.e. a framework that will facilitate development of wide range of parallel and distributed applications on networks of heterogeneous workstations (NOW). Parallel and distributed application developers are provided an infrastructure (consisting of pieces of executable software developed in Java and a Java software library) that allows them to build and run their distributed applications on their heterogeneous NOW without worrying about the issues specific to the NOW environments. The results of the extensive set of experiments conducted have shown that Safran is quite scaleable and responds well to compute intensive parallel and distributed applications.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
3

Paula, Nilton Cézar de. „Um ambiente de monitoramento de recursos e escalonamento cooperativo de aplicações paralelas em grades computacionais“. Universidade de São Paulo, 2009. http://www.teses.usp.br/teses/disponiveis/3/3141/tde-31032009-102441/.

Der volle Inhalt der Quelle
Annotation:
Grade computacional é uma alternativa para melhorar o desempenho de aplicações paralelas, por permitir o uso simultâneo de vários recursos distribuídos. Entretanto, para que a utilização de uma grade seja adequada, é necessário que os recursos sejam utilizados de maneira a permitir a otimização de algum critério. Para isto, várias estratégias de escalonamento têm sido propostas, mas o grande desafio é extrair o potencial que os recursos oferecem para a execução de aplicações paralelas. Uma estratégia bastante usada em sistemas de escalonamento atuais é escalonar uma aplicação paralela nos recursos de um único cluster. Contudo, apesar da estratégia ser simples, ela é muito limitada, devido principalmente a baixa utilização dos recursos. Este trabalho propõe e implementa o sistema GCSE (Grid Cooperative Scheduling Environment) que provê uma estratégia de escalonamento cooperativo para usar eficientemente os recursos distribuídos. Os processos de uma aplicação paralela podem ser distribuídos em recursos de vários clusters e computadores, todos conectados a redes de comunicação públicas. GCSE também gerencia a execução das aplicações, bem como oferece um conjunto de primitivas que fornece informações sobre os ambientes de execução para o suporte à comunicação entre processos. Além disto, uma estratégia de antecipação de dados é proposta para aumentar ainda mais o desempenho das aplicações. Para realizar um bom escalonamento é preciso descobrir os recursos distribuídos. Neste sentido, o sistema LIMA (Light-weIght Monitoring Architecture) foi projetado e implementado. Este sistema provê um conjunto de estratégias e mecanismos para o armazenamento distribuído e acesso eficiente às informações sobre os recursos distribuídos. Além disto, LIMA adiciona facilidades de descobrimento e integração com o GCSE e outros sistemas. Por fim, serão apresentados os testes e avaliações dos resultados com o uso integrado dos sistemas GCSE e LIMA, compondo um ambiente robusto para a execução de aplicações paralelas.
Computing grid is an alternative for improving the parallel application performance, because it allows the simultaneous use of many distributed resources. However, in order to take advantage of a grid, the resources must be used in such a way that some criteria can be optimized. Thus, various scheduling strategies have been proposed, but the great challenge is the exploitation of the potential that the resources provide to the parallel application execution. A strategy often used in current scheduling systems is to schedule a parallel application on resources of a single cluster. Even though this strategy is simple, it is very limited, mainly due to low resource utilization. This thesis proposes and implements the GCSE system (Grid Cooperative Scheduling Environment) that provides a cooperative scheduling strategy for efficiently using the distributed resources. The processes of a parallel application can be distributed in resources of many clusters and computers, and they are all connected by public communication networks. GCSE also manages the application execution, as well as offering a primitive set that provide information about the execution environments for ensuring the communication between processes. Moreover, a data advancement strategy is proposed for improving the application performance. In order to perform a good scheduling, the distributed resources must be discovered. Therefore, the LIMA system (Light-weIght Monitoring Architecture) was designed and implemented. This system provides both strategy and mechanism set for distributed storage and efficient access to information about the distributed resources. In addition, LIMA offers facilities for resource discovering and integrating its functionalities both GCSE and other systems. Finally, the tests and result evaluations are presented with the integrated use of both GCSE and LIMA systems, composing a robust environment for executing parallel application.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
4

Wang, Chen. „Chemistry Inspired Middleware for Flexible Service Composition and Application“. Phd thesis, INSA de Rennes, 2013. http://tel.archives-ouvertes.fr/tel-00932085.

Der volle Inhalt der Quelle
Annotation:
Les Architectures Orientées Services (SOA) sont adoptées aujourd'hui par de nombreuses entreprises car elles représentent une solution flexible pour la construction d'applications distribuées. Une Application Basée sur des Services (SBA) peut se définir comme un workflow qui coordonne de manière dynamique l'exécution distribuée d'un ensemble de services. Les services peuvent être sélectionnés et intégrés en temps réel en fonction de leur Qualité de Service (QoS), et la composition de services peut être dynamiquement modifiée pour réagir à des défaillances imprévues pendant l'exécution. Les besoins des architectures orientées services présentent des similarités avec la nature: dynamicité, évolutivité, auto-adaptabilité, etc. Ainsi, il n'est pas surprenant que les métaphores inspirées par la nature soient considérées comme des approches appropriées pour la modélisation de tels systèmes. Nous allons plus loin en utilisant le paradigme de programmation chimique comme base de construction d'un middleware. Dans cette thèse, nous présentons un middleware "chimique'' pour l'exécution dynamique et adaptative de SBA. La sélection, l'intégration, la coordination et l'adaptation de services sont modélisées comme une série de réactions chimiques. Tout d'abord, l'instantiation de workflow est exprimée par une série de réactions qui peuvent être effectuées de manière parallèle, distribuée et autonome. Ensuite, nous avons mis en oeuvre trois modèles de coordination pour exécuter une composition de service. Nous montrons que les trois modèles peuvent réagir aux défaillances de type panne franche. Enfin, nous avons évalué et comparé ces modèles au niveau d'efficacité et complexité sur deux workflows. Nous montrons ainsi dans cette thèse que le paradigme chimique possède les qualités nécessaires à l'introduction de la dynamicité et de l'adaptabilité dans la programmation basée sur les services.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
5

Ghafoor, Sheikh Khaled. „Modeling of an adaptive parallel system with malleable applications in a distributed computing environment“. Diss., Mississippi State : Mississippi State University, 2007. http://sun.library.msstate.edu/ETD-db/theses/available/etd-11092007-145420.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
6

Helal, Manal Computer Science &amp Engineering Faculty of Engineering UNSW. „Indexing and partitioning schemes for distributed tensor computing with application to multiple sequence alignment“. Awarded by:University of New South Wales. Computer Science & Engineering, 2009. http://handle.unsw.edu.au/1959.4/44781.

Der volle Inhalt der Quelle
Annotation:
This thesis investigates indexing and partitioning schemes for high dimensional scientific computational problems. Building on the foundation offered by Mathematics of Arrays (MoA) for tensor-based computation, the ultimate contribution of the thesis is a unified partitioning scheme that works invariant of the dataset dimension and shape. Consequently, portability is ensured between different high performance machines, cluster architectures, and potentially computational grids. The Multiple Sequence Alignment (MSA) problem in computational biology has an optimal dynamic programming based solution, but it becomes computationally infeasible as its dimensionality (the number of sequences) increases. Even sub-optimal approximations may be unmanageable for more than eight sequences. Furthermore, no existing MSA algorithms have been formulated in a manner invariant over the number of sequences. This thesis presents an optimal distributed MSA method based on MoA. The latter offers a set of constructs that help represent multidimensional arrays in memory in a linear, concise and efficient way. Using MoA allows the partitioning of the dynamic programming algorithm to be expressed independently of dimension. MSA is the highest dimensional scientific problem considered for MoA-based partitioning to date. Two partitioning schemes are presented: the first is a master/slave approach which is based on both master/slave scheduling and slave/slave coupling. The second approach is a peer-to-peer design, in which the scheduling and dependency communication are calculated independently by each process, with no need for a master scheduler. A search space reduction technique is introduced to cater for the exponential expansion as the problem dimensionality increases. This technique relies on defining a hyper-diagonal through the tensor space, and choosing a band of neighbouring partitions around the diagonal to score. In contrast, other sub-optimal methods in the literature only consider projections on the surface of the hyper-cube. The resulting massively parallel design produces a scalable solution that has been implemented on high performance machines and cluster architectures. Experimental results for these implementations are presented for both simulated and real datasets. Comparisons between the reduced search space technique of this thesis with other sub-optimal methods for the MSA problem are presented.
APA, Harvard, Vancouver, ISO und andere Zitierweisen
7

King, Randall. „WARPED Redesigned: An API and Implementation for Discrete Event Simulation Analysis and Application Development“. University of Cincinnati / OhioLINK, 2011. http://rave.ohiolink.edu/etdc/view?acc_num=ucin1298040848.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
8

Singh, Neeta S. „An automatic code generation tool for partitioned software in distributed computing“. [Tampa, Fla.] : University of South Florida, 2005. http://purl.fcla.edu/fcla/etd/SFE0001129.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
9

Kachirski, Oleg. „AN INTERACTIVE DISTRIBUTED SIMULATION FRAMEWORK WITH APPLICATION TO WIRELESS NETWORKS AND INTRUSION DETECTION“. Doctoral diss., University of Central Florida, 2005. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/2531.

Der volle Inhalt der Quelle
Annotation:
In this dissertation, we describe the portable, open-source distributed simulation framework (WINDS) targeting simulations of wireless network infrastructures that we have developed. We present the simulation framework which uses modular architecture and apply the framework to studies of mobility pattern effects, routing and intrusion detection mechanisms in simulations of large-scale wireless ad hoc, infrastructure, and totally mobile networks. The distributed simulations within the framework execute seamlessly and transparently to the user on a symmetric multiprocessor cluster computer or a network of computers with no modifications to the code or user objects. A visual graphical interface precisely depicts simulation object states and interactions throughout the simulation execution, giving the user full control over the simulation in real time. The network configuration is detected by the framework, and communication latency is taken into consideration when dynamically adjusting the simulation clock, allowing the simulation to run on a heterogeneous computing system. The simulation framework is easily extensible to multi-cluster systems and computing grids. An entire simulation system can be constructed in a short time, utilizing user-created and supplied simulation components, including mobile nodes, base stations, routing algorithms, traffic patterns and other objects. These objects are automatically compiled and loaded by the simulation system, and are available for dynamic simulation injection at runtime. Using our distributed simulation framework, we have studied modern intrusion detection systems (IDS) and assessed applicability of existing intrusion detection techniques to wireless networks. We have developed a mobile agent-based IDS targeting mobile wireless networks, and introduced load-balancing optimizations aimed at limited-resource systems to improve intrusion detection performance. Packet-based monitoring agents of our IDS employ a CASE-based reasoner engine that performs fast lookups of network packets in the existing SNORT-based intrusion rule-set. Experiments were performed using the intrusion data from MIT Lincoln Laboratories studies, and executed on a cluster computer utilizing our distributed simulation system.
Ph.D.
School of Computer Science
Engineering and Computer Science
Computer Science
APA, Harvard, Vancouver, ISO und andere Zitierweisen
10

Scriven, Ian Michael. „Derivation and Application of Approximate Electromagnetic Noise Source Models using Decentralised Parallel Particle Swarm Optimisation“. Thesis, Griffith University, 2012. http://hdl.handle.net/10072/367576.

Der volle Inhalt der Quelle
Annotation:
Distributed computing and computational optimisation algorithms are often applied to design problems in various engineering disciplines. In electromagnetic compatibility, an area of electrical engineering concerned with the unintentional generation of and susceptibility to electromagnetic energy, distributed optimisation techniques could be used to ensure compliance with various civilian and military standards that most electrical devices must meet. While there are a variety of numerical techniques that can be applied to problems in computational electromagnetic compatibility, the finite-difference time-domain (FDTD) method stands out as it allows a wide range of frequencies to be examined in a single simulation using Fourier analysis. This property is important for electromagnetic compatibility problems, as resonant frequencies are usually not known or easily calculable. The efficacy of the application of the FDTD method to electromagnetic compatibility problems is limited, however, by a need for excitation sources which can accurately represent the device being modelled. Computational electromagnetic simulations have long been used in many engineering fields to numerically examine the performance of a wide variety of systems. Using such numerical methods, it is possible to examine the influence that any number of input parameters or stimuli have on the design. Optimisation algorithms are often used in conjunction with these numerical simulations to automate the design process.
Thesis (PhD Doctorate)
Doctor of Philosophy (PhD)
Griffith School of Engineering
Science, Environment, Engineering and Technology
Full Text
APA, Harvard, Vancouver, ISO und andere Zitierweisen

Bücher zum Thema "Distributed Parallel Application"

1

IFIP WG10.3 Working Conference on Applications in Parallel and Distributed Computing (1994 Caracas, Venezuela). Applications in parallel and distributed computing: Proceedings of the IFIP WG10.3 Working Conference on Applications in Parallel and Distributed Computing, Caracas, Venezuela, 18-22 April, 1994. Amsterdam: North-Holland, 1994.

Den vollen Inhalt der Quelle finden
APA, Harvard, Vancouver, ISO und andere Zitierweisen
2

Lawrence, Sterling Thomas, Hrsg. How to build a Beowulf: A guide to the implementation and application of PC clusters. Cambridge, Mass: MIT Press, 1999.

Den vollen Inhalt der Quelle finden
APA, Harvard, Vancouver, ISO und andere Zitierweisen
3

Stojmenovic, Ivan, Ruppa K. Thulasiram, Laurence T. Yang, Weijia Jia, Minyi Guo und Rodrigo Fernandes de Mello, Hrsg. Parallel and Distributed Processing and Applications. Berlin, Heidelberg: Springer Berlin Heidelberg, 2007. http://dx.doi.org/10.1007/978-3-540-74742-0.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
4

Guo, Minyi, und Laurence Tianruo Yang, Hrsg. Parallel and Distributed Processing and Applications. Berlin, Heidelberg: Springer Berlin Heidelberg, 2003. http://dx.doi.org/10.1007/3-540-37619-4.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
5

Guo, Minyi, Laurence T. Yang, Beniamino Di Martino, Hans P. Zima, Jack Dongarra und Feilong Tang, Hrsg. Parallel and Distributed Processing and Applications. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11946441.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
6

Pan, Yi, Daoxu Chen, Minyi Guo, Jiannong Cao und Jack Dongarra, Hrsg. Parallel and Distributed Processing and Applications. Berlin, Heidelberg: Springer Berlin Heidelberg, 2005. http://dx.doi.org/10.1007/11576235.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
7

Cao, Jiannong, Laurence T. Yang, Minyi Guo und Francis Lau, Hrsg. Parallel and Distributed Processing and Applications. Berlin, Heidelberg: Springer Berlin Heidelberg, 2005. http://dx.doi.org/10.1007/b104574.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
8

Institute Of Electrical and Electronics Engineers. IEEE parallel & distributed technology: Systems & applications. Los Alamitos, CA: IEEE Computer Society, 1993.

Den vollen Inhalt der Quelle finden
APA, Harvard, Vancouver, ISO und andere Zitierweisen
9

Shen, Hong, Yingpeng Sang, Yong Zhang, Nong Xiao, Hamid R. Arabnia, Geoffrey Fox, Ajay Gupta und Manu Malek, Hrsg. Parallel and Distributed Computing, Applications and Technologies. Cham: Springer International Publishing, 2022. http://dx.doi.org/10.1007/978-3-030-96772-7.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
10

Xie, Guoqi, Gang Zeng, Renfa Li und Keqin Li. Scheduling Parallel Applications on Heterogeneous Distributed Systems. Singapore: Springer Singapore, 2019. http://dx.doi.org/10.1007/978-981-13-6557-7.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen

Buchteile zum Thema "Distributed Parallel Application"

1

Lovas, Róbert, Péter Kacsuk, Ákos Horváth und Ándrás Horányi. „Application of P-Grade Development Environment in Meteorology“. In Distributed and Parallel Systems, 109–16. Boston, MA: Springer US, 2002. http://dx.doi.org/10.1007/978-1-4615-1167-0_13.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
2

Yu, Shui, und Wanlei Zhou. „An Efficient Reliable Architecture for Application Layer Anycast Service“. In Distributed and Parallel Computing, 376–85. Berlin, Heidelberg: Springer Berlin Heidelberg, 2005. http://dx.doi.org/10.1007/11564621_44.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
3

Marosi, Attila Csaba, Gábor Gombás und Zoltán Balaton. „Secure application deployment in the Hierarchical Local Desktop Grid“. In Distributed and Parallel Systems, 145–53. Boston, MA: Springer US, 2007. http://dx.doi.org/10.1007/978-0-387-69858-8_15.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
4

Kovács, József, Rafal Mikolajczak, Radoslaw Januszewski und Gracjan Jankowski. „Application and Middleware Transparent Checkpointing with TCKPT on Clustergrid“. In Distributed and Parallel Systems, 179–89. Boston, MA: Springer US, 2007. http://dx.doi.org/10.1007/978-0-387-69858-8_18.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
5

Walters, John Paul, und Vipin Chaudhary. „Application-Level Checkpointing Techniques for Parallel Programs“. In Distributed Computing and Internet Technology, 221–34. Berlin, Heidelberg: Springer Berlin Heidelberg, 2006. http://dx.doi.org/10.1007/11951957_21.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
6

Fürlinger, Karl, und Michael Gerndt. „Distributed Application Monitoring for Clustered SMP Architectures“. In Euro-Par 2003 Parallel Processing, 127–34. Berlin, Heidelberg: Springer Berlin Heidelberg, 2003. http://dx.doi.org/10.1007/978-3-540-45209-6_20.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
7

Zhang, Wenbo, Bo Yang, Beihong Jin, Ningjing Chen und Tao Huang. „Performance Tuning for Application Server OnceAS“. In Parallel and Distributed Processing and Applications, 451–62. Berlin, Heidelberg: Springer Berlin Heidelberg, 2004. http://dx.doi.org/10.1007/978-3-540-30566-8_54.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
8

Shetty, Keerthi S., und Sanjay Singh. „Cloud Based Application Development for Mobile Devices for Accessing LBS“. In Advances in Parallel Distributed Computing, 532–43. Berlin, Heidelberg: Springer Berlin Heidelberg, 2011. http://dx.doi.org/10.1007/978-3-642-24037-9_53.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
9

Leblanc, Philippe, und Jean-Luc Roux. „Extended SDL-Based Tools for Rapid Prototyping of Application Specific Signal Processors“. In Distributed and Parallel Embedded Systems, 149–56. Boston, MA: Springer US, 1999. http://dx.doi.org/10.1007/978-0-387-35570-2_13.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
10

Sakaguchi, Tomoya, Hitoshi Sakagami, Manabu Nii und Yutaka Takahashi. „Implementation of Application Collaboration Protocol“. In Parallel and Distributed Computing: Applications and Technologies, 90–93. Berlin, Heidelberg: Springer Berlin Heidelberg, 2004. http://dx.doi.org/10.1007/978-3-540-30501-9_22.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen

Konferenzberichte zum Thema "Distributed Parallel Application"

1

Li, Yusen, Wentong Cai und Xueyan Tang. „Application Layer Multicast in P2P Distributed Interactive Applications“. In 2013 International Conference on Parallel and Distributed Systems (ICPADS). IEEE, 2013. http://dx.doi.org/10.1109/icpads.2013.62.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
2

Ataie, Mohammad Zaman, und Omid Elahi. „Analysis of a Parallel/Distributed Application Using a Cycle-Accurate Parallel/Distributed Simulator“. In 2018 Iranian Conference on Electrical Engineering (ICEE). IEEE, 2018. http://dx.doi.org/10.1109/icee.2018.8472447.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
3

Watson, Gregory R., Craig E. Rasmussen und Beth R. Tibbitts. „An integrated approach to improving the parallel application development process“. In Distributed Processing (IPDPS). IEEE, 2009. http://dx.doi.org/10.1109/ipdps.2009.5160941.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
4

Fatoohi, Rod. „Performance evaluation of NSF application benchmarks on parallel systems“. In Distributed Processing Symposium (IPDPS). IEEE, 2008. http://dx.doi.org/10.1109/ipdps.2008.4536498.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
5

Derbel, B., M. Mosbah und A. Zemmari. „Fast distributed graph partition and application“. In Proceedings 20th IEEE International Parallel & Distributed Processing Symposium. IEEE, 2006. http://dx.doi.org/10.1109/ipdps.2006.1639362.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
6

Wang, Alf Inge, und Erik Andreas Larsen. „Using Brain-Computer Interfaces in an Interactive Multimedia Application“. In Parallel and Distributed Computing and Systems. Calgary,AB,Canada: ACTAPRESS, 2012. http://dx.doi.org/10.2316/p.2012.790-046.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
7

Filgueira, Rosa, David E. Singh, Florin Isaila, Jesu's Carretero und Antonio Garcia Loureiro. „Optimization and evaluation of parallel I/O in BIPS3D parallel irregular application“. In 2007 IEEE International Parallel and Distributed Processing Symposium. IEEE, 2007. http://dx.doi.org/10.1109/ipdps.2007.370585.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
8

Faber, Clayton J., und Roger D. Chamberlain. „Application of Network Calculus Models to Heterogeneous Streaming Applications“. In 2024 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW). IEEE, 2024. http://dx.doi.org/10.1109/ipdpsw63119.2024.00057.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
9

Martel, E., J. Miranda, L. Hernandez und F. Guerra. „Remote management of distributed application“. In 12th Euromicro Conference on Parallel, Distributed and Network-Based Processing, 2004. Proceedings. IEEE, 2004. http://dx.doi.org/10.1109/empdp.2004.1271441.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
10

Padmanabhan, S., R. K. Cytron, R. D. Chamberlain und J. W. Lockwood. „Automatic application-specific microarchitecture reconfiguration“. In Proceedings 20th IEEE International Parallel & Distributed Processing Symposium. IEEE, 2006. http://dx.doi.org/10.1109/ipdps.2006.1639457.

Der volle Inhalt der Quelle
APA, Harvard, Vancouver, ISO und andere Zitierweisen
Wir bieten Rabatte auf alle Premium-Pläne für Autoren, deren Werke in thematische Literatursammlungen aufgenommen wurden. Kontaktieren Sie uns, um einen einzigartigen Promo-Code zu erhalten!

Zur Bibliographie