Dissertations / Theses on the topic 'Software architecture – Evaluation'

To see the other types of publications on this topic, follow the link: Software architecture – Evaluation.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Software architecture – Evaluation.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Mårtensson, Frans, and Per Jönsson. "Software Architecture Simulation." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik och datavetenskap, 2002. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-4087.

Full text
Abstract:
A software architecture is one of the first steps towards a software system. A software architecture can be designed in different ways. During the design phase, it is important to select the most suitable design of the architecture, in order to create a good foundation for the system. The selection process is performed by evaluating architecture alternatives against each other. We investigate the use of continuous simulation of a software architecture as a support tool for architecture evaluation. For this purpose, we study a software architecture of an existing software system in an experiment, where we create a model of it using a tool for continuous simulation, and simulate the model. Based on the results from the simulation, we conclude that the system is too complex to be modeled for continuous simulation. Problems we identify are that we need discrete functionality to be able to correctly simulate the system, and that it is very time-consuming to develop a model for evaluation purposes. Thus, we find that continuous simulation is not appropriate for evaluating a software architecture, but that the modeling process is a valuable tool for increasing knowledge and understanding about an architecture.
APA, Harvard, Vancouver, ISO, and other styles
2

Shepperd, Martin John. "System architecture metrics : an evaluation." n.p, 1990. http://ethos.bl.uk/.

Full text
APA, Harvard, Vancouver, ISO, and other styles
3

Zhu, Liming Computer Science &amp Engineering Faculty of Engineering UNSW. "Software architecture evaluation for framework-based systems." Awarded by:University of New South Wales. Computer Science and Engineering, 2007. http://handle.unsw.edu.au/1959.4/28250.

Full text
Abstract:
Complex modern software is often built using existing application frameworks and middleware frameworks. These frameworks provide useful common services, while simultaneously imposing architectural rules and constraints. Existing software architecture evaluation methods do not explicitly consider the implications of these frameworks for software architecture. This research extends scenario-based architecture evaluation methods by incorporating framework-related information into different evaluation activities. I propose four techniques which target four different activities within a scenario-based architecture evaluation method. 1) Scenario development: A new technique was designed aiming to extract general scenarios and tactics from framework-related architectural patterns. The technique is intended to complement the current scenario development process. The feasibility of the technique was validated through a case study. Significant improvements of scenario quality were observed in a controlled experiment conducted by another colleague. 2) Architecture representation: A new metrics-driven technique was created to reconstruct software architecture in a just-in-time fashion. This technique was validated in a case study. This approach has significantly improved the efficiency of architecture representation in a complex environment. 3) Attribute specific analysis (performance only): A model-driven approach to performance measurement was applied by decoupling framework-specific information from performance testing requirements. This technique was validated on two platforms (J2EE and Web Services) through a number of case studies. This technique leads to the benchmark producing more representative measures of the eventual application. It reduces the complexity behind the load testing suite and framework-specific performance data collecting utilities. 4) Trade-off and sensitivity analysis: A new technique was designed seeking to improve the Analytical Hierarchical Process (AHP) for trade-off and sensitivity analysis during a framework selection process. This approach was validated in a case study using data from a commercial project. The approach can identify 1) trade-offs implied by an architecture alternative, along with the magnitude of these trade-offs. 2) the most critical decisions in the overall decision process 3) the sensitivity of the final decision and its capability for handling quality attribute priority changes.
APA, Harvard, Vancouver, ISO, and other styles
4

Shepperd, Martin John. "System architecture metrics : an evaluation." Thesis, Open University, 1991. http://oro.open.ac.uk/57340/.

Full text
Abstract:
The research described in this dissertation is a study of the application of measurement, or metrics for software engineering. This is not in itself a new idea; the concept of measuring software was first mooted close on twenty years ago. However, examination of what is a considerable body of metrics work, reveals that incorporating measurement into software engineering is rather less straightforward than one might pre-suppose and despite the advancing years, there is still a lack of maturity. The thesis commences with a dissection of three of the most popular metrics, namely Haistead's software science, McCabe's cyclomatic complexity and Henry and Kafura's information flow - all of which might be regarded as having achieved classic status. Despite their popularity these metrics are all flawed in at least three respects. First and foremost, in each case it is unclear exactly what is being measured: instead there being a preponderance of such metaphysical terms as complexIty and qualIty. Second, each metric is theoretically doubtful in that it exhibits anomalous behaviour. Third, much of the claimed empirical support for each metric is spurious arising from poor experimental design, and inappropriate statistical analysis. It is argued that these problems are not misfortune but the inevitable consequence of the ad hoc and unstructured approach of much metrics research: in particular the scant regard paid to the role of underlying models. This research seeks to address these problems by proposing a systematic method for the development and evaluation of software metrics. The method is a goal directed, combination of formal modelling techniques, and empirical ealiat%or. The met\io s applied to the problem of developing metrics to evaluate software designs - from the perspective of a software engineer wishing to minimise implementation difficulties, faults and future maintenance problems. It highlights a number of weaknesses within the original model. These are tackled in a second, more sophisticated model which is multidimensional, that is it combines, in this case, two metrics. Both the theoretical and empirical analysis show this model to have utility in its ability to identify hardto- implement and unreliable aspects of software designs. It is concluded that this method goes some way towards the problem of introducing a little more rigour into the development, evaluation and evolution of metrics for the software engineer.
APA, Harvard, Vancouver, ISO, and other styles
5

Borowski, Jimmy. "Software Architecture Simulation : Performance evaluation during the design phase." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik och datavetenskap, 2004. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-5882.

Full text
Abstract:
Due to the increasing size and complexity of software systems, software architectures have become a crucial part in development projects. A lot of effort has been put into defining formal ways for describing architecture specifications using Architecture Description Languages (ADLs). Since no common ADL today offers tools for evaluating the performance, an attempt to develop such a tool based on an event-based simulation engine has been made. Common ADLs were investigated and the work was based on the fundamentals within the field of software architectures. The tool was evaluated both in terms of correctness in predictions as well as usability to show that it actually is possible to evaluate the performance using high-level architectures as models.
APA, Harvard, Vancouver, ISO, and other styles
6

Mårtensson, Frans. "Software architecture quality evaluation : approaches in an industrial context /." Karlskrona : Blekinge Institute of Technology, 2006. http://www.bth.se/fou/Forskinfo.nsf/allfirst2/3e821fbd7a66542cc1257169002ad63c?OpenDocument.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Mårtensson, Frans. "Software Architecture Quality Evaluation : Approaches in an Industrial Context." Licentiate thesis, Karlskrona : Blekinge Institute of Technology, 2006. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-00313.

Full text
Abstract:
Software architecture has been identified as an increasingly important part of software development. The software architecture helps the developer of a software system to define the internal structure of the system. Several methods for evaluating software architectures have been proposed in order to assist the developer in creating a software architecture that will have a potential to fulfil the requirements on the system. Many of the evaluation methods focus on evaluation of a single quality attribute. However, in an industrial system there are normally requirements on several quality aspects of the system. Therefore, an architecture evaluation method that addresses multiple quality attributes, e.g., performance, maintainability, testability, and portability, would be more beneficial. This thesis presents research towards a method for evaluation of multiple quality attributes using one software architecture evaluation method. A prototype-based evaluation method is proposed that enables evaluation of multiple quality attributes using components of a system and an approximation of its intended runtime environment. The method is applied in an industrial case study where communication components in a distributed realtime system are evaluated. The evaluation addresses performance, maintainability, and portability for three alternative components using a single set of software architecture models and a prototype framework. The prototype framework enables the evaluation of different components and component configurations in the software architecture while collecting data in an objective way. Finally, this thesis presents initial work towards incorporating evaluation of testability into the method. This is done through an investigation of how testability is interpreted by different organizational roles in a software developing organization and which measures of source code that they consider affecting testability.
APA, Harvard, Vancouver, ISO, and other styles
8

Graser, Thomas Jeffrey. "Reference architecture representation environment (RARE) : systematic derivation and evaluation of domain-specific, implementation-independent software architectures /." Access restricted to users with UT Austin EID Full text (PDF) from UMI/Dissertation Abstracts International, 2001. http://wwwlib.umi.com/cr/utexas/fullcit?p3023549.

Full text
APA, Harvard, Vancouver, ISO, and other styles
9

Zapalowski, Vanius. "Evaluation of code-based information to architectural module identification." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2014. http://hdl.handle.net/10183/94691.

Full text
Abstract:
Arquitetura de software desempenha um importante papel no desenvolvimento de software, quando explicitamente documentada, ela melhora o entendimento sobre o sistema implementado e torna possível entender a forma com que requisitos não funcionais são tratados. Apesar da relevância da arquitetura de software, muitos sistemas não possuem uma arquitetura documentada, e nos casos em que a arquitetura existe, ela pode estar desatualizada por causa da evolução descontrolada do software. O processo de recuperação de arquitetura de um sistema depende principalmente do conhecimento que as pessoas envolvidas com o software tem. Isso acontece porque a recuperação de arquitetura é uma tarefa que demanda muita investigação manual do código fonte. Pesquisas sobre recuperação de arquitetura objetivam auxiliar esse processo. A maioria dos métodos de recuperação existentes são baseados em dependência entre elementos da arquitetura, padrões arquiteturais ou similaridade semântica do código fonte. Embora as abordagem atuais ajudem na identi cação de módulos arquiteturais, os resultados devem ser melhorados de forma signi cativa para serem considerados con áveis. Então, nesta dissertação, objetivamos melhorar o suporte a recuperação de arquitetura explorando diferentes fontes de informação e técnicas de aprendizado de máquina. Nosso trabalho consiste de uma análise, considerando cinco estudo de casos, da utilidade de usar um conjunto de características de código (features, no contexto de aprendizado de máquina) para agrupar elementos em módulos da arquitetura. Atualmente não são conhecidas as características que afetam a identificação de papéis na arquitetura de software. Por isso, nós avaliamos a relação entre diferentes conjuntos de características e a acurácia obtida por um algoritmo não supervisionado na identificação de módulos da arquitetura. Consequentemente, nós entendemos quais dessas características revelam informação sobre a organização de papéis do código fonte. Nossa abordagem usando características de elementos de software atingiu uma acurácia média significativa. Indicando a relevância das informações selecionadas para recuperar a arquitetura. Além disso, nós desenvolvemos uma ferramenta para auxílio ao processo de recuperação de arquitetura de software. Nossa ferramenta tem como principais funções a avaliação da recuperação de arquitetura e apresentação de diferentes visualizações arquiteturais. Para isso, apresentamos comparações entre a arquitetura concreta e a arquitetura sugerida.
Software architecture plays an important role in the software development, and when explicitly documented, it allows understanding an implemented system and reasoning about how non-functional requirements are addressed. In spite of that, many developed systems lack proper architecture documentation, and if it exists, it may be outdated due to software evolution. The process of recovering the architecture of a system depends mainly on developers' knowledge requiring a manual inspection of the source code. Research on architecture recovery provides support to this process. Most of the existing approaches are based on architectural elements dependency, architectural patterns or source code semantics, but even though they help identifying architectural modules, the obtained results must be signi cantly improved to be considered reliable. We thus aim to support this task by the exploitation of di erent code-oriented information and machine learning techniques. Our work consists of an analysis, involving ve case studies, of the usefulness of adopting a set of code-level characteristics (or features, in the machine learning terminology) to group elements into architectural modules. The characteristics mainly source code metrics that a ect the identi cation of what role software elements play in software architecture are unknown. Then, we evaluate the relationship between di erent sets of characteristics and the accuracy achieved by an unsupervised algorithm the Expectation Maximization in identifying architectural modules. Consequently, we are able to understand which of those characteristics reveal information about the source code structure. By the use of code-oriented information, our approach achieves a signi cant average accuracy, which indicates the importance of the selected information to recover software architecture. Additionally, we provide a tool to support research on architecture recovery providing software architecture measurements and visualizations. It presents comparisons between predicted architectures and concrete architectures.
APA, Harvard, Vancouver, ISO, and other styles
10

Hawley, Jeffrey Allan. "Software architecture of the non-rigid image registration evaluation project." Thesis, University of Iowa, 2011. https://ir.uiowa.edu/etd/1229.

Full text
Abstract:
In medical image registration the goal is to find point by point correspondences between a source image and a target image such that the two images are aligned. There are rigid and non-rigid registration algorithms. Rigid registration uses rigid transformation methods which preserve distances between every pair of points. Non-rigid registration uses transformation methods that do not have to preserve the distances. Image registration has many medical applications -tracking tumors, anatomical changes over time, differences between characteristics like age and gender, etc. A gold standard transformation to compare and evaluate the registration algorithms would be ideal to use to verify if the two images are perfectly aligned. However, there is hardly if ever a gold standard transformation for non-rigid registration algorithms. The reason why there is no gold standard transformation for non-rigid registration algorithms is that pointwise correspondence between two registered points is not unique. In the absence of a gold standard various evaluation methods are used to gauge registration performance. However, each evaluation method only evalutes the error in the transformation from a limited perspective and therefore has its advantages and drawbacks. The Non-Rigid Image Registration Evaluation Project (NIREP) was was created to provide one central tool that has a collection of evaluation methods to perform the evaluations on non-rigid image registration algorithms and rank the registration algorithms based on the outputs of the evaluation methods in the absence of without having to use a gold standard.
APA, Harvard, Vancouver, ISO, and other styles
11

Krishnamurthy, Likhita. "Comparative Assessment of Network-Centric Software Architectures." Thesis, Virginia Tech, 2006. http://hdl.handle.net/10919/32377.

Full text
Abstract:
The purpose of this thesis is to characterize, compare and contrast four network-centric software architectures, namely Client-Server Architecture (CSA), Distributed Objects Architecture (DOA), Service-Oriented Architecture (SOA) and Peer-to-Peer Architecture (PPA) and seven associated frameworks consisting of .NET, Java EE, CORBA, DCOM, Web Services, Jini and JXTA with respect to a set of derived criteria. Network-centric systems are gaining in popularity as they have the potential to solve more complex problems than we have been able to in the past. However, with the rise of SOA, Web Services, a set of standards widely used for implementing service-oriented solutions, is being touted as the â silver bulletâ to all problems afflicting the software engineering domain with the danger of making other architectures seem obsolete. Thus, there is an urgent need to study the various architectures and frameworks in comparison to each other and understand their relative merits and demerits for building network-centric systems. The architectures studied here were selected on the basis of their fundamentality and generality. The frameworks were chosen on the basis of their popularity and representativeness to build solutions in a particular architecture. The criteria used for comparative assessment are derived from a combination of two approaches â by a close examination of the unique characteristics and requirements of network-centric systems and then by an examination of the constraints and mechanisms present in the architectures and frameworks under consideration that may contribute towards realizing the requirements of network-centric systems. Not all of the criteria are equally relevant for the architectures and frameworks. Some, when relevant, are relevant in a different sense from one architecture (or framework) to another. One of the conclusions that can be drawn from this study is that the different architectures are not completely different from each other. In fact, CSA, DOA and SOA are a natural evolution in that order and share several characteristics. At the same time, significant differences do exist, so it is clearly possible to judge/differentiate one from the other. All three architectures can coexist in a single system or system of systems. However, the advantages of each architecture become apparent only when they are used in their proper scope. At the same time, a sharp difference can be perceived between these three architectures and the peer-to-peer architecture. This is because PPA aims to solve a totally different class of problems than the other three architectures and hence has certain unique characteristics not observed in the others. Further, all of the frameworks have certain unique architectural features and mechanisms not found in the others that contribute towards achieving network-centric quality characteristics. The two broad frameworks, .NET and Java EE offer almost equivalent capabilities and features; what can be achieved in one can be achieved in the other. This thesis deals with the study of all the four architectures and their related frameworks. The criteria used, while fairly comprehensive, are not exhaustive. Variants of the fundamental architectures are not considered. However, system/software architects seeking an understanding of the tradeoffs involved in using the various architectures and frameworks and their subtle nuances should benefit considerably from this work.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
12

AL, Halbouni Hadi, and Frank Hansen. "A Scenario-Based evaluation of Game Architecture." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-20241.

Full text
Abstract:
When developers or organizations need to develop a game, simulation or a similar project, they phase the question of whether or not to use a game engine as well as the question on which one to use. Are all game engines the same or does the architecture change and how is the game design different between various game engines? The objective of this thesis is to research these questions as well as giving a concrete understanding of the impact of picking one engine over the other and how each engine influences the way games are developed and answer some more specific questions regarding architecture and usability.  A project was designed with the goal of developing a game. This game was developed by two separate teams over a period of 6 weeks, using two different game engines. The development was split into separate iterations done simultaneously between the teams and questionnaires were filled in to gather data. The game engines used for projects had similarities but also things which were different. Each engine offered ways to speed up development by allowing the developer to reuse and distribute changes among objects to reduce work. The differences caused one engine’s code architecture to be more complex than the other while allowing a better code structure as well as adding more time to learn how the engine handles certain things such as collisions. In conclusion, there is an importance to properly evaluating different game engines depending on the project a developer or organization is creating, not evaluating this properly will impact development speed and project complexity. Even though each engine has their differences, there is no superior game engine as it all depends on the project being developed. The game developed for this project was only touching on certain areas related to 2D games.
APA, Harvard, Vancouver, ISO, and other styles
13

Yan, Jie. "An Evaluation of Current Applications of 3D Visualization Software in Landscape Architecture." DigitalCommons@USU, 2014. https://digitalcommons.usu.edu/etd/3882.

Full text
Abstract:
The design process is important to all landscape architects. It helps generate ideas to solve problems in an efficient amount of time and insure that all stages of a project are completed. Generally, a design process includes project acquisition, inventory and site analysis, conflicts identification, public involvement, draft products, and final presentation. Among these elements, public involvement has been recognized as one of the most important elements in the landscape design process. It not only helps professionals get projects done smoothly, but it also helps with long-term client retention. Traditional two-dimensional communication methods using renderings, design plans, and maps have not been fully successful in their ability to engage and sufficiently inform clients and stakeholders. While professional planners are able to rely on their experience to help them visualize proposed landscapes, the average client is often overwhelmed by the relatively complex and abstract information, and unable to translate this information into landscape visions. Developments in the field of 3D graphics have dramatically extended possibilities to overcome this barrier by providing a tool that produces designs that are easy to comprehend and helps clients better visualize the end product that the designer has put forth. Some people argue that the high investment cost of 3D software such as ArcGIS, 3Dmax, etc., and its time-consuming process to master, is too great an obstacle for most designers and firms to use the software in their work with the average client or stakeholder. However, little research has been done to investigate the extent to which landscape architects have adopted the 3D software. We know even less about their opinions on the suitability of existing 3D software packages to meet their professional needs. A nationwide survey about current use and future demand for 3D simulation software within the landscape profession was conducted for this thesis. Comprehensive online surveys were sent to two groups: (1) landscape architecture firms and freelance landscape architects; and (2) institutions with landscape architecture programs. In total, 3,434 firms and freelance architects were identified based on data from the American Society of Landscape Architecture. Names and contact information of ninety-one faculty members from institutions with landscape architecture programs were found on the Council of Educators in Landscape Architecture website. The opportunities and challenges of 3D visualization technology and its potential applications in landscape and environmental planning have been examined based on the findings from survey results. The results are relevant to the future improvement and innovation of 3D visualization software in the landscape architecture profession and can assist landscape architecture educators with future curriculum development.
APA, Harvard, Vancouver, ISO, and other styles
14

Powale, Kalkin. "Automotive Powertrain Software Evaluation Tool." Master's thesis, Universitätsbibliothek Chemnitz, 2018. http://nbn-resolving.de/urn:nbn:de:bsz:ch1-qucosa-233186.

Full text
Abstract:
The software is a key differentiator and driver of innovation in the automotive industry. The major challenges for software development are increasing in complexity, shorter time-to-market, increase in development cost and demand of quality assurance. The complexity is increasing due to emission legislations, variants of product and new communication technologies being interfaced with the vehicle. The shorter development time is due to competition in the market, which requires faster feedback loops of verification and validation of developed functionalities. The increase in development cost is contributed by two factors; the first is pre-launch cost, this involves the cost of error correction in development stages. Another is post-launch cost; this involves warranty and guarantees cost. As the development time passes the cost of error correction also increases. Hence it is important to detect the error as early as possible. All these factors affect the software quality; there are several cases where Original Equipment Manufacturer (OEM) have callbacks their product because of the quality defect. Hence, there is increased in the requirement of software quality assurance. The solution for these software challenges can be the early quality evaluation in continuous integration framework environment. The most prominent in today\'s automotive industry AUTomotive Open System ARchitecture (AUTOSAR) reference architecture is used to describe software component and interfaces. AUTOSAR provides the standardised software component architecture elements. It was created to address the issues of growing complexity; the existing AUTOSAR environment does have software quality measures, such as schema validations and protocols for acceptance tests. However, it lacks the quality specification for non-functional qualities such as maintainability, modularity, etc. The tool is required which will evaluate the AUTOSAR based software architecture and give the objective feedback regarding quality. This thesis aims to provide the quality measurement tool, which will be used for evaluation of AUTOSAR based software architecture. The tool reads the AUTOSAR architecture information from AUTOSAR Extensible Markup Language (ARXML) file. The tool provides configuration ability, continuous evaluation and objective feedback regarding software quality characteristics. The tool was utilised on transmission control project, and results are validated by industry experts.
APA, Harvard, Vancouver, ISO, and other styles
15

Guerra, Adrien, and Sebastien Peirone. "Evaluation and Selection of Software Architectures : A Case Study of Positioning Systems." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik och datavetenskap, 2002. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-3634.

Full text
Abstract:
The explosion of the mobile telecommunications market has resulted in the apparition of multiple applications and services. However the particular business pressure has often lead to that these systems have poor quality design and, consequently, implementation. Software system builders have realized the importance of the overall system organization to address quality requirements. Practically, they started to use, idiomatically, a number of commonly recognized solutions to guide their design of system structures. This report analyses and evaluates these solutions and comments on the result obtained when one of them was carefully selected and applied to a system of mobile positioning services.
+46708576448
APA, Harvard, Vancouver, ISO, and other styles
16

Wnukiewicz, Karol Kazimierz. "The role of quality requirements in software architecture design." Thesis, Blekinge Tekniska Högskola, Avdelningen för programvarusystem, 2006. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-2253.

Full text
Abstract:
An important issue during architectural design is that besides functional requirements, software architecture is influenced greatly by quality requirements [9][2][7], which often are neglected. The earlier quality requirements are considered, the less effort is needed later in the software lifecycle to ensure a sufficient software quality levels. Errors due to lack of their fulfilment are the most expensive and difficult to correct. Therefore, attention to quality requirements is crucial during an architectural design. The problem is not only to gather the system’s quality requirements, but to establish a methodology that helps to deal with them during the software development. Literature has paid some attention to software architecture in the context of quality requirements, but there is still lack of effective solutions in this area. To alleviate the problem, this paper lays out important concepts and notions of quality requirements in a way they can be used to drive design decisions and evaluate the architecture to estimate whether these requirements are fulfilled. Important concepts of software architecture area are presented to indicate how important quality requirements are during the design and what are the consequences of their lack in a software system. Moreover, a quality requirement-oriented design method is proposed as an outcome of the literature survey. This method is a model taking quality requirements into account at first, before the core functionality is placed. Besides the conceptual solution to the identified problems, this paper also suggests a practical method of handling quality requirements during a design. A recommendation framework for choosing the most suitable architectural pattern from a set of quality attributes is also proposed. Since the literature provides insufficient qualitative information about quality requirement issues in terms of software architectures, an empirical research is conducted as means for gathering the required data. Therefore, a systematic approach to support and analyze architectural designs in terms of quality requirements is prepared. Finally, quality requirement-oriented and pattern-based design method is further proposed as a result of investigating patterns as a tool for addressing quality requirements at different abstraction levels of a design. The research is concerned with the analysis of software architectures against one or more desired software qualities that ought to be achieved at the architectural level.
http://wnukiewicz.pl kareel@gmail.com
APA, Harvard, Vancouver, ISO, and other styles
17

Holmström, Jesper. "Distributed Queries: An Evaluation of the Microservice Architecture." Thesis, Linköpings universitet, Programvara och system, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-166314.

Full text
Abstract:
The microservice architecture is a new architectural style that structures an application into a set of small, independently deployable microservices, as opposed to the traditional monolith approach with a single executable. The microservice architecture is a distributed system that results in new challenges and increased complexity. This study expands the previous related research and investigates the implications of using the one-database-per-service pattern and a solution to the introduced need of queries spanning multiple microservices. In this thesis, two applications are presented, one with the microservice architecture and one monolithic counterpart, which are compared in terms of response time and throughput. As a solution for the distributed queries, the API Composition pattern was chosen. The results of the experiments conclude in a greater understanding of the difficulty in distributed queries as well as the benefits and limitations of the API composition pattern. It shows that the API composition pattern is a valid solution for distributed queries. However, it does perform worse in terms of response time and throughput than the monolith prototype. This results in the insights that one must carefully choose, with respect to the requirements of the system, when to apply it.
APA, Harvard, Vancouver, ISO, and other styles
18

Dandekar, Ashok V. "A procedural approach to the evaluation of software development methodologies." Thesis, Virginia Polytechnic Institute and State University, 1987. http://hdl.handle.net/10919/94502.

Full text
Abstract:
This thesis presents a procedural approach to evaluating software development methodologies. The evaluation procedure adopts a unique approach based on the fundamental assumption that the requirements govern the process by which software is constructed. To begin with, this research partitions desirable software characteristics into three categories, viz., objectives, principles and attributes. The thesis claims (claims are substantiated with literature references) that there exist definitive relationships (or linkages) among the software objectives, principles and attributes. These linkages form the foundation of the evaluation procedure. The procedure constitutes two processes, top-down process and bottom-up process. These processes are used to assess the software product and the employed software development methodology. The top-down process begins by identifying the objectives and travels down through principles and product attributes; thus, evaluating the adequacy of the methodology. The bottom-up process, on the other hand, starts at the attribute level and goes up through principles and objectives. The bottom-up process highlights the effectiveness of the methodology. Attributes are identified in the product via properties. This research establishes several properties (called factors in the report) for each attribute. A measurement approach is also presented to help assess the extent to which attributes are present. The feasibility and validity of the evaluation procedure are illustrated through the analysis of two real life methodologies.
M.S.
APA, Harvard, Vancouver, ISO, and other styles
19

Kvamme, Trond Blomholm, and Jan-Erik Strøm. "Evaluation and Extension of an XNA Game Library used in Software Architecture Projects." Thesis, Norwegian University of Science and Technology, Department of Computer and Information Science, 2008. http://urn.kb.se/resolve?urn=urn:nbn:no:ntnu:diva-8889.

Full text
Abstract:

For most young people growing up today, video games have been a part of their life on the same level as music, films, and other entertainment. They regard video games as a fun, exciting, and absorbing source of entertainment and stimulation. Transferring these properties into an educational context can prove to be very valuable and motivational. In this master thesis, the introduction of video game development with the XNA game development platform in software architecture projects at the Norwegian University of Science and Technology (NTNU) is evaluated. This includes an evaluation of a 2D XNA game library used in the projects. In addition, we present an assessment of the effort and time spent required to grasp the necessary 3D concepts and techniques involved in producing 3D games with XNA. We also describe our improvements and extensions of the game library to support and include 3D features, based on the evaluation and assessment. The students of the course had the choice between the traditional project (a robot simulation) and the new XNA project. We find that the students who chose the XNA project were more motivated, struggled less, and thus required less assistance. On the other hand, the XNA students admitted to over focusing on the gameplay of their game, at the expense of the software architecture. This should even out when more learning material specific to the XNA project becomes available. 40% of the XNA students used the game library in their project. Overall, they were satisfied with the usefulness and usability of it, but did not think it helped them focus less on technical matters and more on the architecture.

APA, Harvard, Vancouver, ISO, and other styles
20

Staite, Christopher. "Identity management architecture and implementation : evaluation and improvement." Thesis, University of Birmingham, 2012. http://etheses.bham.ac.uk//id/eprint/3388/.

Full text
Abstract:
The definition of identity varies, and on the Internet it can be difficult to keep track of. Rather than trying to discuss the philosophical question of "who am I?", I like to define your digital identity as the information you place on the Internet (actively or passively). Managing this identity comes down to what information you give out and how to protect and modify that information. This thesis focuses on the latter half, the protection and modification of online identities and only skims the realms of protecting the information given to third parties. A distinct lack of drive in the development of technologies for managing authentication has dogged the Internet for some time. Numerous efforts have been made to simplify administration, but open protocols meant for simplifying the user experience have had little promotion and ended up forgotten or used to simplify administration. The question that needs to be answered, as usual in research, is why? Studies have shown that password fatigue is a very real issue and identity theft is increasing. Companies will always optimise their time and resources, but academics need to focus their work on optimising the user experience. In this thesis, a study of existing work produces a methodology to evaluate previous developments. This aids in determining where progress has been made in previous iterations and how, leading to a new development in identity management focussed on the needs of the end user. Finally, two implementations are created to realise this new form of identity management.
APA, Harvard, Vancouver, ISO, and other styles
21

Waters, Robert Lee. "Obtaining Architectural Descriptions from Legacy Systems: The Architectural Synthesis Process (ASP)." Diss., Available online, Georgia Institute of Technology, 2004:, 2004. http://etd.gatech.edu/theses/available/etd-10272004-160115/unrestricted/waters%5Frobert%5Fl%5F200412%5Fphd.pdf.

Full text
Abstract:
Thesis (Ph. D.)--Computing, Georgia Institute of Technology, 2005.
Rick Kazman, Committee Member ; Colin Potts, Committee Member ; Mike McCracken, Committee Member ; Gregory Abowd, Committee Chair ; Spencer Rugaber, Committee Member. Includes bibliographical references.
APA, Harvard, Vancouver, ISO, and other styles
22

Panunzio, Marco <1982&gt. "Definition, realization and evaluation of a software reference architecture for use in space applications." Doctoral thesis, Alma Mater Studiorum - Università di Bologna, 2011. http://amsdottorato.unibo.it/3847/.

Full text
Abstract:
A recent initiative of the European Space Agency (ESA) aims at the definition and adoption of a software reference architecture for use in on-board software of future space missions. Our PhD project placed in the context of that effort. At the outset of our work we gathered all the industrial needs relevant to ESA and all the main European space stakeholders and we were able to consolidate a set of technical high-level requirements for the fulfillment of them. The conclusion we reached from that phase confirmed that the adoption of a software reference architecture was indeed the best solution for the fulfillment of the high-level requirements. The software reference architecture we set on building rests on four constituents: (i) a component model, to design the software as a composition of individually verifiable and reusable software units; (ii) a computational model, to ensure that the architectural description of the software is statically analyzable; (iii) a programming model, to ensure that the implementation of the design entities conforms with the semantics, the assumptions and the constraints of the computational model; (iv) a conforming execution platform, to actively preserve at run time the properties asserted by static analysis. The nature, feasibility and fitness of constituents (ii), (iii) and (iv), were already proved by the author in an international project that preceded the commencement of the PhD work. The core of the PhD project was therefore centered on the design and prototype implementation of constituent (i), a component model. Our proposed component model is centered on: (i) rigorous separation of concerns, achieved with the support for design views and by careful allocation of concerns to the dedicated software entities; (ii) the support for specification and model-based analysis of extra-functional properties; (iii) the inclusion space-specific concerns.
APA, Harvard, Vancouver, ISO, and other styles
23

Tyan, Hung-ying. "Design, realization and evaluation of a component-based compositional software architecture for network simulation /." The Ohio State University, 2002. http://rave.ohiolink.edu/etdc/view?acc_num=osu1486402544591791.

Full text
APA, Harvard, Vancouver, ISO, and other styles
24

SANTANA, Alixandre Thiago Ferreira de. "Enterprise architecture analysis based on network paradigm: a framework proposal and empirical evaluation." Universidade Federal de Pernambuco, 2017. https://repositorio.ufpe.br/handle/123456789/25224.

Full text
Abstract:
Submitted by Pedro Barros (pedro.silvabarros@ufpe.br) on 2018-07-25T20:09:25Z No. of bitstreams: 2 license_rdf: 811 bytes, checksum: e39d27027a6cc9cb039ad269a5db8e34 (MD5) TESE Alixandre Thiago Ferreira de Santana.pdf: 11775779 bytes, checksum: c96584c019fac4d26f44797654fcb6e3 (MD5)
Approved for entry into archive by Alice Araujo (alice.caraujo@ufpe.br) on 2018-07-26T21:34:30Z (GMT) No. of bitstreams: 2 license_rdf: 811 bytes, checksum: e39d27027a6cc9cb039ad269a5db8e34 (MD5) TESE Alixandre Thiago Ferreira de Santana.pdf: 11775779 bytes, checksum: c96584c019fac4d26f44797654fcb6e3 (MD5)
Made available in DSpace on 2018-07-26T21:34:30Z (GMT). No. of bitstreams: 2 license_rdf: 811 bytes, checksum: e39d27027a6cc9cb039ad269a5db8e34 (MD5) TESE Alixandre Thiago Ferreira de Santana.pdf: 11775779 bytes, checksum: c96584c019fac4d26f44797654fcb6e3 (MD5) Previous issue date: 2017-02-15
CAPES
While the literature on Enterprise Architecture (EA) models, frameworks, and methodologies for EA implementation has many exemplars, the field is still missing mechanisms of EA analysis. EA analysis is the process which uses any technique or method to extract information from EA models about a particular concern, in order to support EA management by the experts or inform stakeholders. In this thesis, we model the EA as a complex network, a concept discussed in network science, to analyze EA structural aspects. During our exploratory study about EA network analysis (EANA), it was clear that the field was still lacking foundational aspects. First, no common language was shared by researchers. Secondly, there was no clarity about what concerns could be analyzed with network analysis initiatives and thirdly, the techniques and methods´ implementation were not clear in the papers. We solve those gaps in order to describe how to perform analysis of EA components and their relationships supported by network measures. The research approach comprehends qualitative methods such as systematic literature review, thematic analysis and design science research method. The research is conducted in three complementary and interrelated phases, aiming at first, to collect and synthesize the available knowledge about the analysis approaches existent in the literature. Next, we aim to trace a comprehensive understanding of the main concepts involved in EANA such as their analysis concerns, modeling decisions, inputs required and steps necessary to perform it. Altogether, this resulted in a set of six proposed artifacts: EANA meta-model, EANA library, EANA process, EANA data derivation strategy. Finally, we investigate the use of those artifacts, evaluating them empirically through their instantiations and/ or with the help of EA experts of three German multinational companies. The evaluation results were positive regarding, among other criteria, the efficacy and utility of the proposed artifacts in their respective contexts. As contributions, we claim the definition of the conceptual foundations of the EANA research field. Complementary, the study is not limited to the theoretical findings since it advances the understanding of empirical network analysis, whereas it offers a library of analysis initiatives, methods to derive EA data and guidelines to help experts through the analysis process (EANA process). Finally, we also add to the EANA knowledge base two new EANA methods which were also empirically evaluated. We expect that results can enhance the awareness researchers and practitioners about the EA network-based analysis´ efficacy and utility, a step necessary to develop more rationally grounded methods and tools to support the EA management considering structural aspects.
técnicas ou métodos para extrair informações arquiteturais sobre um aspecto de interesse sobre a AE, a partir de modelos, e com o objetivo de dar suporte aos especialistas no gerenciamento da AE ou ainda pra informar seus stakeholders. Nesta tese, a AE é modelada como uma rede complexa, um conceito originário da teoria de redes, com o objetivo de analisar aspectos estruturais de AE. Durante o estudo exploratório sobre a análise de redes aplicada ao contexto de AE, constatamos a ausência de fundamentos conceituais básicos como, por exemplo, um nivelamento conceitual entre os autores dos trabalhos; desconhecimento sobre a abrangência dos estudos de análise estrutural no contexto de AE e finalmente, carência de informação acerca do processo de análise estrutural realizado nos trabalhos. Nosso objetivo principal na tese é investigar como as métricas e métodos de análise de redes podem sem aplicados no contexto de análise de AE. Métodos qualitativos de pesquisa como revisão sistemática de literatura, análise temática e design science research foram utilizados em três fases complementares e inter-relacionadas. Primeiramente, para coletar e consolidar o conhecimento sobre abordagens de análise de AE existentes na literatura. Numa segunda etapa, o objetivo foi traçar um entendimento abrangente sobre os principais conceitos envolvidos na análise estrutural de AE, mapeando seus métodos e técnicas utilizados, culminando com o design de quatro artefatos propostos: um meta-modelo para análise de redes no contexto de AE; uma biblioteca reunindo as iniciativas de análise extraídas dos artigos; um processo de análise de redes para AE e uma estratégia para derivação de dados. Na terceira e última etapa, investigou-se o uso dos artefatos avaliando-os empiricamente por meio de suas instanciações e da opinião de especialistas em AE de três organizações multinacionais alemãs. Os resultados foram positivos considerando, dentre outros critérios, a eficácia e utilidade dos artefatos propostos nos seus respectivos contextos. Como contribuições, esta pesquisa define os conceitos fundamentais para análise de redes em AE, além de avançar no entendimento acerca da análise empírica de redes naquele contexto, uma vez que apresenta um catálogo de métricas e métodos para derivação de dados, além de um processo para auxiliar os especialistas ao longo da execução da análise. Finalmente, a pesquisa também contribui para a base de conhecimento com dois métodos de análise validados também empiricamente. Com base nos resultados, espera-se corroborar o potencial da análise de AE baseada em redes, sua eficácia e utilidade para pesquisadores e práticos, além de estimular a adoção e desenvolvimento de ferramental para suportar o gerenciamento de AE, considerando seus aspectos estruturais.
APA, Harvard, Vancouver, ISO, and other styles
25

Fan, Yao-Long. "Re-engineering the solicitation management system." CSUSB ScholarWorks, 2006. https://scholarworks.lib.csusb.edu/etd-project/3179.

Full text
Abstract:
The scope of this project includes a re-engineering of the internal architecture of the Solicitation Management System (SMS), a web-based application that facilitates the running of grant proposal solicitations for the Office of Technology Transfer and Commercialization at California State University San Bernardino (CSUSB). A goal of the project is to increase consistency and efficiency of the code base of the system, making it easier to understand, maintain, and extend. The previous version of SMS was written to rely on the Spring and Hibernate frameworks. The project includes a restructuring of the system to remove reliance on the Spring framework, but maintain reliance on Hibernate. The result is an updated version of the SMS. The system was written using current technologies such as Java, JSP, and CSS.
APA, Harvard, Vancouver, ISO, and other styles
26

El-Geder, Suad. "Performance evaluation using multiple controllers with different flow setup modes in the software defined network architecture." Thesis, Brunel University, 2017. http://bura.brunel.ac.uk/handle/2438/15737.

Full text
Abstract:
In this thesis, a scheme of using multiple controllers which handle multiple network devices has been proposed, while using OpenFlow controllers in the proactive operations paradigm, and this in order to face the problem of using a single controller in the SDN model, including the lack of reliability and scalability on such a model. The main characteristic of this new approach are focused on the ability to design a dynamic and highly programmable network, moving the intelligence from the underlying systems to the network itself through a controller. To evaluate the proper effects of this new approach, different dynamic and programmable networks that could simulate real scenarios and measure their performance contrasting the obtained results with the pragmatic theory has been implemented. The SDN (Software-Defined Network) controller (Open Daylight), has been utilized, and thoroughly examined. Different sort of nets has been worked out through diverse Open Daylight functionalities, either implementing the intelligence of the controller (bundle), or going through it by an outside intelligent application (External Orchestrator), and eventually sending it through Open Daylight (by making Open Daylight work as an interpreter/translator from its language to OpenFlow or another protocol language). Summing up, the scheme that has been proposed in this research which is the multiple-proactive mode approach and the single proactive controller has scored no packet loss at all, in which implies the strength of reliability of this scheme, while the multiple reactive mode approach has a range of 1-8% packet loss ratio and the single reactive mode approach has a range of 1-25% packet loss ratio. Also, in case of delay the improvement which was obtained from our approach scored an average reduction of 13% comparing with other tested schemes. Thus, these new and interesting technologies show an astonishing capability to add more efficiency in different types of Networks.
APA, Harvard, Vancouver, ISO, and other styles
27

Li, Haorui. "Efficiency of hospitals : Evaluation of Cambio COSMIC system." Thesis, Växjö University, School of Mathematics and Systems Engineering, 2007. http://urn.kb.se/resolve?urn=urn:nbn:se:vxu:diva-1625.

Full text
Abstract:

In this modern world, healthcare has becoming a popular word in human life. People pay their attention on their health protection and treatment, but at the same time, they need to bear the high expenditure for their healthcare processing.

It is a serious problem that the government income can not afford the large expense in healthcare industry. Especially in some developing countries, healthcare problem has become the problem for the nation development.

We would like to choose this basic way to solve this problem directly, to provide the channel to improve the efficiency of healthcare system, Cambio COSMIC.

The aim to analysis COSMIC for my case study is to find out the conclusion that how does the architect design the system from the stakeholders requirement to achieve the success of improving the efficiency of healthcare system. And how to measure the success for the system achieving to improve the efficiency of healthcare system is still required to indicate.

APA, Harvard, Vancouver, ISO, and other styles
28

Roßbach, André Christian. "Evaluation of Software Architectures in the Automotive Domain for Multicore Targets in regard to Architectural Estimation Decisions at Design Time." Master's thesis, Universitätsbibliothek Chemnitz, 2015. http://nbn-resolving.de/urn:nbn:de:bsz:ch1-qucosa-163372.

Full text
Abstract:
In this decade the emerging multicore technology will hit the automotive industry. The increasing complexity of the multicore-systems will make a manual verification of the safety and realtime constraints impossible. For this reason, dedicated methods and tools are utterly necessary, in order to deal with the upcoming multicore issues. A lot of researchprojects for new hardware platforms and software frameworks for the automotive industry are running nowadays, because the paradigms of the “High-Performance Computing” and “Server/Desktop Domain” cannot be easily adapted for the embedded systems. One of the difficulties is the early suitability estimation of a hardware platform for a software architecture design, but hardly a research-work is tackling that. This thesis represents a procedure to evaluate the plausibility of software architecture estimations and decisions at design stage. This includes an analysis technique of multicore systems, an underlying graph-model – to represent the multicore system – and a simulation tool evaluation. This can guide the software architect, to design a multicore system, in full consideration of all relevant parameters and issues
In den nächsten Jahren wird die aufkommende Multicore-Technologie auf die Automobil-Branche zukommen. Die wachsende Komplexität der Multicore-Systeme lässt es nicht mehr zu, die Verifikation von Sicherheits- und Echtzeit-Anforderungen manuell auszuführen. Daher sind spezielle Methoden und Werkzeuge zwingend notwendig, um gerade mit den bevorstehenden Multicore-Problemfällen richtig umzugehen. Heutzutage laufen viele Forschungsprojekte für neue Hardware-Plattformen und Software-Frameworks für die Automobil-Industrie, weil die Paradigmen des “High-Performance Computings” und der “Server/Desktop-Domäne” nicht einfach so für die Eingebetteten Systeme angewendet werden können. Einer der Problemfälle ist das frühe Erkennen, ob die Hardware-Plattform für die Software-Architektur ausreicht, aber nur wenige Forschungs-Arbeiten berücksichtigen das. Diese Arbeit zeigt ein Vorgehens-Model auf, welches ermöglicht, dass Software-Architektur Abschätzungen und Entscheidungen bereits zur Entwurfszeit bewertet werden können. Das beinhaltet eine Analyse Technik für Multicore-Systeme, ein grundsätzliches Graphen-Model, um ein Multicore-System darzustellen, und eine Simulatoren Evaluierung. Dies kann den Software-Architekten helfen, ein Multicore System zu entwerfen, welches alle wichtigen Parameter und Problemfälle berücksichtigt
APA, Harvard, Vancouver, ISO, and other styles
29

Neel, James O'Daniell. "Simulation of an Implementation and Evaluation of the Layered Radio Architecture." Thesis, Virginia Tech, 2002. http://hdl.handle.net/10919/30862.

Full text
Abstract:
Software radio is a radio that is substantially defined in software and whose physical layer behavior can be significantly altered through changes to its software. As a primary goal of software radio is the ability to support existing and future wireless protocols, software radio necessitates the use of a rapidly reprogrammable baseband processing solution. However third generation wireless protocols represent a significant increase in complexity over second generation protocols. Due to the natural performance sacrifices that must be made when moving an application from an Application Specific Integrated Circuit (ASIC) to a general purpose processor or a digital signal processor, it is feared that reprogrammable processing solutions may not suffice for the emerging wireless protocols, which would significantly hinder the realization of software radio, particularly in the handheld domain where power consumption and chip area are critical. Recently, the Configurable Computing Lab at Virginia Tech developed a new breed of reprogrammable processor which they called â custom computing machinesâ (CCM). Representing a dramatic departure from traditional architectures used for baseband processing solutions, CCMs utilize a large number of optimized and programmable processing cores connected through a programmable mesh. Due to this architectural approach, CCMs have been promoted as supplying a level of processing power and power efficiency similar to ASICs while providing a level of reconfigurability similar to that of a DSP. Subsequently, Dr. Srikathyayani Srikanteswara proposed a new software radio architecture, known as the Layered Radio Architecture, which is intended to facilitate the inclusion of CCMs into a software radio. The primary goal of the research presented in this thesis is to demonstrate how a particular CCM, Stallion, can be used within the Layered Radio Architecture to provide sufficient processing performance, power efficiency, and reconfigurability to meet the constraints of the handheld domain through implementations of a single user adaptive receiver with adaptive complex filtering and a W-CDMA downlink rake receiver. These metrics are measured from a detailed simulation of Stallion and the Configuration Layer of the Layered Radio Architecture using advanced object oriented programming techniques that facilitate the inclusion of statistics gathering routines into normal operation. To provide perspective, these statistics are compared to the performance that could be expected from an implementation on a top-of-the-line DSP.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
30

Pontes, Danielle Pompeu Noronha. "Evolução de software baseada em avaliação de arquiteturas." Universidade de São Paulo, 2012. http://www.teses.usp.br/teses/disponiveis/3/3141/tde-14092012-122012/.

Full text
Abstract:
Este trabalho discorre sobre o estudo da utilização do método de avaliação ATAM como referência para um roteiro para evolução arquitetural. O estudo apresentado está dividido em duas partes: a elaboração de um roteiro para evolução de software e a aplicação do roteiro em um ambiente real de um sistema para automação de linhas aéreas. O objetivo é avaliar o uso do método de avaliação de arquitetura para direcionar a evolução do software. As diretrizes geradas neste trabalho orientam as ações a serem tomadas com base em evidências obtidas pela avaliação, possibilitando ao software que exiba os atributos de qualidade desejados.
This paper discusses the study of the use of ATAM evaluation method as a reference to a roadmap for architectural evolution. The present study is divided into two parts: the preparation of a roadmap for software development and implementation of the roadmap in a real environment of a system for automation of airlines. The goal is to evaluate the use of architecture evaluation method to direct the evolution of software. The guidelines generated in this work have guided the actions to be taken based on evidence obtained by the evaluation, enabling the software that displays the desired quality attributes.
APA, Harvard, Vancouver, ISO, and other styles
31

Domanos, Kyriakos. "Integrating Data Distribution Service in an Existing Software Architecture: Evaluation of the performance with different Quality of Service configurations." Thesis, Linköpings universitet, Fysik, elektroteknik och matematik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-171556.

Full text
Abstract:
The Data Distribution Service (DDS) is a flexible, decentralized, peer-to-peer communication middle-ware. This thesis presents a performance analysis of the DDS usage in the Toyota Smartness platform that is used in Toyota’s Autonomous Guided Vehicles (AGVs). The purpose is to find if DDS is suitable for internal communication between modules that reside within the Smartness platform and for external communication between AGVs that are connected in the same network. An introduction to the main concepts of DDS and the Toyota Smartness platform architecture is given together with a presentation of some earlier research that has been done in DDS. A number of different approaches of how DDS can be integrated to the Smartness platform are explored and a set of different configurations that DDS provides are evaluated. The tests that were performed in order to evaluate the usage of DDS are described in detail and the results that were collected are presented, compared and discussed. The advantages and disadvantages of using DDS are listed, and some ideas for future work are proposed.
APA, Harvard, Vancouver, ISO, and other styles
32

Johansson, Gustav. "Investigating differences in response time and error rate between a monolithic and a microservice based architecture." Thesis, KTH, Skolan för elektroteknik och datavetenskap (EECS), 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-264840.

Full text
Abstract:
With great advancements in cloud computing, the microservice architecture has become a promising architectural style for enterprise software. It has been proposed to cope with problems of the traditional monolithic architecture which includes slow release cycles, limited scalability and low developer productivity. Therefore, this thesis aims to investigate the affordances and challenges of adopting microservices as well as the difference in performance compared to the monolithic approach at one of Sweden’s largest banks, SEB - the Scandinavian Individual Bank. The investigation consisted of a literature study of research papers and official documentation of microservices. Moreover, two applications were developed and deployed using two different system architectures - a monolithic architecture and a microservice architecture. Performance tests were executed on both systems to gather quantitative data for analysis. The two metrics investigated in this study were response time and error rate. The results indicate the microservice architecture has a significantly higher error rate but a slower response time than the monolithic approach, further strengthening the results of Ueda et. al. [47] and Villamizar et. al. [48]. The findings have then been discussed with regards to the challenges and complexity involved in implementing distributed systems. From this study, it becomes clear the complexity shifts from inside the application out towards infrastructure with a microservice architecture. Therefore, microservices should not be seen as a silver bullet. Rather, the type of architecture is highly dependent on the scope of the project and the size of the organization.
Med stora framstegen inom molntjänster har microservice arkitekturen kommit att bli en lämplig kandidat för utveckling av företagsprogramvara. Denna typ av systemarkitektur har föreslagits att lösa de problem som den traditionella monolitiska arkitekturen medför; långsamma lanseringar, begränsad skalbarhet och låg produktivitet. Således fokuserar denna avhandling på att utforska de möjligheter samt utmaningar som följer vid adoptering av microservices samt skillnaden i prestanda jämfört med den monolitiska arkitekturen. Detta undersöktes på en av Sveriges största banker, SEB, den Skandinaviska Enskilda Banken. Utredningen bestod av en litteraturstudie av vetenskapliga artiklar samt officiell dokumentation för microservices. Dessutom utvecklades och lanserades två applikationer byggt med två olika typer av systemarkitektur - en som monolitisk arkitektur och den andra som en microservice arkitektur. Prestandatest utfördes sedan på båda systemen för att samla kvantitativ data för analys. De två nyckelvardena som undersöktes i denna studie var responstid och felfrekvens. Resultaten indikerar att microservice arkitekturen har en signifikant högre felfrekvens men en långsammare responstid än den monolitiska arkitekturen, vilket stärker resultaten av Ueda et. al. [47] och Villamizar et. al. [48]. Forskningsresultaten har diskuterats med hänsyn till den komplexitet och de utmaningar som följer vid implementering av distribuerade system. Från denna studie blir det tydligt att komplexiteten i en microservice arkitektur skiftar från inuti applikationen ut till infrastrukturen. Således borde microservices inte ses som en silverkula. Istället är valet av systemarkitektur strikt beroende på omfattningen av projektet samt storleken på organisationen i fråga.
APA, Harvard, Vancouver, ISO, and other styles
33

Johansson, Per, and Henric Holmberg. "On the Modularity of a System." Thesis, Malmö högskola, Teknik och samhälle, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20183.

Full text
Abstract:
Den här uppsatsen behandlar skapandet och designen av en arkitektur över ett system för behandling av depression och andra psykiska sjukdomar via internet, kallat Melencolia. Ett av kraven för detta projekt är att skapa ett system som kan utökas i framtiden. Vi har härlett detta krav till begreppet modularitet och för att skapa en modulär arkitektur för Melencolia har vi undersökt vad begreppet innebär och härlett det till att vara ett kvalitetsdrag hos flera kvalitetsattribut däribland ”maintainability” och ”reusability”. Med hjälp av ”Attribute Driven Design” kan en arkitektur skapas som fokuserar kring en viss typ av kvalitetsattribut. Eftersom modularitet inte är ett kvalitetsattribut utan en kvalitetsegenskap har vi varit tvungna att ändra indata till denna metod, från kvalitetsattribut till kvalitetsegenskap. Vidare har vi härlett och lagt fram en ny metod för att mäta kvalitetsegenskaper i en mjukvaruarkikektur.Slutligen har vi använt vår metod för att mäta graden av modularitet i Melencolias arkitektur.
This thesis considers the problem of creating and designing an architecture for a software project that will result in a system for treatment of depression on the Internet. One of the requirements for this project, named by Melencolia, is to create a system which can be extended in the future. From this requirement we have derived the concept of modularity. In order to create a modular architecture we have concluded that modularity is a quality characteristic of multiple quality attributes such as "maintainability" and "reusability".We deploy Attribute-Driven Design (ADD) in this Melencolia project. By doing this, an architecture that is focused around modularity can be created. Since modularity is not a quality attribute, but rather a quality characteristic, we had to change the input to ADD from a quality attribute to a quality characteristic.Furthermore, we derive and propose a new method for quality characteristic evaluation of software architectures.Finally we apply our aforementioned method on the architecture of Melencolia and by doing this we get an indication on how well our proposed architecture satisfies modularity.
APA, Harvard, Vancouver, ISO, and other styles
34

Al-Naeem, Tariq Abdullah Computer Science &amp Engineering Faculty of Engineering UNSW. "A quality-driven decision-support framework for architecting e-business applications." Awarded by:University of New South Wales. Computer Science and Engineering, 2006. http://handle.unsw.edu.au/1959.4/23419.

Full text
Abstract:
Architecting e-business applications is a complex design activity. This is mainly due to the numerous architectural decisions to be made, including the selection of alternative technologies, software components, design strategies, patterns, standards, protocols, platforms, etc. Further complexities arise due to the fact that these alternatives often vary considerably in their support for different quality attributes. Moreover, there are often different groups of stakeholders involved, with each having their own quality goals and criteria. Furthermore, different architectural decisions often include interdependent alternatives, where the selection of one alternative for one particular decision impacts the selections to be made for alternatives from other different decisions. There have been several research efforts aiming at providing sufficient mechanisms and tools for facilitating the architectural evaluation and design process. These approaches, however, address architectural decisions in isolation, where they focus on evaluating a limited set of alternatives belonging to one architectural decision. This has been the primary motivation behind the development of the Architectural DEcision-Making Support (ADEMS) framework, which basically aims at supporting stakeholders and architects during the architectural decision-making process by helping them determining a suitable combination of architectural alternatives. ADEMS framework is an iterative process that leverages rigorous quantitative decision-making techniques available in the literature of Management Science, particularly Multiple Attribute Decision-Making (MADM) methods and Integer Programming (IP). Furthermore, due to the number of architectural decisions involved as well as the variety of available alternatives, the architecture design space is expected to be huge. For this purpose, a query language has been developed, known as the Architecture Query Language (AQL), to aid architects in exploring and analyzing the design space in further depth, and also in examining different ???what-if??? architectural scenarios. In addition, in order to support leveraging ADEMS framework, a support tool has been implemented for carrying out the sophisticated set of mathematical computations and comparisons of the large number of architectural combinations, which might otherwise be hard to conduct using manual techniques. The primary contribution of the tool is in its help to identify, evaluate, and rank all potential combinations of alternatives based on their satisfaction to quality preferences provided by the different stakeholders. Finally, to assess the feasibility of ADEMS, three different case studies have been conducted relating to the architectural evaluation of different e-business and enterprise applications. Results obtained for the three case studies were quite positive as they showed an acceptable accuracy level for the decisions recommended by ADEMS, and at a reasonable time and effort costs for the different system stakeholders.
APA, Harvard, Vancouver, ISO, and other styles
35

Knutsson, Tor. "Implementation and evaluation of data persistence tools for temporal versioned data models." Thesis, Department of Computer and Information Science, 2009. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-19979.

Full text
Abstract:

The purpose of this thesis was to investigate different concepts and tools which could support the development of a middleware which persists a temporal and versioned relational data model in an enterprise environment. Further requirements for the target application was that changes to the data model had to be facilitated, so that a small change to the model would not result in changes in several files and application layers. Other requirements include permissioning and audit tracing. In the thesis the reader is presented with a comparison of a set of tools for enterprise development and object/relational mapping. One of the tools, a code generator, is chosen as a good candidate to match the requirements of the project. An implementation is presented, where the chosen tool is used. An XML-based language which is used to define a data model and to provide input data for the tool is presented. Other concepts concerning the implementation is then described in detail. Finally, the author discusses alternative solutions and future improvements.

APA, Harvard, Vancouver, ISO, and other styles
36

Pereira, Carlos Eduardo da Silva. "Dynamic evaluation for reactive scenarios." Doctoral thesis, Universidade de Aveiro, 2016. http://hdl.handle.net/10773/18043.

Full text
Abstract:
Doutoramento em Informática
A natureza dinâmica de cenários como Ambient Assisting Living e ambientes pervasivos e ubíquos criam contextos de avaliação exigentes que não são completamente considerados pelos métodos existentes. Esta tese defende que são possíveis avaliações que tenham em consideração a natureza dinâmica e heterogénea de ambientes reactivos, integrando aspectos como percepção e dependência de contexto, adaptabilidade ao utilizador, gestão de eventos complexos e diversidade de ambientes. O principal objectivo deste trabalho foi desenvolver uma solução que forneça aos avaliadores a possibilidade de definir e aplicar avaliações a utilizadores suportadas por um modelo de avaliação flexível, permitindo a criação e reutilização de instrumentos e especificações de avaliação sem modificar a infraestrutura geral. Para atingir este objectivo foi seguida uma abordagem de engenharia envolvendo: a) definição de requisitos; b) conceptualização de uma solução geral contendo um paradigma, uma metodologia, um modelo e uma arquitectura; c) implementação dos componentes nucleares; d) desenvolvimento e teste de provas de conceito. Como resultado principal obteve-se uma solução de avaliação dinâmica para ambientes reactivos integrando três partes essenciais: um paradigma, uma metodologia e uma arquitectura de suporte. No seu conjunto, esta solução permite a criação de sistemas de avaliação escaláveis, flexíveis e modulares para concepção de avaliações e aplicação em ambientes reactivos.
The dynamic nature of scenarios such as Ambient Assisting Living and Ubiquitous and Pervasive environments turns them into challenging evaluation contexts not properly addressed by existing methods. We argue that it is possible to have evaluations that take into consideration the dynamic and heterogeneous nature of reactive environments by integrating aspects such as context-awareness, user adaptability, complex event handling, and environment diversity. In this context, the main objective of this work was to develop a solution providing evaluators with the ability to de ne and apply evaluation tests to end-users supported by a exible evaluation model allowing them to create or reuse evaluation instruments and speci cations without changing the infrastructure or the need for other logistical necessities. To pursue this goal, we adopted an engineering approach encompassing: a) requirements de nition; b) conceptualization of a general solution comprising paradigm, methodology, model, and architecture; c) implementation of its core components; and d) development and deployment of a proof of concept. The result was a dynamic evaluation solution for reactive environments based on three major parts: a paradigm, a methodology and its model, and a support architecture. Altogether, they enable the creation of scalable, exible and modular evaluation systems for evaluation design and application in reactive environments. Overall, we consider that the proposed approach, due to its exibility and scope, widely surpasses the goals considered on the onset of this work. With a broad range of features it establishes itself as a general purpose evaluation solution, potentially applicable to a wider range of scenarios, and fostering the creation of ubiquitous and continuous evaluation systems.
APA, Harvard, Vancouver, ISO, and other styles
37

Kreku, J. (Jari). "Early-phase performance evaluation of computer systems using workload models and SystemC." Doctoral thesis, Oulun yliopisto, 2012. http://urn.fi/urn:isbn:9789514299902.

Full text
Abstract:
Abstract Novel methods and tools are needed for the performance evaluation of future embedded systems due to the increasing system complexity. Systems accommodate a large number of on-terminal and or downloadable applications offering the users with numerous services related to telecommunication, audio and video, digital television, internet and navigation. More flexibility, scalability and modularity is expected from execution platforms to support applications. Digital processing architectures will evolve from the current system-on-chips to massively parallel computers consisting of heterogeneous subsystems connected by a network-on-chip. As a consequence, the overall complexity of system evaluation will increase by orders of magnitude. The ABSOLUT performance simulation approach presented in this thesis combats evaluation complexity by abstracting the functionality of the applications with workload models consisting of instruction-like primitives. Workload models can be created from application specifications, measurement results, execution traces, or the source code. Complexity of execution platform models is also reduced since the data paths of processing elements need not be modelled in detail and data transfers and storage are simulated only from the performance point of view. The modelling approach enables early evaluation since mature hardware or software is not required for the modelling or simulation of complete systems. ABSOLUT is applied to a number of case studies including mobile phone usage, MP3 playback, MPEG4 encoding and decoding, 3D gaming, virtual network computing, and parallel software-defined radio applications. The platforms used in the studies represent both embedded systems and personal computers, and at the same time both currently existing platforms and future designs. The results obtained from simulations are compared to measurements from real platforms, which reveals an average difference of 12% in the results. This exceeds the accuracy requirements expected from virtual system-based simulation approaches intended for early evaluation
Tiivistelmä Sulautettujen tietokonejärjestelmien suorituskyvyn arviointi muuttuu yhä haastavammaksi järjestelmien kasvavan kompleksisuuden vuoksi. Järjestelmissä on suuri määrä sovelluksia, jotka tarjoavat käyttäjälle palveluita liittyen esimerkiksi telekommunikaatioon, äänen ja videokuvan toistoon, internet-selaukseen ja navigaatioon. Tästä johtuen suoritusalustoilta edellytetään yhä enemmän joustavuutta, skaalautuvuutta ja modulaarisuutta. Suoritusarkkitehtuurit kehittyvät nykyisistä System-on-Chip (SoC) -ratkaisuista Network-on-Chip (NoC) -rinnakkaistietokoneiksi, jotka koostuvat heterogeenisistä alijärjestelmistä. Sovellusten ja suoritusalustan muodostaman järjestelmän suorituskyvyn arviointiin tarvitaan uusia menetelmiä ja työkaluja, joilla kompleksisuutta voidaan hallita. Tässä väitöskirjassa esitettävä ABSOLUT-simulointimenetelmä pienentää suorituskyvyn arvioinnin kompleksisuutta abstrahoimalla sovelluksen toiminnallisuutta työkuormamalleilla, jotka koostuvat kuormaprimitiiveistä suorittimen käskyjen sijaan. Työkuormamalleja voidaan luoda sovellusten spesifikaatioista, mittaustuloksista, suoritusjäljistä tai sovellusten lähdekoodeista. Suoritusalustoista ABSOLUT-menetelmä käyttää yksinkertaisia kapasiteettimalleja toiminnallisten mallien sijaan: suoritinarkkitehtuurit mallinnetaan korkealla tasolla ja tiedonsiirto ja tiedon varastointi mallinnetaan vain suorituskyvyn näkökulmasta. Menetelmä mahdollistaa aikaisen suorituskyvyn arvioinnin, koska malleja voidaan luoda ja simuloida jo ennen valmiin sovelluksen tai suoritusalustan olemassaoloa. ABSOLUT-menetelmää on käytetty useissa erilaisissa kokeiluissa, jotka sisälsivät esimerkiksi matkapuhelimen käyttöä, äänen ja videokuvan toistoa ja tallennusta, 3D-pelin pelaamista ja digitaalista tiedonsiirtoa. Esimerkeissä käytetiin tyypillisiä suoritusalustoja sekä kotitietokoneiden että sulautettujen järjestelmien maailmasta. Lisäksi osa esimerkeistä pohjautui tuleviin tai keksittyihin suoritusalustoihin. Osa simuloinneista on varmennettu vertaamalla simulointituloksia todellisista järjestelmistä saatuihin mittaustuloksiin. Niiden välillä huomattiin keskimäärin 12 prosentin poikkeama, mikä ylittää aikaisen vaiheen suorituskyvyn simulointimenetelmiltä vaadittavan tarkkuuden
APA, Harvard, Vancouver, ISO, and other styles
38

Toresson, Gabriel. "Documenting and Improving the Design of a Large-scale System." Thesis, Linköpings universitet, Programvara och system, 2019. http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-157733.

Full text
Abstract:
As software systems become increasingly larger and more complex, the need to make them easily maintained increases, as large systems are expected to last for many years. It has been estimated that system maintenance is a large part of many IT-departments’ software develop­ment costs. In order to design a complex system to be maintainable it is necessary to introduce structure, often as models in the form of a system architecture and a system design. As development of complex large-scale systems progresses over time, the models may need to be reconstructed. Perhaps because development may have diverted from the initial plan, or because changes had to be made during implementation. This thesis presents a reconstructed documentation of a complex large-scale system, as well as suggestions for how to improve the existing design based on identified needs and insufficiencies. The work was performed primarily using a qualitative manual code review approach of the source code, and the proposal was generated iteratively. The proposed design was evaluated and it was concluded that it does address the needs and insufficiencies, and that it can be realistically implemented.
APA, Harvard, Vancouver, ISO, and other styles
39

Harbaoui, Ahmed. "Vers une modélisation et un dimensionnement automatique des systèmes répartis." Phd thesis, Université de Grenoble, 2011. http://tel.archives-ouvertes.fr/tel-00649967.

Full text
Abstract:
De nos jours, les systèmes distribués sont caractérisés par une complexité croissante de l'architecture, des fonctionnalités et de la charge soumise. Cette complexité induit souvent une perte de la qualité de service offerte, ou une saturation des ressources, voire même l'indisponibilité des services en ligne, en particulier lorsque la charge est importante. Afin d'éviter les désagrèments causés par d'importantes charges et remplir le niveau attendu de la qualité de service, les systèmes nécessitent une auto-gestion, en optimisant par exemple un tier ou en le renforçant à travers la réplication. Cette propriété autonome requiert une modélisation des performances de ces systèmes. Visant cet objectif, nous développons un framework basé sur une méthodologie théorique et expérimentale d'identification automatique de modèle et de dimensionnement, fournissant en premier un modèle de réseau de file d'attente pour un système distribué. Ensuite, ce Modèle est utilisé au sein de notre framwork pour dimensionner le système à travers une analyse ou une simulation du réseau de file d'attente.
APA, Harvard, Vancouver, ISO, and other styles
40

Bromberg, David. "Résolution de l'hétérogénéité des intergiciels d'un environnement ubiquitaire." Phd thesis, Université de Versailles-Saint Quentin en Yvelines, 2006. http://tel.archives-ouvertes.fr/tel-00469445.

Full text
Abstract:
nombre croissant de dispositifs informatiques par le biais de technologies réseaux sans fil basées ou non sur des infrastructures (WLAN, Bluetooth, GSM, GPRS, UMTS). Une des problématiques majeures de l'informatique diffuse est de faire communiquer de façon dynamique, spontanée et transparente ces différents dispositifs entre eux indépendamment de leurs hétérogénéités matérielle et logicielle. Les intergiciels ont été introduits dans cet objectif, cependant étant donné leur diversité, une nouvelle source d'hétérogénéité de plus haut niveau apparaît, notamment au niveau de leur protocole d'interaction. Actuellement, deux méthodes permettent de résoudre ces incompatibilités : la substitution et la traduction de protocoles. La première requiert la conception de nouveaux intergiciels capables de s'adapter en fonction de leur environnement d'exécution afin de résoudre dynamiquement l'hétérogénéité des intergiciels existants. L'avantage de cette méthode est de fournir une interopérabilité dynamique. En revanche, son inconvénient est d'être non transparente : elle crée une nouvelle source d'hétérogénéité entre ces nouveaux intergiciels, et nécessite de développer des applications qui leur sont spécifiques. La seconde méthode, quant à elle, est transparente : elle ne requiert ni la conception de nouveaux intergiciels, ni le développement de nouvelles applications. Cependant, elle reste statique et planifiée contrairement à la précédente méthode. Dans le contexte de l'informatique diffuse, ces deux méthodes sont complémentaires. Notre contribution consiste à combiner ces deux approches. A l'aide des langages de processus, nous proposons, dans un premier temps, une spécification formelle de notre solution qui permet de résoudre l'hétérogénéité des intergiciels quels que soient la spécificité de leurs caractéristiques, de leurs protocoles et de leurs technologies. Dans un second temps, nous présentons deux systèmes, basés sur cette spécification, conçus pour résoudre : (i) les incompatibilités des protocoles de découverte de services, (ii) les incompatibilités des protocoles de communication. Leur particularité est d'assurer une interopérabilité dynamique et transparente sans requérir de modifications des applications et des intergiciels existants. A partir de nos différentes expérimentations, il apparaît que le surcoût de cette solution pour résoudre les incompatibilités de protocoles est raisonnable.
APA, Harvard, Vancouver, ISO, and other styles
41

Junior, Edson Alves de Oliveira. "SystEM-PLA: um método sistemático para avaliação de arquitetura de linha de produto de software baseada em UML." Universidade de São Paulo, 2010. http://www.teses.usp.br/teses/disponiveis/55/55134/tde-07102010-111622/.

Full text
Abstract:
A abordagem de linha de produto de software (LP) tem como objetivo principal promover a geração de produtos específicos de um determinado domínio com base na reutilização de uma infraestrutura central, chamada núcleo de artefatos. Um dos principais artefatos do núcleo de uma LP é a Arquitetura de LP (ALP), que representa a abstração de todas as arquiteturas de sistemas únicos que podem ser gerados, para um domínio específico. Avaliações de ALP são importantes, pois permitem aumentar a produtividade e a qualidade dos produtos da LP, bem como, seus resultados permitem a análise de metas de negócio e de retorno de investimento. Este trabalho propõe um método sistemático para avaliação de ALP, o SystEM-PLA (a Systematic Evaluation Method for Software Product Line Architectures). Tal método considera modelos de ALP em UML, por ser uma notação amplamente conhecida e consolidada. SystEM-PLA é composto por um metaprocesso de avaliação, diretrizes que guiam o usuário em como avaliar uma ALP e métricas básicas para modelos UML e atributos de qualidade. O método utiliza a abordagem SMarty (Stereotype-based Management of Variability), para gerenciar variabilidades em LP baseadas em UML. Análises de trade-off com o objetivo de priorizar atributos de qualidade para o desenvolvimento e evolução dos produtos de uma LP são realizadas com base na aplicação e coleta das métricas do SystEM-PLA em configurações de uma ALP. As métricas propostas para os atributos de qualidade complexidade e extensibilidade foram validadas por meio de um estudo experimental. Evidências indicaram a viabilidade de aplicação do método SystEM-PLA na indústria com base em um estudo experimental realizado com profissionais de uma empresa de grande porte no setor de desenvolvimento de software
The software product line (PL) approach aims at promoting the generation of specific products from a particular domain based on the reuse of a central infra-structure, so-called core assets. One of the main assets of a PL is the PL Architecture (PLA) that represents the abstraction of all possible single-product architectures that can be generated for a particular domain. PLA evaluations are important due to allow the increasing of the productivity and the quality of PL products, as well as their results allow business drivers and return on investment analyzes. This work proposes a Systematic Evaluation Method for Software Product Line Architectures, the SystEM-PLA. This method takes into account UML models with PLA variability explicitly represented, since UML is a widely known and consolidated notation. SystEM-PLA is composed of an evaluation meta-process, guidelines that drive the user on how to evaluate a PLA, and basic and quality attribute metrics. This method uses the proposed approach Stereotype-based Management of Variability (SMarty) to manage variabilities in UML-based PLs. Trade-off analyses to prioritize quality attributes for the development and evolution of PL products are carried out based on the application and collection of the SystEM-PLA metrics from PLA configurations. The quality attribute metrics were validated trough an experimental study. Evidences indicated the SystEM-PLA application feasibility in industry based on an experimental study, planned and conducted with professionals from a large software development organization
APA, Harvard, Vancouver, ISO, and other styles
42

Bozakov, Zdravko [Verfasser]. "Architectures for virtualization and performance evaluation in software defined networks / Zdravko Bozakov." Hannover : Technische Informationsbibliothek (TIB), 2016. http://d-nb.info/1122592132/34.

Full text
APA, Harvard, Vancouver, ISO, and other styles
43

Viklund, Joel. "Performance Evaluation of Digital Signa lProcessing Architectures Running an Acoustic Echo Cancellation Software." Thesis, Umeå universitet, Institutionen för datavetenskap, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-128573.

Full text
Abstract:
TrueVoice is a speech enhancement software by Limes Audio AB. It improves the quality of the speech signal by using methods such as acoustic echo cancellation and noise reduction. The real time signal processing of TrueVoice puts some specic requirements on the hardware, which makes some processors more suitable than others to run the code. This thesis looks for good digital signal processor candidates to port the TrueVoice code to. This is done by evaluating memory and number representation aspects in theory and processing power in practice. The conclusion is that Cirrus WM8281 is a good candidate to port TrueVoice to
APA, Harvard, Vancouver, ISO, and other styles
44

Hamada, Skander. "Design and evaluation of a visual rapid prototyping environment in an existing smart home platform." Thesis, Malmö högskola, Fakulteten för teknik och samhälle (TS), 2017. http://urn.kb.se/resolve?urn=urn:nbn:se:mau:diva-20267.

Full text
Abstract:
Recent advances in the field of Internet of Things (IoT) are placing its own vision, as a platform of interconnected services and devices, at the heart of the smart home concept. This consolidation promises a new wave of innovative products designed in the open, and in which the user takes center stage starting from the very first steps. Therefore, researchers as well as product designers in these increasingly related fields are now tasked with a more complex mission when investigating user behavior. In this thesis we consider rapid prototyping as the upcoming standard process for investigating user interactions in the future smart home. Although past research contributed with several self-contained solutions (built from scratch) to allow such investigations, no accounts were found tackling the problem from our perspective, in which the focus is on how to enable rapid prototyping in an existing proprietary smart home platform by using open standards, software and hardware. To answer this question, we conducted our research with participation of academic researchers and professional designers in the context of an academic and industrial partnership, in an ongoing smart home research project. We used an approach based on the design science research process in combination with the user centered design (UCD) and agile software development methodologies. During this thesis we performed an end to end design process starting from ideation to implementation and evaluation; an architectural blueprint was proposed and a working prototype of our visual smart home rapid prototyping environment (SHRPE) was implemented and tested. The obtained results demonstrate the feasibility of enabling visual rapid prototyping capabilities in an existing smart home platform, by using the system integration process to introduce available open standards, software and hardware tools into the platform. In addition, evaluation results of user testing confirmed that using UCD to iteratively capture user needs in such complex context is a solid approach.
APA, Harvard, Vancouver, ISO, and other styles
45

Pennycook, Simon J. "Evaluating the performance of legacy applications on emerging parallel architectures." Thesis, University of Warwick, 2012. http://wrap.warwick.ac.uk/57050/.

Full text
Abstract:
The gap between a supercomputer's theoretical maximum ("peak") floating-point performance and that actually achieved by applications has grown wider over time. Today, a typical scientific application achieves only 5-20% of any given machine's peak processing capability, and this gap leaves room for significant improvements in execution times. This problem is most pronounced for modern "accelerator" architectures - collections of hundreds of simple, low-clocked cores capable of executing the same instruction on dozens of pieces of data simultaneously. This is a significant change from the low number of high-clocked cores found in traditional CPUs, and effective utilisation of accelerators typically requires extensive code and algorithmic changes. In many cases, the best way in which to map a parallel workload to these new architectures is unclear. The principle focus of the work presented in this thesis is the evaluation of emerging parallel architectures (specifically, modern CPUs, GPUs and Intel MIC) for two benchmark codes - the LU benchmark from the NAS Parallel Benchmark Suite and Sandia's miniMD benchmark - which exhibit complex parallel behaviours that are representative of many scientific applications. Using combinations of low-level intrinsic functions, OpenMP, CUDA and MPI, we demonstrate performance improvements of up to 7x for these workloads. We also detail a code development methodology that permits application developers to target multiple architecture types without maintaining completely separate implementations for each platform. Using OpenCL, we develop performance portable implementations of the LU and miniMD benchmarks that are faster than the original codes, and at most 2x slower than versions highly-tuned for particular hardware. Finally, we demonstrate the importance of evaluating architectures at scale (as opposed to on single nodes) through performance modelling techniques, highlighting the problems associated with strong-scaling on emerging accelerator architectures.
APA, Harvard, Vancouver, ISO, and other styles
46

Källström, Hugo. "Increasing Maintainability for Android Applications : Implementation and Evaluation of Three Software Architectures on the Android Framework." Thesis, Umeå universitet, Institutionen för datavetenskap, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:umu:diva-128572.

Full text
Abstract:
Developing maintainable Android applications has been a difficult area ever since the first android smartphone was released. With better hardware and an ever growing community of open source contributors, Android development has seen drastic changes over the past years. Because of this, applications continue to grow more complex and the need to develop maintainable software is increasing. Model View Controller, Model View Presenter and Model View View-Model are three popular and well-studied software architectures which are widely used in GUI-heavy applications, these architectures have now also emerged in Android development. Together with functional and qualitative requirements from a sample application these architectures are implemented in the Android framework. The sample applications are evaluated based on modifiability, testability and performance and compared to each other in order to come to a conclusion with respect to maintainability and under what circumstances aparticular architecture should be used. Evaluation results show that Model View Controller is preferable for smaller, less complex projects where modifiability and testability are not prioritized qualitative properties for the application. Model View Presenter and Model View ViewModel both show good testability as well as modifiable properties and are well suited for applications where these properties are important.
APA, Harvard, Vancouver, ISO, and other styles
47

Khuat, Quang Hai. "Definition and evaluation of spatio-temporal scheduling strategies for 3D multi-core heterogeneous architectures." Thesis, Rennes 1, 2015. http://www.theses.fr/2015REN1S007/document.

Full text
Abstract:
Empilant une couche multiprocesseur (MPSoC) et une couche de FPGA pour former un système sur puce reconfigurable en trois dimension (3DRSoC), est une solution prometteuse donnant un niveau de flexibilité élevé en adaptant l'architecture aux applications visées. Pour une application exécutée sur ce système, l'un des principaux défis vient de la gestion de haut niveau des tâches. Cette gestion est effectuée par le service d'ordonnancement du système d'exploitation et elle doit être en mesure de déterminer, lors de l'exécution de l'application, quelle tâche est exécutée logiciellement et/ou matériellement, quand (dimension temporelle) et sur quelles ressources (dimension spatiale, c'est à dire sur quel processeur ou quelle région du FPGA) pour atteindre la haute performance du système. Dans cette thèse, nous proposons des stratégies d'ordonnancement spatio-temporel pour les architectures 3DRSoCs. La première stratégie décide la nécessité de placer une tâche matérielle et une tâche logicielle en face-à-face afin que le coût de la communication entre tâches soit minimisé. La deuxième stratégie vise à minimiser le temps d'exécution globale de l'application. Cette stratégie exploits la présence de processeurs de la couche MPSoC afin d'anticiper, en temps-réel, l'exécution d'une tâche logicielle quand sa version matérielle ne peut pas être allouée sur le FPGA. Ensuite, un outil de simulation graphique a été développé pour vérifier le bon fonctionnement des stratégies développées et aussi nous permettre de produire des résultats
Stacking a multiprocessor (MPSoC) layer and a FPGA layer to form a 3D Reconfigurable System-on- Chip (3DRSoC) is a promising solution giving a high flexibility level in adapting the architecture to the targeted application. For an application defined as a graph of parallel tasks running on the 3DRSoC system, one of the main challenges comes from the high-level management of tasks. This management is done by the scheduling service of the Operating System and it must be able to determine, on the fly, what task should be run in software and/or hardware, when (temporal dimension) and where (spatial dimension, i.e. on what processor or what area of the FPGA) in order to achieve high performance of the system. In this thesis, we propose online spatio-temporal scheduling strategies for 3DRSoCs. The first strategy decides, during the task scheduling, the need for a SW task and a HW task to communicate in face-to-face so that the communication cost between tasks is minimized. The second strategy aims at minimizing the overall execution time of the application. It exploits the presence of processors in the MPSoC layer in order to anticipate, at run-time, the SW execution of a task when its HW version cannot be allocated to the FPGA. Then, a graphical simulation tool has been developed to verify the proper functioning of the developed strategies and also enable us to produce results
APA, Harvard, Vancouver, ISO, and other styles
48

Salber, Daniel. "De l'interaction homme-machine individuelle aux systèmes multi-utilisateurs : l'exemple de la communication homme-homme mediatisée." Grenoble 1, 1995. http://tel.archives-ouvertes.fr/tel-00005060.

Full text
Abstract:
Cette thèse s'inscrit dans le domaine de l'ingénierie des systèmes multi-utilisateurs. Dans notre modèle conceptuel, un système multi-utilisateur repose sur la combinaison de trois espaces fonctionnels : l'espace de production, l'espace de coordination et l'espace de communication. Ce dernier espace, la communication entre individus, fait l'objet de notre étude et définit la communication homme-homme médiatisée. L'approche adoptée s'articule en trois niveaux : les principes, issus des sciences non-informatiques (psychologie, sociologie, éthique, etc. ), les propriétés issues des principes et destinées à guider la conception et évaluer la réalisation, et enfin les techniques de mise en oeuvre informatique. Les principes traduisent les contributions de sciences non-informatiques comme les sciences sociales ou la psychologie cognitive à notre domaine d'étude. L'expérimentation psychologique Garden Movie, qui étudie l'influence de la disposition des caméras et de la surface de travail dans une tâche collecticielle, illustre l'utilisation des principes. Les propriétés sont des caractéristiques objectives et vérifiables d'un système informatique dont le choix est guidé par les principes. Nous proposons des propriétés pour les systèmes multi-utilisateurs et les systèmes de communication homme-homme médiatisée. Les propriétés permettent de guider la conception ainsi que l'étude de l'utilisabilité d'un système. Nous présentons la plate-forme d'observation du comportement des utilisateurs et Magicien d'Oz NEIMO pour l'étude expérimentale de l'utilisabilité. Nous illustrons son utilisation pour les systèmes de communication homme-homme médiatisée avec l'expérience Supratel. Les techniques comprennent deux volets : les modèles d'architecture logicielle guident la réalisation et les outils permettent la réalisation effective. Le choix des techniques est guide par les propriétés que le système doit vérifier. Nous présentons une grille d'analyse des modèles d'architecture logicielle pour les systèmes multi-utilisateurs que nous utilisons pour évaluer les modèles proposés dans la littérature. Constatant qu'aucun n'intègre harmonieusement les trois espaces de notre modèle conceptuel et l'insuffisance des modèles pour la communication homme-homme médiatisée, nous présentons CoPAC, un modèle d'architecture logicielle pour les systèmes multi-utilisateurs et la communication médiatisée. Nous illustrons sa mise en oeuvre avec la réalisation de notre médiaspace VideoPort. En ce qui concerne les outils, nous décrivons la réalisation de la bibliothèque UserLink pour la communication de médias continus. Nous proposons aussi la taxonomie IMPACT pour l'analyse des outils de communication homme-homme médiatisée
APA, Harvard, Vancouver, ISO, and other styles
49

Coste, Nicolas. "Vers la prédiction de performance de modèles compositionnels dans les architectures GALS." Phd thesis, Université de Grenoble, 2010. http://tel.archives-ouvertes.fr/tel-00538425.

Full text
Abstract:
La validation, incluant vérification fonctionnelle et évaluation de performance, est un processus critique pour la conception de designs matériels complexes : un design fonctionnellement correct peut s'avérer incapable d'atteindre la performance ciblée. Plus un problème dans un design est identifié tard, plus son coût de correction est élevé. La validation de designs devrait donc être entreprise le plus tôt possible dans le flot de conception. Cette thèse présente un formalisme de modélisation par composition, couvrant les aspects fonctionnels et temporisés des systèmes matériels, et définit une approche d'évaluation de performance afin d'analyser les modèles construits. Le formalisme de modélisation défini, appelé Interactive Probabilistic Chain (IPC), est une algèbre de processus a temps discret. Nous avons défini une bisimulation de branchement et prouvé sa congruence par rapport à l'opérateur de composition parallèle, nous permettant une approche compositionnelle. les IPCs peuvent être vues comme une transposition des Interactive Markov Chains dans un espace de temps discret. Pour l'évaluation de performance, une IPC complètement spécifiée est transformée en une chaîne de Markov à temps discret, qui peut être analysée. De plus, nous avons défini une mesure de perfor- mance, appelée latence, et un algorithme permettant de calculer sa distribution moyenne sur le long terme. A l'aide d'outils permettant de traiter les IPCs, développés sur la base de la boîte à outils CADP, nous avons étudié les aspects de communication d'un design industriel, l'architecture xSTream, développée chez STMicroelectronics.
APA, Harvard, Vancouver, ISO, and other styles
50

Eaves, Hugh L. "Evaluating and Improving the Efficiency of Software and Algorithms for Sequence Data Analysis." VCU Scholars Compass, 2016. http://scholarscompass.vcu.edu/etd/4295.

Full text
Abstract:
With the ever-growing size of sequence data sets, data processing and analysis are an increasingly large portion of the time and money spent on nucleic acid sequencing projects. Correspondingly, the performance of the software and algorithms used to perform that analysis has a direct effect on the time and expense involved. Although the analytical methods are widely varied, certain types of software and algorithms are applicable to a number of areas. Targeting improvements to these common elements has the potential for wide reaching rewards. This dissertation research consisted of several projects to characterize and improve upon the efficiency of several common elements of sequence data analysis software and algorithms. The first project sought to improve the efficiency of the short read mapping process, as mapping is the most time consuming step in many data analysis pipelines. The result was a new short read mapping algorithm and software, demonstrated to be more computationally efficient than existing software and enabling more of the raw data to be utilized. While developing this software, it was discovered that a widely used bioinformatics software library introduced a great deal of inefficiency into the application. Given the potential impact of similar libraries to other applications, and because little research had been done to evaluate library efficiency, the second project evaluated the efficiency of seven of the most popular bioinformatics software libraries, written in C++, Java, Python, and Perl. This evaluation showed that two of libraries written in the most popular language, Java, were an order of magnitude slower and used more memory than expected based on the language in which they were implemented. The third and final project, therefore, was the development of a new general-purpose bioinformatics software library for Java. This library, known as BioMojo, incorporated a new design approach resulting in vastly improved efficiency. Assessing the performance of this new library using the benchmark methods developed for the second project showed that BioMojo outperformed all of the other libraries across all benchmark tasks, being up to 30 times more CPU efficient than existing Java libraries.
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography