Dissertations / Theses on the topic 'Software quality, processes and metrics'

To see the other types of publications on this topic, follow the link: Software quality, processes and metrics.

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the top 50 dissertations / theses for your research on the topic 'Software quality, processes and metrics.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Browse dissertations / theses on a wide variety of disciplines and organise your bibliography correctly.

1

Popelka, Vladimír. "Aplikace procesní analýzy při řízení kvality a testování software." Master's thesis, Vysoká škola ekonomická v Praze, 2011. http://www.nusl.cz/ntk/nusl-72729.

Full text
Abstract:
This thesis deals with questions regarding quality assurance and software testing. The subject of its theoretical part is the specification of the general concept of quality, description of standards used in the field of software product quality evaluation and finally the evaluation of software development process itself. The thesis intends to introduce the theoretical framework of software quality assurance, especially the detailed analysis of the whole software testing branch. An added value to the theoretical part constitutes the characterization of procedural approach and selected methods used towards the improvement of processes. The practical part of the thesis comprises of the exemplification -- it shows the procedural approach at software quality management, applied to a selected IT company. The main aim of the practical part is to create a purposeful project for optimization of quality assurance and software testing processes. The core of the matter is to accomplish the process analysis of the present condition of software testing methodology. For the purpose of process analysis and optimization project, the models of key processes will be created; these processes will then be depicted based on defined pattern. The description of the state-of-the-art of software product quality assurance processes is further supplemented by the evaluation of such processes maturity. The project for optimization of software testing and quality assurance processes comes from the process analysis of the present condition of software testing methodology, as well as from the evaluation of procedural models maturity. The essence of processes optimization is the incorporation of change requests and innovative intentions of individual processes into the resulting state of methodology draft. For the measurement of selected quality assurance and software testing processes, the configuration of efficiency indicators and their application on particular processes is implemented. The research on the of the state-of-the-art, as well as the elaboration of this whole project for optimization of software testing and quality assurance processes runs in conformity with the principles of DMAIC model of Six Sigma method.
APA, Harvard, Vancouver, ISO, and other styles
2

Gurbuz, Ozge. "An Automated Quality Measurement Approach For Business Process Models." Master's thesis, METU, 2011. http://etd.lib.metu.edu.tr/upload/12613665/index.pdf.

Full text
Abstract:
Business process modeling has become a common need for organizations. Therefore process quality is also having an important role for the organizations. The most of the quality studies are based on cost and time which can be analyzed during or after the execution of the business processes. There are also quality measures which help analyzing measures before the execution of the business processes. This type of measures can give early feedback about the processes. There are three frameworks defined in the literature for a more comprehensive measurement. One of the frameworks is adapted from software programs and it aims to enable process design to be less error-prone, understandable and maintainable. The second framework is adapted from object-oriented software designs and it provides object-oriented view to the design of the business process. The last framework is adapted from ISO/IEC Software Product Quality enabling to measure the quality of process itself rather than the design. By conducting a case study, the measures defined in the frameworks are explored in terms of applicability, automation potential and required time and effort on a set of business process model. As a result of this study it is observed that measurement takes time and requires effort and is always error-prone. Therefore, an approach is implemented by automating the measures which have automation potential, in order to decrease the required time and effort and also to increase the accuracy of the measurement. The second case study is then conducted on a set of another business process models in order to validate the approach.
APA, Harvard, Vancouver, ISO, and other styles
3

Masoud, F. A. "Quality metrics in software engineering." Thesis, University of Liverpool, 1987. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.381358.

Full text
Abstract:
In the first part of this study software metrics are classified into three categories: primitive, abstract and structured. A comparative and analytical study of metrics from these categories was performed to provide software developers, users and management with a correct and consistent evaluation of a representative sample of the software metrics available in the literature. This analysis and comparison was performed in an attempt to: assist the software developers, users and management in selecting suitable quality metric(s) for their specific software quality requirements and to examine various definitions used to calculate these metrics. In the second part of this study an approach towards attaining software quality is developed. This approach is intended to help all the people concerned with the evaluation of software quality in the earlier stages of software systems development. The approach developed is intended to be uniform, consistent, unambiguous and comprehensive and one which makes the concept of software quality more meaningful and visible. It will help the developers both to understand the concepts of software quality and to apply and control it according to the expectations of users, management, customers etc.. The clear definitions provided for the software quality terms should help to prevent misinterpretation, and the definitions will also serve as a touchstone against which new ideas can be tested.
APA, Harvard, Vancouver, ISO, and other styles
4

Jah, Muzamil. "Software metrics : usability and evaluation of software quality." Thesis, University West, Department of Economics and IT, 2008. http://urn.kb.se/resolve?urn=urn:nbn:se:hv:diva-548.

Full text
Abstract:

It is difficult to understand, let alone improve, the quality of software without the knowledge of its software development process and software products. There must be some measurement process to predict the software development, and to evaluate the software products. This thesis provides a brief view on Software Quality, Software Metrics, and Software Metrics methods that will predict and measure the specified quality factors of software. It further discusses about the Quality as given by the standards such as ISO, principal elements required for the Software Quality and Software Metrics as the measurement technique to predict the Software Quality. This thesis was performed by evaluating a source code developed in Java, using Software Metrics, such as Size Metrics, Complexity Metrics, and Defect Metrics. Results show that, the quality of software can be analyzed, studied and improved by the usage of software metrics.

APA, Harvard, Vancouver, ISO, and other styles
5

Wake, Steven A. "Predicting maintainability with software quality metrics." Thesis, Virginia Tech, 1988. http://hdl.handle.net/10919/43067.

Full text
Abstract:
Maintenance of software makes up a large fraction of the time and money spent in the software life cycle. By reducing the need for maintenance these costs can also be reduced. Predicting where maintenance is likely to occur can, help to reduce maintenance by prevention. This thesis details a study of the use of software quality;metrics to determine high complexity components in a software system. By the use of a history of maintenance done on a particular system, it is shown that a predictor equation can be developed to identify components which needed maintenance activities. This same equation can also be used to determine which components are likely to need maintenance in the future. Through the use of.these predictions and software metric complexities it should be possible to reduce the likelihood of a component needing maintenance. This might be accomplished by reducing the complexity of that component through further decomposition.
Master of Science
APA, Harvard, Vancouver, ISO, and other styles
6

Kwan, Pak Leung. "Design metrics forensics : an analysis of the primitive metrics in the Zage design metrics." Virtual Press, 1994. http://liblink.bsu.edu/uhtbin/catkey/897490.

Full text
Abstract:
The Software Engineering Research Center (SERC) Design Metrics Research Team at Ball State University has developed a design metric D(G) of the form:D(G) = D~ + DiWhere De is the architectural design metric (external design metric) and D; is the detailed design metric (internal design metric).Questions to be investigated in this thesis are:Why can D, be an indicator of the potential error modules?Why can D; be an indicator of the potential error modules?Are there any significant factors that dominate the design metrics?In this thesis, the report of the STANFINS data is evaluated by using correlation analysis, regression analysis, and several other statistical techiques. The STANFINS study is chosen because it contains approximately 532 programs, 3,000 packages and 2,500,000 lines of Ada.The design metrics study was completed on 21 programs (approximately 24,000 lines of code) which were selected by CSC development teams. Error reports were also provided by CSC personnel.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
7

Burdett, Yan Liu. "Correlation of Software Quality Metrics and Performance." NSUWorks, 2012. http://nsuworks.nova.edu/gscis_etd/109.

Full text
Abstract:
Performance is an aspect of software quality that is often not considered at early stages of software design. Common approaches to performance analysis include utilizing profiling tools after the software has been developed to find bottlenecks and executing simulation models that are either manually constructed or derived from UML design diagrams. Many projects attempt to correct performance issues by adding more powerful hardware instead of attacking the root cause. Software metrics have been used to predict many aspects of software quality such as maintainability and fault-tolerance by correlation and regression analysis. Metrics proposed by Chidamber and Kemerer, also known as the CK metric suite, have been studied extensively in software quality model analyses. These studies examined maintainability, fault tolerance, error proneness, and scalability of the software as it evolved through time. Correlations were made between metrics and the likely quality models they represented. Other metrics such as Cyclomatic Complexity by McCabe and class couplings by Martin have also been used in quality predictions. No research has been conducted to analyze relationship between performance and metrics. The goal of this research was to construct a decision tree that used software metrics to forecast performance characteristics. The relationship between metrics and performance was derived by correlation between static code metrics and three runtime variables: number of objects, method call frequency, and average method call lengths on selected software benchmarks. The decision tree was constructed using the C4.5 algorithm implemented in the WEKA software. Pearson correlation coefficients were obtained for the combined datasets from all benchmarks. The decision trees and Pearson statistics showed that weighted methods per class (WMC), total lines of code (TLOC), and coupling between objects (CBO) have significant correlation with software performance. WMC showed positive correlation with number of objects and calls. CBO correlated positively with average method call lengths and negatively with number of objects. TLOC correlated positively with number of calls.
APA, Harvard, Vancouver, ISO, and other styles
8

Yerneni, Ashok. "A reliability model incorporating software quality metrics." Thesis, Virginia Polytechnic Institute and State University, 1989. http://hdl.handle.net/10919/50098.

Full text
Abstract:
Large scale software development efforts in the past decade have posed a problem in terms of the reliability of the software. The size and complexity of software that is being developed is growing rapidly and integrating diverse pieces of software in the operational environment also poses severe reliability issues, resulting in increased development and operational costs. A number of reliability models have been defined in the literature to deal with problems of this kind. However, most of these models treat the system as a "black box" and do not consider the complexity of the software in its reliability predictions. Also, reliability is predicted after the system had been completely developed leaving little scope for any major design changes to improve system reliability. This thesis reports on an effort to develop a reliability model based on complexity metrics which characterize a software system and runtime metrics which reflect the degree of testing of the system. A complete development of the reliability model is presented here. The model is simple and reflects on our intuition of the software development process and our understanding of the significance of the complexity metrics. Credibility analysis is done on the model by simulating a number of systems and applying the model. Data collected from three FORTRAN coded systems developed for NASA Goddard was used as representative of the actual software systems. An analysis of the results is finally presented.
Master of Science
incomplete_metadata
APA, Harvard, Vancouver, ISO, and other styles
9

Sears, Kenneth. "Viewpoint quality model : a software quality model for the application of software quality metrics." Thesis, University of Birmingham, 1992. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.341698.

Full text
APA, Harvard, Vancouver, ISO, and other styles
10

Afzal, Wasif. "Metrics in Software Test Planning and Test Design Processes." Thesis, Blekinge Tekniska Högskola, Avdelningen för programvarusystem, 2007. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-6193.

Full text
Abstract:
Software metrics plays an important role in measuring attributes that are critical to the success of a software project. Measurement of these attributes helps to make the characteristics and relationships between the attributes clearer. This in turn supports informed decision making. The field of software engineering is affected by infrequent, incomplete and inconsistent measurements. Software testing is an integral part of software development, providing opportunities for measurement of process attributes. The measurement of software testing process attributes enables the management to have better insight in to the software testing process. The aim of this thesis is to investigate the metric support for software test planning and test design processes. The study comprises of an extensive literature study and follows a methodical approach. This approach consists of two steps. The first step comprises of analyzing key phases in software testing life cycle, inputs required for starting the software test planning and design processes and metrics indicating the end of software test planning and test design processes. After establishing a basic understanding of the related concepts, the second step identifies the attributes of software test planning and test design processes including metric support for each of the identified attributes. The results of the literature survey showed that there are a number of different measurable attributes for software test planning and test design processes. The study partitioned these attributes in multiple categories for software test planning and test design processes. For each of these attributes, different existing measurements are studied. A consolidation of these measurements is presented in this thesis which is intended to provide an opportunity for management to consider improvement in these processes.
00 92 51 4430327
APA, Harvard, Vancouver, ISO, and other styles
11

Lim, Edwin C. "Software metrics for monitoring software engineering projects." Thesis, Edith Cowan University, Research Online, Perth, Western Australia, 1994. https://ro.ecu.edu.au/theses/1100.

Full text
Abstract:
As part of the undergraduate course offered by Edith Cowan University, the Department of Computer Science has (as part of a year's study) a software engineering group project. The structure of this project was divided into two units, Software Engineering l and Software Engineering 2. ln Software Engineering 1, students were given the group project where they had to complete and submit the Functional Requirement and Detail System Design documentation. In Software Engineering 2, students commenced with the implementation of the software, testing and documentation. The software was then submitted for assessment and presented to the client. To aid the students with the development of the software, the department had adopted EXECOM's APT methodology as its standard guideline. Furthermore, the students were divided into groups of 4 to 5, each group working on the same problem. A staff adviser was assigned to each project group. The purpose of this research exercise was to fulfil two objectives. The first objective was to ascertain whether there is a need to improve the final year software engineering project for future students by enhancing any aspect that may be regarded as deficient. The second objective was to ascertain the factors that have the most impact on the quality of the delivered software. The quality of the delivered software was measured using a variety of software metrics. Measurement of software has mostly been ignored until recently or used without true understanding of its purpose. A subsidiary objective was to gain an understanding of the worth of software measurement in the student environment One of the conclusions derived from the study suggests that teams who spent more time on software design and testing, tended to produce better quality software with less defects. The study also showed that adherence to the APT methodology led to the project being on schedule and general team satisfaction with the project management. One of the recommendations made to the project co-ordinator was that staff advisers should have sufficient knowledge of the software engineering process.
APA, Harvard, Vancouver, ISO, and other styles
12

Alshammari, Bandar M. "Quality metrics for assessing security-critical computer programs." Thesis, Queensland University of Technology, 2011. https://eprints.qut.edu.au/49780/1/Bandar_Alshammari_Thesis.pdf.

Full text
Abstract:
Existing secure software development principles tend to focus on coding vulnerabilities, such as buffer or integer overflows, that apply to individual program statements, or issues associated with the run-time environment, such as component isolation. Here we instead consider software security from the perspective of potential information flow through a program’s object-oriented module structure. In particular, we define a set of quantifiable "security metrics" which allow programmers to quickly and easily assess the overall security of a given source code program or object-oriented design. Although measuring quality attributes of object-oriented programs for properties such as maintainability and performance has been well-covered in the literature, metrics which measure the quality of information security have received little attention. Moreover, existing securityrelevant metrics assess a system either at a very high level, i.e., the whole system, or at a fine level of granularity, i.e., with respect to individual statements. These approaches make it hard and expensive to recognise a secure system from an early stage of development. Instead, our security metrics are based on well-established compositional properties of object-oriented programs (i.e., data encapsulation, cohesion, coupling, composition, extensibility, inheritance and design size), combined with data flow analysis principles that trace potential information flow between high- and low-security system variables. We first define a set of metrics to assess the security quality of a given object-oriented system based on its design artifacts, allowing defects to be detected at an early stage of development. We then extend these metrics to produce a second set applicable to object-oriented program source code. The resulting metrics make it easy to compare the relative security of functionallyequivalent system designs or source code programs so that, for instance, the security of two different revisions of the same system can be compared directly. This capability is further used to study the impact of specific refactoring rules on system security more generally, at both the design and code levels. By measuring the relative security of various programs refactored using different rules, we thus provide guidelines for the safe application of refactoring steps to security-critical programs. Finally, to make it easy and efficient to measure a system design or program’s security, we have also developed a stand-alone software tool which automatically analyses and measures the security of UML designs and Java program code. The tool’s capabilities are demonstrated by applying it to a number of security-critical system designs and Java programs. Notably, the validity of the metrics is demonstrated empirically through measurements that confirm our expectation that program security typically improves as bugs are fixed, but worsens as new functionality is added.
APA, Harvard, Vancouver, ISO, and other styles
13

Wilburn, Cathy A. "Using the Design Metrics Analyzer to improve software quality." Virtual Press, 1994. http://liblink.bsu.edu/uhtbin/catkey/902489.

Full text
Abstract:
Effective software engineering techniques are needed to increase the reliability of software systems, to increase the productivity of development teams, and to reduce the costs of software development. Companies search for an effective software engineering process as they strive to reach higher process maturity levels and produce better software. To aid in this quest for better methods of software engineering. the Design Metrics Research Team at Ball State University has analyzed university and industry software to be able to detect error-prone modules. The research team has developed, tested and validated their design metrics and found them to be highly successful. These metrics were typically collected and calculated by hand. So that these metrics can be collected more consistently, more accurately and faster, the Design Metrics Analyzer for Ada (DMA) was created. The DMA collects metrics from the files submitted based on a subprogram level. The metrics results are then analyzed to yield a list of stress points, which are modules that are considered to be error-prone or difficult for developers. This thesis describes the Design Metrics Analyzer, explains its output and how it functions. Also, ways that the DMA can be used in the software development life cycle are discussed.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
14

Stineburg, Jeffrey. "Software reliability prediction based on design metrics." Virtual Press, 1999. http://liblink.bsu.edu/uhtbin/catkey/1154775.

Full text
Abstract:
This study has presented a new model for predicting software reliability based on design metrics. An introduction to the problem of software reliability is followed by a brief overview of software reliability models. A description of the models is given, including a discussion of some of the issues associated with them. The intractability of validating life-critical software is presented. Such validation is shown to require extended periods of test time that are impractical in real world situations. This problem is also inherent in fault tolerant software systems of the type currently being implemented in critical applications today. The design metrics developed at Ball State University is proposed as the basis of a new model for predicting software reliability from information available during the design phase of development. The thesis investigates the proposition that a relationship exists between the design metric D(G) and the errors that are found in the field. A study, performed on a subset of a large defense software system, discovered evidence to support the proposition.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
15

Barkmann, Henrike. "Quantitative Evaluation of Software Quality Metrics in Open-Source Projects." Thesis, Växjö University, School of Mathematics and Systems Engineering, 2009. http://urn.kb.se/resolve?urn=urn:nbn:se:vxu:diva-2562.

Full text
Abstract:

The validation of software quality metrics lacks statistical

significance. One reason for this is that the data collection

requires quite some effort. To help solve this problem,

we develop tools for metrics analysis of a large number of

software projects (146 projects with ca. 70.000 classes and

interfaces and over 11 million lines of code). Moreover, validation

of software quality metrics should focus on relevant

metrics, i.e., correlated metrics need not to be validated independently.

Based on our statistical basis, we identify correlation

between several metrics from well-known objectoriented

metrics suites. Besides, we present early results of

typical metrics values and possible thresholds.

APA, Harvard, Vancouver, ISO, and other styles
16

Glasberg, Daniela. "Evaluating the quality of object-oriented software using design metrics." Thesis, McGill University, 2000. http://digitool.Library.McGill.CA:80/R/?func=dbin-jump-full&object_id=33402.

Full text
Abstract:
In this thesis, we present a cognitive theory of object-oriented metrics and an empirical study which has as objectives to formally test this theory while validating the metrics and to build a post-release fault proneness prediction model. The four cognitive mechanisms which we apply in this study to object-oriented metrics are based on contemporary models of human memory. They are: familiarity, interference, fan effects and the span of working memory. Our empirical study was performed with data from a commercial Java application. We found that Depth of Inheritance Tree (DIT) is a good measure of familiarity and, as predicted, has a quadratic relationship with fault-proneness. Our hypotheses were confirmed for Import Coupling to other classes, Export Coupling and Number of Children metrics. The prediction model constructed had a good accuracy. Finally, we formulated a cost savings model and applied it to our predictive model. This demonstrated a 42% reduction in post-release costs if the prediction model is used to identify the classes that should be inspected. (Abstract shortened by UMI.)
APA, Harvard, Vancouver, ISO, and other styles
17

Pipkin, Jeffrey A. "Applying design metrics to large-scale telecommunications software." Virtual Press, 1996. http://liblink.bsu.edu/uhtbin/catkey/1036178.

Full text
Abstract:
The design metrics developed by the Design Metrics team at Ball State University are a suite of metrics that can be applied during the design phase of software development. The benefit of the metrics lies in the fact that the metrics can be applied early in the software development cycle. The suite includes the external design metric De,the internal design metric D27 D(G), the design balance metric DB, and the design connectivity metric DC.The suite of design metrics have been applied to large-scale industrial software as well as student projects. Bell Communications Research of New Jersey has made available a software system that can be used to apply design metrics to large-scale telecommunications software. This thesis presents the suite of design metrics and attempts to determine if the characteristics of telecommunications software are accurately reflected in the conventions used to compute the metrics.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
18

Jain, Aman, and Raghu ram Aduri. "Quality metrics in continuous delivery : A mixed approach." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-12804.

Full text
Abstract:
Context. Continuous delivery deals with concept of deploying the user stories as soon as they are finished rather than waiting for the sprint to end. This concept increases the chances of early improvement to the software and provides the customer with a clear view of the final product that is expected from the software organization, but little research has been done on the quality of product developed and the ways to measure it. This research is conducted in the context of presenting a checklist of quality metrics that can be used by the practitioners to ensure good quality product delivery. Objectives. In this study, the authors strive towards the accomplishment of the following objectives: the first objective is to identify the quality metrics being used in agile approaches and continuous delivery by the organizations. The second objective is to evaluate the usefulness of the identified metrics, limitations of the metrics and identify new metrics. The final objective is to is to present and evaluate a solution i.e., checklist of metrics that can be used by practitioners to ensure quality of product developed using continuous delivery. Methods. To accomplish the objectives, the authors used mixture of approaches. First literature review was performed to identify the quality metrics being used in continuous delivery. Based on the data obtained from the literature review, the authors performed an online survey using a questionnaire posted over an online questionnaire hosting website. The online questionnaire was intended to find the usefulness of identified metrics, limitations of using metrics and also to identify new metrics based on the responses obtained for the online questionnaire. The authors conducted interviews and the interviews comprised of few close-ended questions and few open-ended questions which helped the authors to validate the usage of the metrics checklist. Results. Based on the LR performed at the start of the study, the authors obtained data regarding the background of continuous delivery, research performed over continuous delivery by various practitioners as well as a list of quality metrics used in continuous delivery. Later, the authors conducted an online survey using questionnaire that resulted in ranking the usefulness of quality metrics and identification of new metrics used in continuous delivery. Based on the data obtained from the online questionnaire, a checklist of quality metrics involved in continuous delivery was generated. Conclusions. Based on the interviews conducted to validate the checklist of metrics (generated as a result of the online questionnaire), the authors conclude that the checklist of metrics is fit for use in industry, but with some necessary changes made to the checklist based on the project requirements. The checklist will act as a reminder to the practitioners regarding the quality aspects that need to be measured during product development and maybe as a starting point while planning metrics that need to be measured during the project.
APA, Harvard, Vancouver, ISO, and other styles
19

Jayachandran, Naveen. "Understanding roi metrics for software test automation." [Tampa, Fla.] : University of South Florida, 2005. http://purl.fcla.edu/fcla/etd/SFE0001241.

Full text
APA, Harvard, Vancouver, ISO, and other styles
20

Beaver, Justin. "A LIFE CYCLE SOFTWARE QUALITY MODEL USING BAYESIAN BELIEF NETWORKS." Doctoral diss., University of Central Florida, 2006. http://digital.library.ucf.edu/cdm/ref/collection/ETD/id/2353.

Full text
Abstract:
Software practitioners lack a consistent approach to assessing and predicting quality within their products. This research proposes a software quality model that accounts for the influences of development team skill/experience, process maturity, and problem complexity throughout the software engineering life cycle. The model is structured using Bayesian Belief Networks and, unlike previous efforts, uses widely-accepted software engineering standards and in-use industry techniques to quantify the indicators and measures of software quality. Data from 28 software engineering projects was acquired for this study, and was used for validation and comparison of the presented software quality models. Three Bayesian model structures are explored and the structure with the highest performance in terms of accuracy of fit and predictive validity is reported. In addition, the Bayesian Belief Networks are compared to both Least Squares Regression and Neural Networks in order to identify the technique is best suited to modeling software product quality. The results indicate that Bayesian Belief Networks outperform both Least Squares Regression and Neural Networks in terms of producing modeled software quality variables that fit the distribution of actual software quality values, and in accurately forecasting 25 different indicators of software quality. Between the Bayesian model structures, the simplest structure, which relates software quality variables to their correlated causal factors, was found to be the most effective in modeling software quality. In addition, the results reveal that the collective skill and experience of the development team, over process maturity or problem complexity, has the most significant impact on the quality of software products.
Ph.D.
School of Electrical Engineering and Computer Science
Engineering and Computer Science
Computer Engineering
APA, Harvard, Vancouver, ISO, and other styles
21

Turk, Tuna. "The Effect Of Software Design Patterns On Object-oriented Software Quality And Maintainability." Master's thesis, METU, 2009. http://etd.lib.metu.edu.tr/upload/12611104/index.pdf.

Full text
Abstract:
This study investigates the connection between design patterns, object oriented (OO) quality metrics and software maintainability. The literature on OO metrics, design patterns and software maintainability are reviewed, the relation between OO metrics and software maintainability is investigated, and then, in terms of obtained maintainability indicator metrics, the maintainability change of an application due to usage of design patterns is observed.
APA, Harvard, Vancouver, ISO, and other styles
22

Kriukov, Illia. "Multi-version software quality analysis through mining software repositories." Thesis, Linnéuniversitetet, Institutionen för datavetenskap och medieteknik (DM), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-74424.

Full text
Abstract:
The main objective of this thesis is to identify how the software repository features influence software quality during software evolution. To do that the mining software repository area was used. This field analyzes the rich data from software repositories to extract interesting and actionable information about software systems, projects and software engineering. The ability to measure code quality and analyze the impact of software repository features on software quality allows us to better understand project history, project quality state, development processes and conduct future project analysis. Existing work in the area of software quality describes software quality analysis without a connection to the software repository features. Thus they lose important information that can be used for preventing bugs, decision-making and optimizing development processes. To conduct the analysis specific tool was developed, which cover quality measurement and repository features extraction. During the research general procedure of the software quality analysis was defined, described and applied in practice. It was found that there is no most influential repository feature and the correlation between software quality and software repository features exist, but it is too small to make a real influence.
APA, Harvard, Vancouver, ISO, and other styles
23

Gray, David Philip Harry. "Software defect prediction using static code metrics : formulating a methodology." Thesis, University of Hertfordshire, 2013. http://hdl.handle.net/2299/11067.

Full text
Abstract:
Software defect prediction is motivated by the huge costs incurred as a result of software failures. In an effort to reduce these costs, researchers have been utilising software metrics to try and build predictive models capable of locating the most defect-prone parts of a system. These areas can then be subject to some form of further analysis, such as a manual code review. It is hoped that such defect predictors will enable software to be produced more cost effectively, and/or be of higher quality. In this dissertation I identify many data quality and methodological issues in previous defect prediction studies. The main data source is the NASA Metrics Data Program Repository. The issues discovered with these well-utilised data sets include many examples of seemingly impossible values, and much redundant data. The redundant, or repeated data points are shown to be the cause of potentially serious data mining problems. Other methodological issues discovered include the violation of basic data mining principles, and the misleading reporting of classifier predictive performance. The issues discovered lead to a new proposed methodology for software defect prediction. The methodology is focused around data analysis, as this appears to have been overlooked in many prior studies. The aim of the methodology is to be able to obtain a realistic estimate of potential real-world predictive performance, and also to have simple performance baselines with which to compare against the actual performance achieved. This is important as quantifying predictive performance appropriately is a difficult task. The findings of this dissertation raise questions about the current defect prediction body of knowledge. So many data-related and/or methodological errors have previously occurred that it may now be time to revisit the fundamental aspects of this research area, to determine what we really know, and how we should proceed.
APA, Harvard, Vancouver, ISO, and other styles
24

Dixon, Mark Brian. "An automated approach to the measurement and evaluation of software quality during development." Thesis, Leeds Beckett University, 1997. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.267449.

Full text
APA, Harvard, Vancouver, ISO, and other styles
25

Perera, Dinesh Sirimal. "Design metrics analysis of the Harris ROCC project." Virtual Press, 1995. http://liblink.bsu.edu/uhtbin/catkey/935930.

Full text
Abstract:
The Design Metrics Research Team at Ball State University has developed a quality design metric D(G), which consists of an internal design metric Di, and an external design metric De. This thesis discusses applying design metrics to the ROCC-Radar On-line Command Control project received from Harris Corporation. Thus, the main objective of this thesis is to analyze the behavior of D(G), and the primitive components of this metric.Error and change history reports are vital inputs to the validation of design metrics' performance. Since correct identification of types of changes/errors is critical for our evaluation, several different types of analyses were performed in an attempt to qualify the metric performance in each case.This thesis covers the analysis of 666 FORTRAN modules with approximately 142,296 lines of code.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
26

Mehmood, Raja Majid, and Gulraiz Iqbal. "Visualization of Statistical Contents." Thesis, Linnéuniversitetet, Institutionen för datavetenskap, fysik och matematik, DFM, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-8583.

Full text
Abstract:
Our project presents the research on visualization of statistical contents. Here wewill introduce the concepts of visualization, software quality metrics andproposed visualization technique (line chart). Our aim to study the existingvisualization techniques for visualization of software metrics and then proposedthe visualization approach that is more time efficient and easy to perceive byviewer.In this project, we focus on the practical aspects of visualization of multipleprojects with respect to the versions and metrics. This project also gives animplementation of proposed visualization techniques of software metrics. In thisresearch based work, we have to compare practically the proposed visualizationapproaches. We will discuss the software development life cycle of our proposedvisualization system, and we will also describe the complete softwareimplementation of implemented software.
APA, Harvard, Vancouver, ISO, and other styles
27

Bhargava, Manjari. "Analysis of multiple software releases of AFATDS using design metrics." Virtual Press, 1991. http://liblink.bsu.edu/uhtbin/catkey/834502.

Full text
Abstract:
The development of high quality software the first time, greatly depends upon the ability to judge the potential quality of the software early in the life cycle. The Software Engineering Research Center design metrics research team at Ball State University has developed a metrics approach for analyzing software designs. Given a design, these metrics highlight stress points and determine overall design quality.The purpose of this study is to analyze multiple software releases of the Advanced Field Artillery Tactical Data System (AFATDS) using design metrics. The focus is on examining the transformations of design metrics at each of three releases of AFATDS to determine the relationship of design metrics to the complexity and quality of a maturing system. The software selected as a test case for this research is the Human Interface code from Concept Evaluation Phase releases 2, 3, and 4 of AFATDS. To automate the metric collection process, a metric tool called the Design Metric Analyzer was developed.Further analysis of design metrics data indicated that the standard deviation and mean for the metric was higher for release 2, relatively lower for release 3, and again higher for release 4. Interpreting this means that there was a decrease in complexity and an improvement in the quality of the software from release 2 to release 3 and an increase in complexity in release 4. Dialog with project personnel regarding design metrics confirmed most of these observations.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
28

MEHMOOD, RAJA MAJID, and GULRAIZ IQBAL. "Visualization of Statistical Contents." Thesis, Växjö University, School of Mathematics and Systems Engineering, 2009. http://urn.kb.se/resolve?urn=urn:nbn:se:vxu:diva-5556.

Full text
Abstract:

Our project presents the research on visualization of statistical contents. Here wewill introduce the concepts of visualization, software quality metrics andproposed visualization technique (line chart). Our aim to study the existingvisualization techniques for visualization of software metrics and then proposedthe visualization approach that is more time efficient and easy to perceive byviewer.In this project, we focus on the practical aspects of visualization of multipleprojects with respect to the versions and metrics. This project also gives animplementation of proposed visualization techniques of software metrics. In thisresearch based work, we have to compare practically the proposed visualizationapproaches. We will discuss the software development life cycle of our proposedvisualization system, and we will also describe the complete softwareimplementation of implemented software.

APA, Harvard, Vancouver, ISO, and other styles
29

Yuceturk, Kerem. "Software Subcontracting Success: A Case Study On The Relationship Between Project Success And Process Metrics." Master's thesis, METU, 2005. http://etd.lib.metu.edu.tr/upload/12606802/index.pdf.

Full text
Abstract:
While software subcontracting is a common business transaction in the information technology field, metrics specifically aimed at effectiveness of software subcontracting arrangements are not commonly available. This thesis makes a review of the literature and derives such metrics from fields of software quality, COTS acquisition and IS success. A case study is performed on software subcontracting projects of a Turkish defense contractor, and the project metrics are compared according to their success. The results suggest that metrics regarding the requirements are good indicators for subcontracting success and that larger projects may enjoy more success compared to smaller ones.
APA, Harvard, Vancouver, ISO, and other styles
30

Nithi, R. V. "Using a quality model approach to define and evaluate metrics for Object-Oriented Programming Systems (MOOPS)." Thesis, University of Southampton, 1999. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.310478.

Full text
APA, Harvard, Vancouver, ISO, and other styles
31

Tkachuk, M. V., and T. F. Abbasov. "An operating model for dynamic requirements management in agile software development." Thesis, NTU "KhPI", 2018. http://repository.kpi.kharkov.ua/handle/KhPI-Press/38080.

Full text
APA, Harvard, Vancouver, ISO, and other styles
32

Ruiz, de Azua Nieto David. "VizzAnalyzer goes Eclipse!" Thesis, Växjö University, School of Mathematics and Systems Engineering, 2007. http://urn.kb.se/resolve?urn=urn:nbn:se:vxu:diva-1323.

Full text
Abstract:

The VizzAnalyzer Framework is a stand-alone tool for analyzing and visualizing the structures of large software systems. Today, it has its own limited Swing based GUI lacking a professional look & feel. Furthermore, the effort needed to extend the VizzAnalyzer with new features like automatic update, progress monitoring, help system, and integration of the Eclipse Java and C/C++ AST API is high.

In order to improve current limitations and ease the future maintenance effort we refactored the VizzAnalyzer to be a plug-in to the Eclipse platform. We removed the burden of GUI development from the authors of the VizzAnalyzer replacing the Swing GUI with a SWT based GUI, which utilizes the rich feature set provided by the Eclipse Platform. Furthermore, the we did not only provide existing features of the VizzAnalyzer as loading and binding graphs, a complex system to load dynamic plug-ins functionalities for analysis, retrieval and visualization. We implemented an update and help manager, allowed for an easy use of third party plug-ins, which are available for Eclipse, and provided product branding.

We propose that the newly created VizzAnalyzer 2.0 solved the aforementioned limitations and provides a good foundation for the future evolution of the VizzAnalyzer tool.

This master thesis documents our how the VizzAnalyzer 2.0 has been developed and implemented for the Eclipse platform, and how developers shall use the new VizzAnalyzer version.

APA, Harvard, Vancouver, ISO, and other styles
33

Lin, Burch. "Neural networks and their application to metrics research." Virtual Press, 1996. http://liblink.bsu.edu/uhtbin/catkey/1014859.

Full text
Abstract:
In the development of software, time and resources are limited. As a result, developers collect metrics in order to more effectively allocate resources to meet time constraints. For example, if one could collect metrics to determine, with accuracy, which modules were error-prone and which were error-free, one could allocate personnel to work only on those error-prone modules.There are three items of concern when using metrics. First, with the many different metrics that have been defined, one may not know which metrics to collect. Secondly, the amount of metrics data collected can be staggering. Thirdly, interpretation of multiple metrics may provide a better indication of error-proneness than any single metric.This thesis researched the accuracy of a neural network, an unconventional model, in building a model that can determine whether a module is error-prone from an input of a suite of metrics. The accuracy of the neural network model was compared with the accuracy of a linear regression model, a standard statistical model, that has the same input and output. In other words, we attempted to find whether metrics correlated with error-proneness. The metrics were gathered from three different software projects. The suite of metrics that was used to build the models was a subset of a larger collection of metrics that was reduced using factor analysis.The conclusion of this thesis is that, from the projects analyzed, neither the neural network model nor the logistic regression model provide acceptable accuracies for real use. We cannot conclude whether one model provides better accuracy than the other.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
34

Mehraban, Mehrdad. "Instant Feedback Loops – for short feedback loops and early quality assurance." Thesis, Blekinge Tekniska Högskola, Institutionen för kommunikationssystem, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-13379.

Full text
Abstract:
Context. In recent years, Software Quality Assurance (SQA) has become a crucial part of software development processes. Therefore, modern software development processes led to an increase in demand for manual and automated code quality assurance. Manual code quality reviews can be a time-consuming and expensive process with varying results. Thus, automated code reviews turn out to be a preferred alternative for mitigating this process. However, commercial and open-source static code analyzer tools often offer deep analysis with long lead time. Objectives. In this thesis work, the main aim is to introduce an early code quality assurance tool, which features a combination of software metrics. The tool should be able to examine code quality and complexity of a telecommunication grade software product such as source code of specific Ericsson product by Ericsson. This tool should encapsulate complexity and quality of a software product with regards to its efficiency, scope, flexibility, and execution time. Methods. For this purpose, the background section of the thesis is dedicated to in-depth research on software metrics included in well-known static code analyzers. Then, development environment, under investigation source code of Ericsson product, and collected software metric for evaluation were presented. Next, according to each software metric’s characteristics, point of interest, and requirement, a set of steps based on a Susman’s action research cycle were defined. Moreover, SWAT, a suitable software analytics toolkit, employed to extract conducted experiment data of each software metric from a static analyzer code named Lizard in order to detect most efficient software metrics. Outcome of conducted experiment demonstrates relationship of selected software metrics with one another. Results. The chosen software metrics were evaluated based on a variety of vital factors especially actual data from number of defects of specific Ericsson product. Highly effective software metrics from investigations in this thesis work were implemented as a new model named hybrid model to be utilized as an early quality assurance. Conclusions. The proposed model, which consist of well-performing software metrics, demonstrate an impressive performance as an early code quality indicator. Consequently, the utilized model in this master thesis could be studied in a future research to further investigate the effectiveness and robustness of this tool an early quality assurance.
APA, Harvard, Vancouver, ISO, and other styles
35

Murthy, Sindhu Dharani. "Ikriya: Simulating Software Quality Enhancement With Selected Replacement Policies." TopSCHOLAR®, 2013. http://digitalcommons.wku.edu/theses/1247.

Full text
Abstract:
The quality of information systems in any organization helps to determine theefficiency of the organization. Many organizations maintain a custom software portfolio, whose quality is important to the organization. Management would like to optimize the portfolio’s quality. Decisions about software replacement or enhancement are made based on organizational needs and priorities. The development resources allocated help in determining the quality of new software, and should be put to optimal use. Enhancing existing software might sound cheap and easy but it is not always efficient. This thesis proposes a simulation model - iKriya - for this problem. It explores the consequences of various development and maintenance policies which might be applied.These depend on the state of existing software portfolio, the queue and properties of proposed projects, and the resources available. Optimal decisions are made by the simulator by taking the above mentioned factors into consideration.
APA, Harvard, Vancouver, ISO, and other styles
36

Palmqvist, Simon. "Validating the Quality of a Big Data Java Corpus." Thesis, Linnéuniversitetet, Institutionen för datavetenskap och medieteknik (DM), 2018. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-75410.

Full text
Abstract:
Recent research within the field of Software Engineering have used GitHub, the largest hub for open source projects with almost 20 million users and 57 million repositories, to mine large amounts of source code to get more trustworthy results when developing machine and deep learning models. Mining GitHub comes with many challenges since the dataset is large and the data does not only contain quality software projects. In this project, we try to mine projects from GitHub based on earlier research by others and try to validate the quality by comparing the projects with a small subset of quality projects with the help of software complexity metrics.
APA, Harvard, Vancouver, ISO, and other styles
37

Wu, Xinhao, and Maike Zhang. "An empirical assessment of the predictive quality of internal product metrics to predict software maintainability in practice." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik, 2020. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-20149.

Full text
Abstract:
Background. Maintainability of software products continues to be an area of im- portance and interest both for practice and research. The time used for maintenance usually exceeds 70% of the whole period of software development process. At present, there is a large number of metrics that have been suggested to indicate the main- tainability of a software product. However, there is a gap in validation of proposed source code metrics and the external quality of software maintainability. Objectives. In this thesis, we aim to catalog the proposed metrics for software maintainability. From this catalog we will validate a subset of commonly proposed maintainability indicators. Methods. Through a literature review with a systematic search and selection ap- proach, we collated maintainability metrics from secondary studies on software main- tainability. A subset of commonly metrics identified in the literature review were validated in a retrospective study. The retrospective study used a large open source software "Elastic Search" as a case. We collected internal source code metrics and a proxy for maintainability of the system for 911 bug fixes in 14 version (11 experi- mental samples, 3 are verification samples) of the product. Results. Following a systematic search and selection process, we identified 11 sec- ondary studies on software maintainability. From these studies we identified 290 source code metrics that are claimed to be indicators of the maintainability of a soft- ware product. We used mean time to repair (MTTR) as a proxy for maintainability of a product. Our analysis reveals that for the "elasticsearch" software, the values of the four indicators LOC, CC, WMC and RFC have the strongest correlation with MTTR. Conclusions. In this thesis, we validated a subset of commonly proposed source code metrics for predicting maintainability. The empirical validation using a popu- lar large-scale open source system reveals that some metrics have shown a stronger correlation with a proxy for maintainability in use. This study provides important empirical evidence towards a better understanding of source code attributes and maintainability in practice. However, a single case and a retrospective study are insufficient to establish a cause effect relation. Therefore, further replications of our study design with more diverse cases can increase the confidence in the predictive ability and thus the usefulness of the proposed metrics.
APA, Harvard, Vancouver, ISO, and other styles
38

West, James F. "An examination of the application of design metrics to the development of testing strategies in large-scale SDL models." Virtual Press, 2000. http://liblink.bsu.edu/uhtbin/catkey/1191725.

Full text
Abstract:
There exist a number of well-known and validated design metrics, and the fault prediction available through these metrics has been well documented for systems developed in languages such as C and Ada. However, the mapping and application of these metrics to SDL systems has not been thoroughly explored. The aim of this project is to test the applicability of these metrics in classifying components for testing purposes in a large-scale SDL system. A new model has been developed for this purpose. This research was conducted using a number of SDL systems, most notably actual production models provided by Motorola Corporation.
Department of Computer Science
APA, Harvard, Vancouver, ISO, and other styles
39

Aniche, Mauricio Finavaro. "Context-based code quality assessment." Universidade de São Paulo, 2016. http://www.teses.usp.br/teses/disponiveis/45/45134/tde-13092016-123733/.

Full text
Abstract:
Two tasks that software engineers constantly perform are writing code that is easy to evolve and maintain, and detecting poorly written pieces of code. For the former, software engineers commonly rely on well-known software architecture styles, such as Model-View-Controller (MVC). To the latter, they rely on code metrics and code smell detection approaches. However, up to now, these code metrics and code smell approaches do not take into account underlying architectureall classes are assessed as if they were the same. In practice, software developers know that classes differ in terms of responsibilities and implementation, and thus, we expect these classes to present different levels of coupling, cohesion, and complexity. As an example, in an MVC system, Controllers are responsible for the flow between the Model and the View, and Models are responsible for representing the systems business concepts. Thus, in this thesis, we evaluate the impact of architectural roles within a system architecture on code metrics and code smells. We performed an empirical analysis in 120 open source systems, and interviewed and surveyed more than 50 software developers. Our findings show that each architectural role has a different code metric values distribution, which is a likely consequence of their specific responsibilities. Thus, we propose SATT, an approach that provides specific thresholds for architectural roles that are significantly different from others in terms of code smells. We also show that classes that play a specific architectural role contain specific code smells, which developers perceive as problems, and can impact class\' change- and defect-proneness. Based on our findings, we suggest that developers understand the responsibilities of each architectural role in their system architecture, so that code metrics and code smells techniques can provide more accurate feedback.
Duas tarefas que desenvolvedores de software constantemente fazem são escrever código fácil de ser mantido e evoluído, e detectar pedaços de código problemáticos. Para a primeira tarefa, desenvolvedores comumente fazem uso de conhecidos padrões arquiteturais, como Model-View-Controller (MVC). Para a segunda tarefa, desenvolvedores fazem uso de métricas de código e estratégias de detecção de maus cheiros de código (code smells). No entanto, até o momento, métricas de código e estratégias de detecção de maus cheiros de código não levam em conta a arquitetura do software em análise. Isso significa que todas classes são avaliadas como se umas fossem iguais às outras. Na prática, sabemos que classes são diferentes em suas responsibilidades e implementação, e portanto, esperamos que elas variem em termos de acoplamento, coesão e complexidade. Por exemplo, em um sistema MVC, Controladores são responsáveis pelo fluxo entre a camada de Modelo e a camada de Visão, e Modelos representam a visão de negócios do sistema. Nesta tese, nós avaliamos o impacto dos papéis arquiteturais em técnicas de medição de métricas de código e de detecção de maus cheiros de código. Nós realizamos um estudo empírico em 120 sistemas de código aberto, e entrevistamos e realizamos questionários com mais de 50 desenvolvedores. Nossos resultados mostram que cada papel arquitetural possui distribuições diferentes de valores de métrica de código, consequência das diferentes responsabilidades de cada papel. Como consequência, propomos SATT, uma abordagem que provê thresholds específicos para papéis arquiteturais que são significantemente diferentes de outros em termos de métricas de código. Mostramos também que classes que cumprem um papel arquitetural específico também contêm maus cheiros de código específicos. Esses maus cheiros são percebidos por desenvolvedores como problemas reais e podem fazer com que essas classes sejam mais modificadas e apresentem mais defeitos do que classes limpas. Sugerimos então que desenvolvedores entendam a arquitetura dos seus sistemas, bem como as responsabilidades de cada papel arquitetural que as classes desempenham, para que tanto métricas de código quanto estratégias de detecção de maus cheiros de código possam prover um melhor retorno.
APA, Harvard, Vancouver, ISO, and other styles
40

Allahyari, Hiva. "On the concept of Understandability as a Property of Data mining Quality." Thesis, Blekinge Tekniska Högskola, Sektionen för datavetenskap och kommunikation, 2010. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-6134.

Full text
Abstract:
This paper reviews methods for evaluating and analyzing the comprehensibility and understandability of models generated from data in the context of data mining and knowledge discovery. The motivation for this study is the fact that the majority of previous work has focused on increasing the accuracy of models, ignoring user-oriented properties such as comprehensibility and understandability. Approaches for analyzing the understandability of data mining models have been discussed on two different levels: one is regarding the type of the models’ presentation and the other is considering the structure of the models. In this study, we present a summary of existing assumptions regarding both approaches followed by an empirical work to examine the understandability from the user’s point of view through a survey. From the results of the survey, we obtain that models represented as decision trees are more understandable than models represented as decision rules. Using the survey results regarding understandability of a number of models in conjunction with quantitative measurements of the complexity of the models, we are able to establish correlation between complexity and understandability of the models.
APA, Harvard, Vancouver, ISO, and other styles
41

Deniz, Berkhan. "Investigation Of The Effects Of Reuse On Software Quality In An Industrial Setting." Master's thesis, METU, 2013. http://etd.lib.metu.edu.tr/upload/12615318/index.pdf.

Full text
Abstract:
Software reuse is a powerful tool in order to reduce development and maintenance time and cost. Any software life cycle product can be reused, not only fragments of source code. A high degree of reuse correlates with a low defect density. In the literature, many theoretical and empirical researches have examined the relationship of software reuse and quality. In this thesis, the effects of reuse on software quality are investigated in an industrial setting. Throughout this study, we worked with Turkey&rsquo
s leading defense industry company: Aselsan&rsquo
s software engineering department. We aimed to explore their real-life software projects and interpret reuse and quality relations for their projects. With this intention, we defined four different hypotheses to determine reuse and quality relations
and in order to confirm these hypotheses
we designed three separate case studies. In these case studies, we collected and calculated reuse and quality metrics i.e. Object-oriented quality metrics, reuse rates and performance measures of individual modules, fault-proneness of software components, and productivity rates of different products. Finally, by analyzing these measurements, we developed suggestions to further benefit from reuse in Aselsan through systematic improvements to the reuse infrastructure and process. Similar case studies have been reported in the literature, however, in Turkey, there are not many case studies using real-life project data, particularly in the defense industry.
APA, Harvard, Vancouver, ISO, and other styles
42

Ponomarenko, Maksym. "Maintenance of the Quality Monitor Web-Application." Thesis, Linnéuniversitetet, Institutionen för datavetenskap (DV), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-28816.

Full text
Abstract:
Applied Research in System Analysis (ARiSA) is a company specialized in the development of the customer-specific quality models and applied research work. In order to improve the quality of the projects and to reduce maintenance costs, ARiSA developed Quality Monitor (QM) – a web application for quality analysis. QM application has been originally developed as a basic program to enable customers to evaluate the quality of the sources. Therefore, the business logic of the application was simplified and certain limitations were imposed on it, which in its turn leads to a number of issues related to user experience, performance and architecture design. These aspects are important for both application as a product, and for its future promotion. Moreover, this is important for customers, as end users. Main application issues, which were added to the maintenance list are: manual data upload, insufficient server resources to handle long-running and resource consuming operations, no background processing and status reporting, simplistic presentation of analysis results and known usability issues, weak integration between analysis back-ends and front-end. ­­­­­­­­­­­In order to address known issues and to make improvements of the existing limitations, a maintenance phase of QM application is initiated. First of all, it is intended to stabilize current version and improve user experience. It also needed for refactoring and implementation of more efficient data uploads processing in the background. In addition, extended functionality of QM would fulfill customer needs and transform application from the project into a product. Extended functionality includes: automated data upload from different build processes, new data visualizations, and improvement of the current functionality according to customer comments. Maintenance phase of QM application has been successfully completed and master thesis goals are met. Current version is more stable and more responsive from user experience perspective. Data processing is more efficient, and now it is implemented as background analysis with automatic data import. User interface has been updated with visualizations for client-side interaction and progress reporting. The solution has been evaluated and tested in close cooperation with QM application customers. This thesis describes requirements analysis, technology stack with choice rationale and implementation to show maintenance results.
APA, Harvard, Vancouver, ISO, and other styles
43

Magapu, Akshay Kumar, and Nikhil Yarlagadda. "Performance, Scalability, and Reliability (PSR) challenges, metrics and tools for web testing : A Case Study." Thesis, Blekinge Tekniska Högskola, Institutionen för programvaruteknik, 2016. http://urn.kb.se/resolve?urn=urn:nbn:se:bth-12801.

Full text
Abstract:
Context. Testing of web applications is an important task, as it ensures the functionality and quality of web applications. The quality of web application comes under non-functional testing. There are many quality attributes such as performance, scalability, reliability, usability, accessibility and security. Among these attributes, PSR is the most important and commonly used attributes considered in practice. However, there are very few empirical studies conducted on these three attributes. Objectives. The purpose of this study is to identify metrics and tools that are available for testing these three attributes. And also to identify the challenges faced while testing these attributes both from literature and practice. Methods. In this research, a systematic mapping study was conducted in order to collect information regarding the metrics, tools, challenges and mitigations related to PSR attributes. The required information is gathered by searching in five scientific databases. We also conducted a case study to identify the metrics, tools and challenges of the PSR attributes in practice. The case study is conducted at Ericsson, India where eight subjects were interviewed. And four subjects working in other companies (in India) were also interviewed in order to validate the results obtained from the case company. In addition to this, few documents of previous projects from the case company are collected for data triangulation. Results. A total of 69 metrics, 54 tools and 18 challenges are identified from systematic mapping study. And 30 metrics, 18 tools and 13 challenges are identified from interviews. Data is also collected through documents and a total of 16 metrics, 4 tools and 3 challenges were identified from these documents. We formed a list based on the analysis of data that is related to tools, metrics and challenges. Conclusions. We found that metrics available from literature are overlapping with metrics that are used in practice. However, tools found in literature are overlapping only to some extent with practice. The main reason for this deviation is because of the limitations that are identified for the tools, which lead to the development of their own in-house tool by the case company. We also found that challenges are partially overlapped between state of art and practice. We are unable to collect mitigations for all these challenges from literature and hence there is a need for further research to be done. Among the PSR attributes, most of the literature is available on performance attribute and most of the interviewees are comfortable to answer the questions related to performance attribute. Thus, we conclude there is a lack of empirical research related to scalability and reliability attributes. As of now, our research is dealing with PSR attributes in particular and there is a scope for further research in this area. It can be implemented on the other quality attributes and the research can be done in a larger scale (considering more number of companies).
APA, Harvard, Vancouver, ISO, and other styles
44

Rodríguez, Martínez Cecilia. "Software quality studies using analytical metric analysis." Thesis, KTH, Kommunikationssystem, CoS, 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-120325.

Full text
Abstract:
Today engineering companies expend a large amount of resources on the detection and correction of the bugs (defects) in their software. These bugs are usually due to errors and mistakes made by programmers while writing the code or writing the specifications. No tool is able to detect all of these bugs. Some of these bugs remain undetected despite testing of the code. For these reasons, many researchers have tried to find indicators in the software’s source codes that can be used to predict the presence of bugs. Every bug in the source code is a potentially failure of the program to perform as expected. Therefore, programs are tested with many different cases in an attempt to cover all the possible paths through the program to detect all of these bugs. Early prediction of bugs informs the programmers about the location of the bugs in the code. Thus, programmers can more carefully test the more error prone files, and thus save a lot of time by not testing error free files. This thesis project created a tool that is able to predict error prone source code written in C++. In order to achieve this, we have utilized one predictor which has been extremely well studied: software metrics. Many studies have demonstrated that there is a relationship between software metrics and the presence of bugs. In this project a Neuro-Fuzzy hybrid model based on Fuzzy c-means and Radial Basis Neural Network has been used. The efficiency of the model has been tested in a software project at Ericsson. Testing of this model proved that the program does not achieve high accuracy due to the lack of independent samples in the data set. However, experiments did show that classification models provide better predictions than regression models. The thesis concluded by suggesting future work that could improve the performance of this program.
Idag spenderar ingenjörsföretag en stor mängd resurser på att upptäcka och korrigera buggar (fel) i sin mjukvara. Det är oftast programmerare som inför dessa buggar på grund av fel och misstag som uppkommer när de skriver koden eller specifikationerna. Inget verktyg kan detektera alla dessa buggar. Några av buggarna förblir oupptäckta trots testning av koden. Av dessa skäl har många forskare försökt hitta indikatorer i programvarans källkod som kan användas för att förutsäga förekomsten av buggar. Varje fel i källkoden är ett potentiellt misslyckande som gör att applikationen inte fungerar som förväntat. För att hitta buggarna testas koden med många olika testfall för att försöka täcka alla möjliga kombinationer och fall. Förutsägelse av buggar informerar programmerarna om var i koden buggarna finns. Således kan programmerarna mer noggrant testa felbenägna filer och därmed spara mycket tid genom att inte behöva testa felfria filer. Detta examensarbete har skapat ett verktyg som kan förutsäga felbenägen källkod skriven i C ++. För att uppnå detta har vi utnyttjat en välkänd metod som heter Software Metrics. Många studier har visat att det finns ett samband mellan Software Metrics och förekomsten av buggar. I detta projekt har en Neuro-Fuzzy hybridmodell baserad på Fuzzy c-means och Radial Basis Neural Network använts. Effektiviteten av modellen har testats i ett mjukvaruprojekt på Ericsson. Testning av denna modell visade att programmet inte Uppnå hög noggrannhet på grund av bristen av oberoende urval i datauppsättningen. Men gjordt experiment visade att klassificering modeller ger bättre förutsägelser än regressionsmodeller. Exjobbet avslutade genom att föreslå framtida arbetet som skulle kunna förbättra detta program.
Actualmente las empresas de ingeniería derivan una gran cantidad de recursos a la detección y corrección de errores en sus códigos software. Estos errores se deben generalmente a los errores cometidos por los desarrolladores cuando escriben el código o sus especificaciones.  No hay ninguna herramienta capaz de detectar todos estos errores y algunos de ellos pasan desapercibidos tras el proceso de pruebas. Por esta razón, numerosas investigaciones han intentado encontrar indicadores en los códigos fuente del software que puedan ser utilizados para detectar la presencia de errores. Cada error en un código fuente es un error potencial en el funcionamiento del programa, por ello los programas son sometidos a exhaustivas pruebas que cubren (o intentan cubrir) todos los posibles caminos del programa para detectar todos sus errores. La temprana localización de errores informa a los programadores dedicados a la realización de estas pruebas sobre la ubicación de estos errores en el código. Así, los programadores pueden probar con más cuidado los archivos más propensos a tener errores dejando a un lado los archivos libres de error. En este proyecto se ha creado una herramienta capaz de predecir código software propenso a errores escrito en C++. Para ello, en este proyecto se ha utilizado un indicador que ha sido cuidadosamente estudiado y ha demostrado su relación con la presencia de errores: las métricas del software. En este proyecto un modelo híbrido neuro-disfuso basado en Fuzzy c-means y en redes neuronales de función de base radial ha sido utilizado. La eficacia de este modelo ha sido probada en un proyecto software de Ericsson. Como resultado se ha comprobado que el modelo no alcanza una alta precisión debido a la falta de muestras independientes en el conjunto de datos y los experimentos han mostrado que los modelos de clasificación proporcionan mejores predicciones que los modelos de regresión. El proyecto concluye sugiriendo trabajo que mejoraría el funcionamiento del programa en el futuro.
APA, Harvard, Vancouver, ISO, and other styles
45

Corrêa, Ulisses Brisolara. "Aplicação de métricas de software na predição de características físicas de software embarcado." reponame:Biblioteca Digital de Teses e Dissertações da UFRGS, 2011. http://hdl.handle.net/10183/28733.

Full text
Abstract:
A complexidade dos dispositivos embarcados propõe novos desafios para o desenvolvimento de software embarcado, além das tradicionais restrições físicas. Então, a avaliação da qualidade do software embarcado e seu impacto nessas propriedades tradicionais torna-se mais importante. Conceitos como reúso abstração, coesão, acoplamento, entre outros atributos de software têm sido usados como métricas de qualidade no domínio da engenharia de software. No entanto, elas não têm sido usadas no domínio do software embarcado. No desenvolvimento de sistemas embarcados outro conjunto de ferramentas é usado para estimar as propriedades físicas, tais como: consumo de energia, ocupação de memória e desempenho. Essas ferramentas geralmente envolvem custosos processos de síntese e simulação. Nos complexos dispositivos embarcados atuais deve-se confiar em ferramentas que possam ajudar na exploração do espaço de projeto ainda nos níveis mais altos de abstração, identificando a solução que representa a melhor estratégia de projeto em termos da qualidade de software, enquanto, simultaneamente, atenda aos requisitos físicos. Neste trabalho é apresentada uma análise da correlação entre métricas de qualidade de software, que podem ser extraídas antes do sistema ser sintetizado, e as métricas físicas do software embarcado. Usando uma rede neural nós investigamos o uso dessas correlações para predizer o impacto que uma determinada modificação no software trará às métricas físicas do mesmo software. Esta estimativa pode ser usada para guiar decisões em direção a melhoria das propriedades físicas dos sistemas embarcados, além de manter um equilíbrio em relação às métricas de software.
The complexity of embedded devices poses new challenges to embedded software development in addition to the traditional physical requirements. Therefore, the evaluation of the quality of embedded software and its impact on these traditional properties becomes increasingly relevant. Concepts such as reuse, abstraction, cohesion, coupling, and other software attributes have been used as quality metrics in the software engineering domain. However, they have not been used in the embedded software domain. In embedded systems development, another set of tools is used to estimate physical properties such as power consumption, memory footprint, and performance. These tools usually require costly synthesis-and-simulation design cycles. In current complex embedded devices, one must rely on tools that can help design space exploration at the highest possible level, identifying a solution that represents the best design strategy in terms of software quality, while simultaneously meeting physical requirements. We present an analysis of the cross-correlation between software quality metrics, which can be extracted before the final system is synthesized, and physical metrics for embedded software. Using a neural network, we investigate the use of these cross-correlations to predict the impact that a given modification on the software solution will have on embedded software physical metrics. This estimation can be used to guide design decisions towards improving physical properties of embedded systems, while maintaining an adequate trade-off regarding software quality.
APA, Harvard, Vancouver, ISO, and other styles
46

Berry, Michael CSE UNSW. "Assessment of software measurement." Awarded by:University of New South Wales. CSE, 2006. http://handle.unsw.edu.au/1959.4/25134.

Full text
Abstract:
Background and purpose. This thesis documents a program of five studies concerned with the assessment of software measurement. The goal of this program is to assist the software industry to improve the information support for managers, analysts and software engineers by providing evidence of where opportunities for improving measurement and analysis exist. Methods. The first study examined the assessment of software measurement frameworks using models of best practice based on performance/success factors. The software measurement frameworks of thirteen organisations were surveyed. The association between a factor and the outcome experienced with the organisations' frameworks was then evaluated. The subsequent studies were more info-centric and investigated using models of information quality to assess the support provided for software processes. For these studies, information quality models targeting specific software processes were developed using practitioner focus groups. The models were instantiated in survey instruments and the responses were analysed to identify opportunities to improve the information support provided. The final study compared the use of two different information quality models for the assessing and improving information support. Assessments of the same quantum of information were made using a targeted model and a generic model. The assessments were then evaluated by an expert panel in order to identify which information quality model was more effective for improvement purposes. Results. The study of performance factors for software measurement frameworks confirmed the association of some factors with success and quantified that association. In particular, it demonstrated the importance of evaluating contextual factors. The conclusion is that factor-based models may be appropriately used for risk analysis and for identifying constraints on measurement performance. Note, however, that a follow-up study showed that some initially successful frameworks subsequently failed. This implied an instability in the dependent variable, success, that could reduce the value of factor-based models for predicting success. The studies of targeted information quality models demonstrated the effectiveness of targeted assessments for identifying improvement opportunities and suggest that they are likely to be more effective for improvement purposes than using generic information quality models. The studies also showed the effectiveness of importance-performance analysis for prioritizing improvement opportunities.
APA, Harvard, Vancouver, ISO, and other styles
47

MONTEIRO, GABRIEL VIEIRA. "THE USE OF ANALOG PROCESSES AS A FACTOR TO IMPROVE SOFTWARE QUALITY: A CASE STUDY OF ORTHO-MEDICAL SOFTWARE." PONTIFÍCIA UNIVERSIDADE CATÓLICA DO RIO DE JANEIRO, 2003. http://www.maxwell.vrac.puc-rio.br/Busca_etds.php?strSecao=resultado&nrSeq=4318@1.

Full text
Abstract:
COORDENAÇÃO DE APERFEIÇOAMENTO DO PESSOAL DE ENSINO SUPERIOR
A presente dissertação avalia um software médico através de parâmetros e critérios ergonômicos. Primeiramente descrevem- se os conceitos relativos a Tecnologia da Informação e conceitua-se o software como um dos subsistemas de qualquer sistema de Tecnologia da Informação. A partir de então, verificou-se as etapas de desenvolvimento de software, relacionados ao contexto de usabilidade. Levantaram-se também os principais problemas de interação encontrados no software. A segunda parte da pesquisa, um estudo de caso, inclui a realização de heurística, entrevistas, grupo de foco e técnica de eliciação do conhecimento (card sorting), para que os usuários pudessem expressar sua opinião com relação ao software, às dificuldades encontradas e à utilização de menus.
This thesis evaluates a medical software according to ergonomic criteria and parameters. Firstly, the concepts related to Information technology were described and the software was defined as one of the subsystems together whith the presentation of Information Technology. After that, the stages of software design were determined: context of usability. The main problems related to software interfaces were also described. The second part of this research, a case study, included heuristic approach, interview, focus groups and a knowledge elicitation (card sorting). So that users would have the opportunity to express their opinions in relation to the difficulties faced with its use.
APA, Harvard, Vancouver, ISO, and other styles
48

McEntee, Simon. "The application of intelligent software for on-line product quality monitoring in manufacturing processes." Thesis, Glasgow Caledonian University, 1996. http://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.295005.

Full text
APA, Harvard, Vancouver, ISO, and other styles
49

Glazunov, Vladimir. "Quality assessment of a large real world industry project." Thesis, Linnéuniversitetet, Institutionen för datavetenskap (DV), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:lnu:diva-31155.

Full text
Abstract:
Quality Monitor is application, which automatically analyzes software projects forquality and makes quality assessment reports. This thesis project aims to instantiate Quality Monitor for a large real-world .Net project and to extend Quality Monitor by considering other data sources than just source code. This extended analysis scope includes bug reports, features, and time reports besides .Net assemblies (code) as artifacts. Different tools were investigated for the analysis of code, bug reports, features and time reports. The analysis of .Net assemblies was implemented as none of the existing tools under evaluation met all requirements. The analysis of .Net assemblies was successfully completed; it allows the extraction data necessary for creating Call and Control Flow graphs. These graphs are used for calculating additional metrics allowing for an improved assessment of quality of the project. Implementation of .Net assembly reader was tested using large real world industrial project. Other data sources were analyzed theoretically, but excluded for further implementation. Altogether the thesis includes an analysis of possible Quality Monitor extensions including their requirements, design, and (partially) their implementation and evaluation.
APA, Harvard, Vancouver, ISO, and other styles
50

Ndenga, Malanga Kennedy. "Predicting post-release software faults in open source software as a means of measuring intrinsic software product quality." Electronic Thesis or Diss., Paris 8, 2017. http://www.theses.fr/2017PA080099.

Full text
Abstract:
Les logiciels défectueux ont des conséquences coûteuses. Les développeurs de logiciels doivent identifier et réparer les composants défectueux dans leurs logiciels avant de les publier. De même, les utilisateurs doivent évaluer la qualité du logiciel avant son adoption. Cependant, la nature abstraite et les multiples dimensions de la qualité des logiciels entravent les organisations de mesurer leur qualités. Les métriques de qualité logicielle peuvent être utilisées comme proxies de la qualité du logiciel. Cependant, il est nécessaire de disposer d'une métrique de processus logiciel spécifique qui peut garantir des performances de prédiction de défaut meilleures et cohérentes, et cela dans de différents contextes. Cette recherche avait pour objectif de déterminer un prédicteur de défauts logiciels qui présente la meilleure performance de prédiction, nécessite moins d'efforts pour la détection et a un coût minimum de mauvaise classification des composants défectueux. En outre, l'étude inclut une analyse de l'effet de la combinaison de prédicteurs sur la performance d'un modèles de prédiction de défauts logiciels. Les données expérimentales proviennent de quatre projets OSS. La régression logistique et la régression linéaire ont été utilisées pour prédire les défauts. Les métriques Change Burst ont enregistré les valeurs les plus élevées pour les mesures de performance numérique, avaient les probabilités de détection de défaut les plus élevées et le plus faible coût de mauvaise classification des composants
Faulty software have expensive consequences. To mitigate these consequences, software developers have to identify and fix faulty software components before releasing their products. Similarly, users have to gauge the delivered quality of software before adopting it. However, the abstract nature and multiple dimensions of software quality impede organizations from measuring software quality. Software quality metrics can be used as proxies of software quality. There is need for a software process metric that can guarantee consistent superior fault prediction performances across different contexts. This research sought to determine a predictor for software faults that exhibits the best prediction performance, requires least effort to detect software faults, and has a minimum cost of misclassifying components. It also investigated the effect of combining predictors on performance of software fault prediction models. Experimental data was derived from four OSS projects. Logistic Regression was used to predict bug status while Linear Regression was used to predict number of bugs per file. Models built with Change Burst metrics registered overall better performance relative to those built with Change, Code Churn, Developer Networks and Source Code software metrics. Change Burst metrics recorded the highest values for numerical performance measures, exhibited the highest fault detection probabilities and had the least cost of mis-classification of components. The study found out that Change Burst metrics could effectively predict software faults
APA, Harvard, Vancouver, ISO, and other styles
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography