Conference Paper

Using GQM for Testing Design Patterns in Real-Time and Embedded Systems on a Software Production Line

DOI: 10.1109/ITNG.2009.267 Conference: Sixth International Conference on Information Technology: New Generations, ITNG 2009, Las Vegas, Nevada, 27-29 April 2009
Source: DBLP


This article describes a methodology named Causal Analysis and Resolution (CAR) based on Goals, Questions, and Metrics (GQM) principles. Indicators are defined based on metrics for a decision-making process. Its main contributions are the construction of an information process system model and a prototype, involving GQM approach, in a quantitative definition and qualitative metrics. The CAR methodology is a process area (PA) of the Capability Maturity Model Integrated (CMMi) for software development from Carnegie Mellon University. This PA was used to eliminate systematic error cases listed in a Technical Report (TR) generated by CAR. An information system model was created to allow the elimination of defects, errors, and failures in a design pattern named IO Manager, during the test phase, and before its publication in a components library. The prototype was created using Rational Rose RealTime (RRRT) with focus on verification tests. It provided a quality assessment to the IO Manager design pattern. The use of this methodology was based on GQM and CAR along with the information process system model. The developed prototype aimed to monitor errors on design pattern tests in real-time embedded system of a software production line.

Download full-text


Available from: Luiz Alberto Vieira Dias
  • [Show abstract] [Hide abstract]
    ABSTRACT: A methodology for implement, design and test, and an instrument for quality inspection complying with the ISO 9126 standard was developed by the Research Group on Software Engineering, GPES, and TATA Consultancy Services at the Aeronautics Institute of Technology, ITA. GPES is a research group based on ITA. The methodology and the instrument use the Causal Analysis and Resolution, CAR, process area, based on the approach Goal, Question, indicator, and Measures (GQ(i)M). The main contributions were: a) the creation and implementation of indicators, funded on metrics, for a decision process, which are based on the construction of a quality assessment methodology for a software product, in a model of a information system; and b) a prototype of an instrument for quality, which involves the approach GQ(i)M in the definition of quantitative and qualitative metrics, by means of a Microsoft Excel spreadsheet. The methodology represents the sophistication of the Causal Analysis and Resolution technique for gain identification, using database techniques. In this article the methodology was applied to two different case studies.
    No preview · Conference Paper · Jan 2010
  • [Show abstract] [Hide abstract]
    ABSTRACT: In anticipation of providing a software product quality, organizations select points to be measured in developing this product but do not have a question if they really are those points that should be measured. This work aims to propose a set of indicators that show how a deterministic error affects the business of the organization. They use the approach Goal-Question-Metric (GQM) which allows the analysis of the goals of the organization questions, and from these questions develop metrics that consider the variables that impact the business of organization. Measurements are performed when justified by a clearly defined goal and the indicators are generated if they are strongly related to the business objectives of this organization. After these indicators, is a report indicating what steps the organization should take to ensure that errors do not occur further. This report is a Technical Report, which enables decision-making more assertions by the organization. After the development of these indicators, they form the basis of historical errors and historical bases of metrics.
    No preview · Conference Paper · Jan 2010
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper describes a hybrid-sampling model for bank fraud diagnosis, including those for multiple frauds in a banking system. The Multi-Layer Perceptron (MLP) network was used to analyze similarity, together with a statistical optimization model for sampling, to reduce the volume of used data in the diagnostics phase. The created MLP was utilized for banking transactions learning, in order to detect frauds. This neural network was tested with different configurations to improve diagnosis. The hybrid-sampling model was also employed to improve training results. The results have shown that the optimization strategy reduced the database volume and improved the learning process, presenting similar precisions to diagnose frauds detection, within this hybrid-sampling model.
    Full-text · Article · Apr 2013
Show more