Data Quality
The quality of the information is based on the quality of the basic data with which measurements and indicators are built. As data quality we means complete, with valid values (not null), correct type (eg Date, Number), and is in the proper range (zero to N , a valid date, a valid name).
You can help ensure data quality as follows:
By entering certain mandatory fields
Monitoring a "data quality index" using the defined indicator. This indicator is calculated for each data import and individual field level obtained as a ratio between the number of field values considered correct vs. the number of values considered incomplete
Fig. 22. Example of variation of quality indicator data entry
...
Process Control
You can monitor the process performance using indicators and measurements, according to their business objectives. These indicators are constructed from basic measurements with high reusability in different contexts and processes: size, effort, reuse, productivity, errors, rework. Have access to a rich set of indicators covering a wide range of processes (management, engineering and process support). Through the use of "process control graphics" can analyze data generated at each iteration of project over time. These graphs allow discrimination "noise" from "signal" using very simple statistical techniques (based on a "normal" distribution, so the data must be "homogeneous"). Given the amount of data available it is important to discriminate information from simple "noise". This can be viewed directly "statistical signal" in the data to determine points in a data series that require "explanation" of its assignable cause variation. These cases are characterized by an unexpected change in the performance of the process. Also known as "assignable causes" because they can be identified, analyzed, and taken into account to prevent recurrence. May be due to:
...
To improve the performance of your processes, the categorization of the variation in the data reported by the indicators is the first step to take. You can discern between different variation patterns, each with its implication including:
Points "dominant": a possible signal process (or fault data) that cause an overflow outside the natural statistical limits, given by + - 3 sigma variation from the mean
Points "moderate": signal a possible change process that moves the average natural boundary near the bottom or top
Points "weak": signal a possible process that causes a shift down, or up, the average
Points "near average": a possible signal that keeps the process close to the average performance
Points "ascendant": signal a possible change process that moves something the average downward or upward
...
Measurements analyst has all the information necessary to determine past performance, current and future estimated that each indicator provides so we can support better decision-making and continuous improvement of processes available through graphics, rules of pattern analysis variation, filters and other facilities like comparing baseline and extrapolation or comparing individual series. You can access this type of analysis directly from the tools
Fig. 22. 23. Normal Distribution
...
Statistical signals and user defined baseline
...
Fig. Fuzzy signals
...
Data Quality
The quality of the information is based on the quality of the basic data with which measurements and indicators are built. As data quality we means complete, with valid values (not null), correct type (eg Date, Number), and is in the proper range (zero to N , a valid date, a valid name).
You can help ensure data quality as follows:
By entering certain mandatory fields
Monitoring a "data quality index" using the defined indicator. This indicator is calculated for each data import and individual field level obtained as a ratio between the number of field values considered correct vs. the number of values considered incomplete
We can create an indicator to monitor this quality index over time, along with detailed data from the associated context for easy review.
Fig. 23. Example of variation of quality indicator data entry
...
The Context
The context is the quantitative / qualitative information associated with a graph / data of an indicator for a project / group in a period of time that gives it meaning and that it must always accompany them in order to be feasible its interpretation. This context is generated automatically by the application.
...
There are several types of context. A type of context, called "Global" is a set of events (dates) on a timeline. A special event is "SourceOrigin" and can be used to filter data from that date. This context is generated and maintained by the user through "Global Notes".
...
The data sources, and certain indicators, contain metadata to represent the components of the related activity system, in particular, the instruments, operations and intervening actors.
Fig. 25. An activity system under analysis
...
Activity patterns
Self-constructed graphical formalism which is a socio-technical systemic view that relates the following components in a solution pattern from the signals and other clues in the data:
...
It is used in Story to explain / argue the set and is based on the theory of Activity Systems and Patterns
Fig. 25. An activity system under analysis
...
Activity Pattern as part of a Story
...
Categories
You can explore the indicator repository in a structured way by using categories. These categories represent dimensions or aspects used to represent reality from a business point of view.
...