| Literature DB >> 26354830 |
Harold Thimbleby1, Patrick Oladimeji2, Paul Cairns3.
Abstract
Number entry is a ubiquitous activity and is often performed in safety- and mission-critical procedures, such as healthcare, science, finance, aviation and in many other areas. We show that Monte Carlo methods can quickly and easily compare the reliability of different number entry systems. A surprising finding is that many common, widely used systems are defective, and induce unnecessary human error. We show that Monte Carlo methods enable designers to explore the implications of normal and unexpected operator behaviour, and to design systems to be more resilient to use error. We demonstrate novel designs with improved resilience, implying that the common problems identified and the errors they induce are avoidable.Entities:
Keywords: dependable systems; evaluating user interfaces; human error; number entry
Mesh:
Year: 2015 PMID: 26354830 PMCID: PMC4614478 DOI: 10.1098/rsif.2015.0685
Source DB: PubMed Journal: J R Soc Interface ISSN: 1742-5662 Impact factor: 4.118
Terminology used in this paper. The table makes clear that the designer has responsibility both at the blunt end and at the sharp end. (In a sense, the regulators, procurers and managers are all designers, since they specify or choose from a set of designs, which itself is a design activity.)
| blunt end | regulator | the organization that specifies high-level design rules and procedures (such as ISO 9241, ISO 19471, etc.) |
| designer | the person or persons who design, create or program the system. Designers are typically remote, as in manufacturers or their sub-contractors. In this paper, we are particularly concerned with designers of interactive systems | |
| system | the environment in which the operator works. The system includes the devices as well as the standard operating procedures, training and other people. (This paper is particularly concerned with the human interface of automated parts of the system.) | |
| procurer | people who choose designed (manufactured, programmed) products and assemble them into local systems | |
| manager or supervisor | people who are responsible for and devise rules within which operators work. Managers typically set requirements for designers | |
| team | in resilient organizations [ | |
| operator | the person ‘at the sharp end’ who is normally (but not always appropriately) considered responsible for outcomes | |
| device | the part of the system that physically causes the incident; for example, the operator may have pressed a button on the device, but the device actually caused the harm | |
| victim | the person or persons immediately suffering from the consequences of unmanaged or inadequately managed error | |
| second victim | operators or others who suffer indirectly, for instance from depression or inappropriate line management response [ |
Detecting error on the Apple iPhone calculator. We illustrate the problem with division by zero in the example where the operator intends to calculate but omits the 7 in error. Division by zero is detected, and is displayed, but the operator continues, and finally reaches a display that appears to show that 10 is the correct answer to the calculation (the correct answer is 11.4285714 to the precision of the iPhone). A more dependable calculator would display continuously until is pressed or the operator otherwise indicates they have recognized the error.
Figure 2.Risk ratio, the ratio of risk divided by vulnerability; compare visualization with figure 1, which is the same data. The distinctively defective designs A and B stand out. They counter-productively make risk ratio increasingly worse as the operator tries to reduce vulnerability: that is, however vigilant the operator (reducing their vulnerability, even to zero) the design defects ensure there is still residual risk (so the risk ratio goes to infinity). Put another way, even a perfect operator might be blamed for the problems these poor designs themselves are creating.
Summary of designs. ABCN are common, commercial designs; DEFG are proposals. Some unusual defective designs [17] are not considered here. Table 4 illustrates the designs on example keystroke errors and recoveries. Appendix C provides specifications of the designs, sufficient for them to be implemented.
| design | brief description |
|---|---|
| A | delete key ignores decimal points, and the design ignores multiple decimal points. Thus |
| B | delete key works correctly, but the design ignores multiple decimal points. Thus |
| C | correct design, exemplified by the Casio fx-85GT and many familiar keyboard-based applications on PCs, such as Microsoft Word |
| D | correct design, which also intercepts key bounce. A number entered with a repetition is blocked, and the operator has to re-enter it |
| E | correct design, which also checks ISMP recommendations. Invalid numbers are intercepted and the operator retypes them, possibly making further errors |
| F | correct design, which also enforces ISMP recommendations and ensures the number is correctly entered |
| G | correct design, which also enforces value to be within a factor of 5 of the intended number |
| N | no delete key. Noticed errors corrected by clearing and starting over |
| — | we know of no design that implements delete incorrectly but which implements decimal point correctly |
Delete key behaviour. Astonishingly, many numerical user interfaces always show a decimal point even if one has not been keyed (regardless of the delete key). For clarity, the right-hand column only shows a decimal point if it has been keyed and not deleted. It matters: if the display always shows a decimal point, if the next keystroke is a 0, it unpredictably leaves the number unchanged or multiplies it by 10. (This table was generated automatically by the Monte Carlo simulation program: hence what it describes is what was evaluated.)
Figure 1.Risk against vulnerability for different designs (figure 2). Risk increasing with vulnerability is expected (lower lines/gradients are safer), but different designs perform differently. Defective designs AB have potentially unacceptable risk even for ‘perfect’ (v = 0) operation; the alternative designs prove such risk is avoidable. The grey region covers designs that additionally cue the operator to manage error; all are safer than conventional designs.