Appendix 3 - Example of Usability Testing
Clasification Society 2024 - Version 9.40
Statutory Documents - IMO Publications and Documents - Circulars - Maritime Safety Committee - MSC.1/Circular.1512 - Guideline on Software Quality Assurance and Human Centred-Design for E-Navigation - (13 July 2015) - Annex - Guideline on Software Quality Assurance and Human Centred-Design for E-Navigation - Appendix 3 - Example of Usability Testing

Appendix 3 - Example of Usability Testing

1 This appendix provides information on Usability Testing (UT) and uses ECDIS as a closely aligned example relevant to future e-navigation systems. This UT example aligns with Stage 4 of the HCD process for evaluating the performance of essential tasks by competent users. The selection of test participants is important and has a bearing on the quality of test results.

2 If tasks require operations based on navigational experience or knowledge, then appropriate participants should be selected. Tasks that are generally performed by less experienced or knowledgeable personnel should be similarly tested.

3 The UT activity involves the following steps:
  • .1 Planning;

  • .2 Preparation;

  • .3 Undertaking and controlling tests;

  • .4 Evaluation of results; and

  • .5 Use of feedback.

4 Only the steps related to planning and evaluation of results are explained in this appendix since these steps are the most important.

5 A UT plan should be developed by defining scenarios and identifying the most important or critical tasks that users must perform. Users and the test environment should also be identified.

6 A goal-based approach should be used when setting the tasks with the aim of facilitating flexible yet practical assessment of the target system.

7 The following steps can be part of the goal-based approach:
  • .1 definition of goals based on the context of use of the system, which may come from functions stipulated in internationally agreed performance standards;

  • .2 specify functional requirements or the criteria to be satisfied in order to conform to the goals, taking into account the relevant performance standards and user requirements;

  • .3 specify "usability" requirements that must be achieved during testing, based on the aspects of effectiveness, efficiency and satisfaction; and .4 prepare tests that will assist in verifying the extent to which the system conforms with the identified goals.

8 In the case of ECDIS goals could include "to plan and display the ship's route for the intended voyage and to plot and monitor positions throughout the voyage", based on SOLAS regulation V/19.2.1.4.

9 Similarly, functional requirements for ECDIS could be defined based on the IMO's ECDIS performance standard (resolution MSC.232(82)). The following example of ECDIS functional requirements relates to nautical data handling necessary for safe navigation, with the following sub-requirements:
  • .1 chart data handling (example: change display orientation, mode, etc.);

  • .2 own ship data handling (example: read position, speed, etc.); and

  • .3 tracked target (TT) and radar data handling (example: show TT symbols overlaid on ECDIS chart screen, etc.).

10 In the case of ECDIS, "usability" can be evaluated in terms of user effectiveness and efficiency for each of the tasks and overall satisfaction of the system (for example through subjective evaluation). As highlighted in table 1, measures of effectiveness relate the selected user goals to the accuracy and completeness with which these goals can be achieved. In this example, the achievement rate is used as a measure of "effectiveness". The four levels and their criteria are listed in table 1. Usability outcomes can be based on the "dialogue principles", as identified under ISO 9421-110, using UT methods based on ISO/TR16982. It is important that methods for evaluating usability are selected when devising the UT plan.

11 Scenarios and test tasks can also be created to satisfy the functional requirements. The following are examples of tasks for a basic display handling scenario:
  • Task 1: Adjust display modes and scale to meet operator's needs

  • Task 2: Obtain information about a lighthouse

  • Task 3: Measure the bearing and distance to a landmark

  • Task 4: Overlay a tracked target symbol and obtain information about the target

12 Criteria should be set to establish the degree to which tasks are achieved and also to capture user feedback on satisfaction with the operation of the system. Table 1 provides simple examples of achievement criteria for each task. Quantitative performance criteria such as time taken to complete tasks can also be included.

13 For the evaluation of system performance the level of task achievement can be useful (e.g. the time required to complete tasks). Questionnaires can assist with overall subjective system evaluation.

Table 1: Examples of achievement criteria for measures of effectiveness

Achievement level Criteria
Achieved 1 ✓ Participants understood the information correctly and operated properly with confidence.
✓ In case participants made some mistakes but noticed the mistakes immediately and achieved the goal smoothly, this should be considered "achieved smoothly".
2 ✓ Participants completed the task properly byt hemselves, even with some hesitation or confusion.
✓ In case participants took time to find the first action or to recover from errors but completed the task, this should be considered "achieved not smoothly".
Not achieved 3 ✓ Even if participants completed the task properly, it should be considered "not achieved with errors" if the participants could not understand the information correctly or if achievement took a large number of interactions.
4 ✓ Participants could not complete the task by themselves and needed suggestions from the moderator.

14 To satisfy quality management system requirements a UT report should be developed. ISO/IEC 25062 provides an example for a template that can be used for a UT report.

UT methods that can be applied at various stages in the life cycle (based on ISO/TR 16982)

Name of the method Direct involvement of users Short description of method Life cycle stage
Observation of users Y Collection of information in a precise and systematic way about the behaviour and the performance of users, in the context of specific tasks during user activity. 4
Performance-related measurements Y Collection of quantifiable performance measurements in order to understand the impacts of usability issues. 4
Critical incident analysis Y Systematic collection of specific events (positive or negative). 1
Questionnaires Y Indirect evaluation methods which gather users' opinions about the user interface in predefined questionnaires. 1 and 2
Interviews Y Similar to questionnaires but with greater flexibility involving face-to-face interaction with the interviewee. 2
Thinking aloud Y Involves having users continuously verbalize their ideas, beliefs, expectations, doubts, discoveries, etc. during their use of the system being tested. 3 and 4
Collaborative design and evaluation Y Methods which allow different types of participants (users, product developers and human factors specialists, etc.) to collaborate in the evaluation or design of systems. Any
Creativity methods Y/N Methods which involve the elicitation of new products and system features, usually extracted from group interactions. In the context of human-centred approaches, members of such groups are often users. 1 and 2
Document-based methods N Examination of existing documents by the usability specialist to form a professional judgement of the system. 1 and 2
Model-based approaches N Use of abstract representations of the evaluated product to allow the prediction of users' performance. 2 and 3
Expert evaluation N Evaluation based on the knowledge, expertise and practical experience in ergonomics of the usability specialist. Any
Automated evaluation N Algorithms focused on usability criteria or using ergonomic knowledge-based systems which diagnose the deficiencies of a product compared to pre-defined rules. 4
Simulation N Use of computer simulation modelling tools used for initial evaluations. 2 and 3

Copyright 2022 Clasifications Register Group Limited, International Maritime Organization, International Labour Organization or Maritime and Coastguard Agency. All rights reserved. Clasifications Register Group Limited, its affiliates and subsidiaries and their respective officers, employees or agents are, individually and collectively, referred to in this clause as 'Clasifications Register'. Clasifications Register assumes no responsibility and shall not be liable to any person for any loss, damage or expense caused by reliance on the information or advice in this document or howsoever provided, unless that person has signed a contract with the relevant Clasifications Register entity for the provision of this information or advice and in that case any responsibility or liability is exclusively on the terms and conditions set out in that contract.