Professional Documents
Culture Documents
Usability Analysis and Evaluation of Mobile ICT Systems
Usability Analysis and Evaluation of Mobile ICT Systems
ICT systems
Asbjørn Følstad, Petter Bae Brandtzæg and Jan Heim
SINTEF Telecom and Informatics, PO Box 124, N-0314 Oslo, Norway
asbjorn.folstad@informatics.sintef.no
Abstract
There is a need for new usability methods tailored to suit the development of mobile ICT.
In response to this, three methods that have been developed to a prototype level are
presented. The methods are designed to be cost-effective and provide information on (1)
user problems as experienced in users’ everyday lives, (2) contexts of use, and (3)
integration of technology in users’ patterns of activity. The methods have been tried out in
a small-scale evaluation in conjunction with the introduction of PDAs and mobile
telephones with PDAs in an organisation.
1.Introduction
A major challenge for future usability research is to develop new methods for usability
analysis and evaluation of ICT (Information Communication Technology) systems
[Abowd & Mynatt (2000); Monk (2000)]. The users of ICT systems are no longer only
business users, but belong to the general population. And the technology is becoming
increasingly mobile.
Existing usability methods and tools are suitable for testing stationary in-office ICT, but
do not cover the necessary span of mobile ICT use. Opposed to traditional desktop
computation, mobile ICT involves issues like use in multiple contexts, multiple users,
and continuous everyday use [Abowd & Mynatt (2000)]. Traditional usability work
typically concerns single users using one single technology to solve well-defined tasks in
an undisturbed environment. Solving tasks in this manner requires different strategies,
other sources of support, and is accompanied by problems other than the solving of tasks
in complex and shifting environments [Zsambok (1997)]. The traditional usability
framework needs to be supplemented by methods tailored to gain knowledge of user
problems in the ‘real world’, actual user contexts, and users’ integration of mobile ICT in
their everyday lives.
Mobile computing systems typically have short product realisation cycles. Consequently,
cost and time efficient methods and tools for usability analysis and evaluation are
required (Millen, 2000). What is needed are ‘quick and clean’ (as opposed to ‘quick and
dirty’) methods of usability analysis and evaluation [Wichansky (2000)]. One direction
with potential that has been identified is that of training users to report their own critical
incidents [Castillo, Hartson, & Hix (1998); Palen (2000)].
The three methods was used in a small-scale investigation, conducted in conjunction with
the introduction of PDAs (Palm IIIc) and mobile phones with PDAs (Nokia 9110
Communicator) in a department of a Norwegian research organisation.
2.Method development
The development and implementation of the methods will be presented in regard to:
• Generic description. Presentation of the method on a general level.
• Purpose. The kind of data that may be collected, and why these data are important.
• Ideal implementation. How the method ultimately may be implemented, given future
time and resources.
• Actual implementation. Implementation of the method at this stage of development.
• Validation. How the method has been-, or may be validated.
The purpose of Angry Face is to collect data on user problems as they appear in real life.
The method is designed to be low-cost, and should be easy to implement in a
development process providing that the product is sufficiently operative.
The prototype of Angry Face used in the investigation was far more mundane than the
ideal implementation. The PDAs were equipped with a screen-dump program, allowing
users to capture screen images associated with user problems by pushing a single button.
Comments were entered in the notebook function of the PDA.
Angry Face was validated through semi-structured interviews with the participants,
following their period as testers. Future validation may include the comparing of results
from Angry Face with results from more structured testing.
Among the above contextual variables, only one (‘time’) is easily logged automatically.
Location, activity, and social situation may only partially be automatically logged, and
this with an overwhelming amount of technical fuzz1.
The method of User Context Logging involves users giving ‘on the fly’ descriptions of
their contexts of use. The technology in question is equipped with a device for recording
users’ context descriptions and given to a predefined set of users. The users are urged to
report the context of every incident of use as they occur, for a given period of time.
An ideal implementation of the data collection tool for User Context Logging would be a
recording device activated by the onset of interaction with the technology, allowing the
user to register the necessary information in the simplest possible way. More down to
earth, the prototype data collection tool consisted of context templates stored in the
notebook function of the participants PDAs. The templates included slots for recording
1
Meaningful location terms (e.g. ‘at the office’, ‘on the bus’ etc) may to a certain degree be inferred from
geographical positioning data. Activities may be partially inferred from what functionality is used in the
mobile device. Social situation may be waverly induced from the combination of time and location.
time, task, location and social situation, and the participants entered their contextual
information by using the stylus. The collected data were validated through semi-
structured interviews.
2.3Patterns of Activity
Mobile technology is often introduced to support users in activities they are already
engaged in. As an example, a PDA may be introduced to help the user updating her
calendar; a task she may be used to do with the help of a Time Manager or PC software.
A beneficial introduction of the PDA requires that she changes her pattern of use in
regard to aids that have previously supported the activity, and that the new technology is
integrated in her pattern of activity.
The questionnaires are divided in parts corresponding to the activities that the mobile
technology is likely to support. The activities are identified through an initial task
analysis. For each identified activity, the respondent is asked to indicate
• How often the technology is used in conjunction with the activity
• How often other identified aids are used in conjunction with the activity, as compared
to before the new technology was introduced
• Satisfaction with the present way of the activity
The questionnaire includes a measure of general user satisfaction, providing ipsative data
on the perceived Usefulness, Control, Ease of Use, and Affect. The supplementing
interviews are conducted with a subset of the participants and should capture user activity
patterns not addressed in the questionnaires.
Data collected through Patterns of Activity may generate useful knowledge for deciding
on future developments of the technology, as well as for deciding the focus of attention of
further usability testing.
For the present investigation of PDAs, the tasks identified through the initial task
analyses included viewing and updating of calendar, contact lists, and task lists, in
addition to reading e-mail. The satisfaction measure consisted of 12 items corresponding
to the four factors mentioned above. The items were presented in four groups of three in a
way that each factor was represented with one item in three of the groups, and no two
groups had the same combination of factors. The participants were asked to rank the
items of each group in regard to how well they described the participants’ relation to the
technology. Factor scores were calculated as the sum of the rankings given to each of its
items.
3.1Angry Face
Angry Face was conducted with four novice users of Palm IIIc, over a period of two
weeks. The users totally reported 21 instances of usability problems. A sample of the
reported problems is given in Table 1.
• Cumbersome to give written input, in particular the Norwegian letters ‘æ’, ‘ø’, and ‘å’
• Cumbersome and difficult to register an event in the calendar for two consecutive days
• Difficult to understand how to categorise to-do lists
• Difficult to understand how to turn off alert function in calendar
• Lack date (in addition to time) on ‘Home’ screen
• Cumbersome to delete notes the hard way (through ‘details’), and difficult to understand how to do it
the easy way (through the menu or using graffiti)
• Dysfunctional when used as shopping-list.
Table 1: A sample of reported usability problems
Five of the reported problems were overlapping, and indicates that these may be
experienced by a majority of users. Nine of the reported problems were accompanied by
a relevant screen image, stored through the screen-dump function. Figure 2 shows the
screen-dump recorded by one user, indicating where she expected to find the function for
registering an event in the calendar for two consecutive days.
The real nature of the reported problems was analysed through subsequent interviews,
and by having an expert user interpret the reported incidents. Through this process,
incidents could be re-categorised. An example of this was that a function reported as
missing was only hard to understand.
In the interviews the participants reported that they were able to use Angry Face to
indicate usability problems they would otherwise forget. Even so, they claimed to be
hampered from giving a complete review of their user problems when using the current
version of Angry Face. Partly because its rudimentary implementation made the input
process cumbersome.
As for tasks, the data did not provide much in the line of surprises: Checking calendar is
the most frequently registered activity on the PDA. What is interesting however, is that
reported use in home and mobile contexts by far outnumbers use in a work setting. Also it
may seem that solitaire use is far more frequent than use in the company of others.
In subsequent interviews, the participants claimed that their assignment in User Context
Logging was easy to carry through. It was held to be easy to remember reporting the
instances of use, in particular since they were urged to glue a yellow post-it reminder on
the cover of the PDA. The participants also reported that the logging procedure did not
interfere with the task at hand.
3.3Patterns of activity
Patterns of activity were conducted with 13 users. Three of the participants were provided
Nokia 9110 while the 10 others got Palm IIIc. Each of the participants responded to a
questionnaire distributed 12 weeks upon receiving the mobile terminals. The participants
had also responded to a similar questionnaire distributed on the day of receiving the
terminals. Four of the respondents participated in interviews.
Regarding activities related to viewing and updating calendar, contact list and task list,
the users reported an increased use of these applications on their PC (MS Outlook), and a
decreased use of all other aids (like filofax, paper notebook etc.). This indicates a mutual
reinforcement between the use of PDAs and corresponding PC applications. Further, the
PDAs were more closely integrated in activities regarding the viewing and updating of
calendar, and less so in other activities. The participants, when asked before the
introduction of the technology, were better at predicting what effect they would get from
the calendar function of the PDAs than they were at predicting the effect of the other
functions. As for satisfaction, Palm IIIc users generally scored high on perceived usability
and low on affect, while users of Nokia 9110 scored high on affect.
Quite a few respondents reported that they were unable to express their individual use of
the technology through the questionnaire, and the interpretation of data generated through
the questionnaires relied heavily on the supplementary interviews.
4.Discussion
Angry Face and User Context Logging seem to be interesting methodological concepts.
They both enable the gathering of valuable information from real users in the world with
a minimum of fuzz and resources, and the information collected could not easily be
gathered through other means. When this is said, both methods have their limitations.
Both methods require that the technology to be evaluated is sufficiently developed to be
tried out on naive users without the benefit of supervision. Contextual User Logging also
has the problem that it may bias the participant, since the method may impose an
exaggerated consciousness regarding the use of the technology. Angry Face on the other
hand may suffer from the possible misconception that it is a regular usability evaluation,
only with the laboratory and mentor removed. This is not so. Data from Angry Face
address usability problems as experienced by the users in the real world, and should not
be expected to have the same focus or detail as data from the usability lab.
The success of Patterns of Activity relied partly on the authors’ ability to generate a
standardised form for collecting the necessary data associated with each participant’s
individual pattern of activity. This proved to be no easy task. The questionnaire generated
interesting data, but the interpretation of these required extensive analyses and
supplementary semi-structured interviews. A possible lesson to be learnt from this is that
since the use of mobile ICT in real life is complex partly unpredictable, structured
questionnaires alone may not be the optimal way to collect data on this.
4.3Future challenges
Future development of methods for analysis and evaluation of mobile ICT involves
several challenges. The methods presented in this paper may be interesting to pursue, but
require further efforts in the line of validation. Furthermore, nothing has yet been said
about the number of users required to get reliable results, or how interesting user groups
should be identified. A final point that needs to be addressed is that the development of
‘quick and clean’ methods alone is not enough to ensure the usability of future mobile
ICT. Future work should also focus on a tight integration of these usability methods in the
product development cycle.
References
Abowd, G. D., & Mynatt, E. D. (2000). Charting Past, Present, and Future Research in
Ubiquitous Computing. ACM Transactions on Computer-Human Interaction,
7(1), 29-58.
Castillo, J. C., Hartson, H. R., & Hix, D. (1998, 18-23 April). Remote usability
evaluation: Can users report their own critical incedents? Paper presented at the
CHI, Conference on Human Factors in Computing Systems.
Millen, D. R. (2000, 17-19 August). Rapid Ethnography: Time Deepening Strategies for
HCI Field Research. Paper presented at the The International DIS 2000
Conference on Designing Interactive Systems, New York.
Monk, A. (2000). User-Centred Design: The Home Use Challenge. In F. v. R. A. Sloane
(Ed.), Home informatics and telematics: Information technology and society (pp.
181-190). Boston: Kluwer Academic Press.
Palen, L., Salzman, M., Youngs, E. (2000, December 2 - 6). Going Wireless: Behavior &
Practice of New Mobile Phone Users. Paper presented at the CSCW '2000, ACM
Conference on Computer Supported Cooperative Work, Philadelphia,
Pennsylvania, USA.
Porteous, M., Kirakowsky, J., & Corbett, M. (1993). SUMI user handbook. Cork: Human
Factors Research Group, University College Cork.
Wichansky, A. M. (2000). Usability testing in 2000 and beyond. Ergonomics, 43(7), 998-
1006.
Zsambok, C. E. (1997). Naturalistic Decision Making: Where Are We Now? In C. E.
Zsambok, Klein, G. (Ed.), Naturalistic Decision Making (pp. 3-16). Mahwah, NJ:
Lawrence Erlbaum Associates.