Professional Documents
Culture Documents
Evaluating Software Engineering Methods and Tool Part 7: Planning Feature Analysis Evaluation 2
Evaluating Software Engineering Methods and Tool Part 7: Planning Feature Analysis Evaluation 2
Evaluating Software Engineering Methods and Tool Part 7: Planning Feature Analysis Evaluation 2
Evaluating Software Engineering Methods 1. the starting point of the evaluation: either the list of
candidate tools for evaluation, or how the candidate list
and Tool should be established (this is discussed in the next arti-
Part 7: Planning Feature Analysis cle);
Evaluation 2 2. the specific software engineering functions t h a t require
support;
Barbara Ann Kitehenham 3. expected scope of usage of the selected tool; (which type
of projects, which type of staff);
Department of Computer Science
4. how the results of the evaluation will be used and by
University of Keele
Staffs ST5 5BG whom.
England
Phone: +44 (0)1782 583079 B A S I S OF E V A L U A T I O N
Fax: +44 (0)1782 713082 The depth of investigation required for the candidates will be
barbara.kit chenham@cs .keele .ac .uk linked to the level of confidence required in the evaluations.
and It may vary as follows:
Lindsay J o n e s
Private Consultant • survey of publicity literature;
10 Clifford Road • attendance at organised demonstrations or trade fairs;
Macclesfield • study of published evaluations in the technical literature;
Cheshire SKll 8QF • study of technical documentation;
Phone: +44 (0)1625 425845 • interviews of current users;
• detailed evaluation of technical documentation plus "hands-
INTRODUCTION
on" usage of demonstration or evaluation versions of soft-
When planning a feature analysis based evaluation of a ware (if appropriate);
method or tool the following areas should be addressed and • practical experience of using the candidate method or tool
included in the evaluation project plan: in a wide range of situations.
• identify each distinct group in the potential user popula- 3. Formal Experiment;
tion; 4. Survey.
• eliciting the requirements of each user population group;
• identifying the features to be assessed and the method by Screening m o d e Approach
which they are scored; This is the cheapest but least reliable form of Feature Anal-
• organising the assessment whereby each of the ysis. The evaluation is performed by a single person based
method(s)/tool(s) is scored; on documentation only. It is the best approach for screening
• collation and analyse of the scores; a large number of methods/tools and is often used as a first
• preparation of the evaluation report. stage in a more complex evaluation exercise involving reduc-
ing a large number of possible methods/tool to a short-list of
The technology users one or two candidate method/tools that can be evaluated in
These are the people who will use the m e t h o d / t o o l chosen as more detail.
a result of the evaluation exercise. The user population for In this case the evaluator is responsible for:
method/tools can be a mixed population (i.e., be composed
of groups that have different requirements). The evaluator • identifying the candidate methods/tools;
needs to identify each group in a mixed user population and • devising the assessment criteria (i.e., the features to be as-
extract the requirements of each group. sessed and the judgement scales) with or without specific
Representatives of each group will need to: help from potential tools users;
• compiling information about the method/tools;
• discuss their technology requirements with the evaluator; • scoring each feature for each method/tool;
• review the features identified by the evaluator for complete- • analysing the scores;
ness with respect to their requirements • presenting a report on the evaluation.
• assess the importance and support required for each of the
different features PTo8
This is the fasted and least costly form of feature analysis.
The m e t h o d / t o o l assessors The number of different roles is minimised.
These are the people who will score each feature for the candi- Cons
date methods/tools. They are responsible for reviewing/using The entire evaluation is based on the subjective evaluation of
specific methods or tools in accordance with the defined eval- a single person (or team), so it may be biased.
uation procedure. Depending on the type of evaluation un- The evaluator may not be representative of potential users of
dertaken, the m e t h o d / t o o l assessor(s) can be: the method/tool.
The Case S t u d y Approach
• the evaluator (or evaluation team);
This is likely to be fairly inexpensive but provides only lim-
• representatives of the technology users;
ited confidence in the reliability of the evaluation. D E S M E T
• current users of the methods/tools. defines a case study evaluation exercise to be one where a
method or tool tried out on a real project. In the context of
EVALUATION PROCEDURES Feature Analysis, this requires a separation of roles between
You should also define how the evaluations are to be carried evaluation staff and software development staff who will act
out. This includes: as assessors:
• how candidate method/tools will be identified; • the evaluator (or evaluation teams) selects the meth-
• how the feature-sets and criteria of rejection or acceptabil- ods/tools to be evaluated;
ity will be derived; • the evaluator devises the assessment criteria and produce an
• how the results will be arrived at and presented; evaluation form working in conjunction with the potential
• how the evaluations are to be organised. users;
• the evaluator selects one or more pilot projects that will try-
It is necessary to decide the level of confidence required in out the methods/tools usually one m e t h o d / t o o l per pilot
the results of the evaluation and how that confidence is to project. The pilot projects should be selected to be typical
be attained. This is often linked to budgetary considerations, of the projects undertaken by the organisation
and to the criticality of the success of the selection. • a software developer (or a development team) in each pilot
project then tries out one of the methods/tools;
There is are four ways of organising evaluations:
• the software developer(s) who tried out the m e t h o d / t o o l
score each feature of the method/tool;
1. Screening mode;
• the evaluato~ analyses the scores and prepares and evalua-
2. Case Study; tion report.
ACM SIGSOFT Software Engineering Notes vol 22 no 4 July 1997 Page 23
curves and so assessing the methods/tools from different conjunction with the potential m e t h o d / t o o l users;
perspectives; • designing and running the survey. This includes:
• different assessors interpreting the evaluation scales differ-
ently - deciding on the type of survey (e.g., a postal survey v.
personal interview)
T h e Formal Experiment Approach -creating the survey documentation (e.g., question-
This is likely to give the most trustworthy results, but is also naire, instructions etc.)
often the most expensive approach to carry out. It involves - identifying who will be asked to participate in the sur-
a separation between evaluation staff and experimental sub- vey;
- running the survey according to the survey design;
jects:
• analysing the replies from the individuals surveyed and
• the evaluator (or evaluation teams) selects the meth- preparing an evaluation report. In the case of a postal sur-
ods/tools to be evaluated; vey, this needs to include an assessment of the response rate
• the evaluator needs to devise the assessment criteria and and an evaluation of the representativeness of the respon-
produce an evaluation form working in conjunction with dents as well as collating the completed evaluation forms.
the potential users;
• the evaluator then needs to plan and run the evaluation Pro8
experiment. The procedures for running the experiment are You can obtain information about the methods/tools on real-
the same as those discussed by Shari Lawrence Pfleeger in lifeprojects.
her series of articles about experimental design in SIGSOFT A survey can provide results in shorter timescales than a case
Notes [1]. They include among other things: study.
A well-designed survey where you understand the population,
- identifying an appropriate experimental design; sample that population at random and obtain replies from all
- identifying subjects (preferably by a random selection
the sampled individuals has all the rigour and validity of a
of potential users from each class of user). Note the
formal experiment.
experimental subjects should not include users who
A survey is less human intensive than a formal experiment.
participated in the development of the feature list;
- identifying the tasks to be performed by the experi- Cons
mental subjects using the method/tool; This approach is only viable if you can identify individuals
- organising any training for experimental subjects; with experience of the methods/tools you want to evaluate.
- running the required experimental trials according to You cannot control the experience, background and capability
the experimental plan of the assessors in the same way that you can in a formal
- the experimental subjects will need to try out the can- experiment.
didate method(s)/tool(s) according to the experimen- With a postal survey results can be biased because:
tal instructions and score each feature;
• the evaluator analyses the forms and prepares and evalua- • the respondents are not typical software developers or
tion report. managers. For example people who have time to reply
to a survey may be better organised than most man-
Pros agers/developers.
A well-designed experiment minimises the effect of individual • the respondents may have an axe to grind. For example,
assessor differences. they may have had a m e t h o d / t o o l imposed on them and
Cons want an opportunity to gripe about it.
It is difficult to ensure that the tasks performed in an exper- • you may get patchy responses (i.e., lots of information
imental situation properly reflect "real-life" tasks, so there is about some of the methods/tool of interest but none about
often a question as to whether results will scale-up. others).
A C M SIGSOFT Software Engineering Notes vol 22 no 4 July 1997 Page 24