Evaluating The Educational Impact of Visualization
Evaluating The Educational Impact of Visualization
124
• Visualization Designers specify the mapping from the • 90%: time it takes to develop visualizations
abstract concepts to their visual representation by cre-
ating specific visualizations – either by using visualiza- • 83%: lack of effective development tools
tion meta tools or by other more direct methods. • 79%: time it takes to adapt visualizations to teaching
• Instructors incorporate visualizations into their teach- approach and/or course content
ing materials and methodology.
Clearly, the amount of time involved in learning how to
• Learners view and hopefully interact with the visual- use visualization tools and in developing demonstrations and
ization in one or more of the engagement levels de- interactive lab materials discourages the use of visualiza-
scribed in [35]. tion. Our thesis is that these impediments can be overcome
by providing high quality support materials for instructors.
An individual may take on many of these roles. For exam- The availability of good support materials will lead to in-
ple, instructors will often also act as visualization designers creased instructor satisfaction, which consequently will lead
to generate their own specialized visualizations. to more widespread usage of visualization.
Working from this basic thesis, Sections 2 and 3 will offer
guidance for enhancing the effective deployment of visual-
Vis. Tool Developer ization tools in the classroom. Since impacting CS edu-
cation requires both widespread use and improved student
outcomes, such guidance will be broken down along these
Concept V Tool two lines. Section 2 will focus on the instructor satisfaction
issues, and Section 3 will cover measurement of student out-
comes. Section 2 will be of particular interest to developers
and designers who seek feedback on how effective their tool
Visualization Designer is in teaching situations.
In future years, the combined results emanating from the
Visualization evaluation techniques described will provide a more informa-
tive measure of the two aspects that combine to influence
educational impact.
125
• Adapt and integrate into a course installing, adapting and integrating into a course, and main-
taining and upgrading are moot. Since the visualization ap-
• Teach students to use visualizations plets are part of the fabric of the hypertextbook, they will
be used quite naturally by instructors and students alike.
• Maintain and upgrade The applets themselves must be designed so that they are
easy to learn by both instructor and students – issues we
Exacerbating this situation is the fact that this effort may discuss below. More can be read about hypertextbooks in
all be done for just a couple of lectures and might need to [6, 40, 41, 27].
be repeated for each concept to be visualized. Indeed, of the Between these two extremes lie a variety of other possibil-
nine top impediments cited by instructors in the survey, six ities for visualization systems and visualization tool devel-
were time issues. opment that address many of the impediments listed earlier.
A second major impediment identified in the literature is It should be noted that these suggestions are not mutually
platform dependence. If visualization systems are designed exclusive. Many can be combined to address more of the
to run on a particular platform (for example, a PC with impediments. We provide some suggestions next.
Windows), it precludes their use on another system (for ex-
ample, a PC with Linux). Indeed, platform dependence has Design for platform independence. This will, obviously,
many more subtle nuances that a visualization tool designer eliminate the impediment of platform dependence. Plat-
must address (for example, version of operating system or form independence is an elusive and likely unattain-
browser used). able goal in its ideal sense, but there are some choices
A third major impediment highlighted in the references that are better than others. For example, designing
is course integration issues. How easy is it to incorporate systems for the Web (or, more precisely, the Java Vir-
a visualization into the fabric of a course and to adapt it tual Machine) is one possibility. An alternative is to
to the way the concepts are presented in the classroom and ensure that a visualization system runs on all of the
in the accompanying textbooks or other class resources? If major platforms likely to be available in academic set-
the visualization does not integrate well into a course, it will tings around the world. Some visualization systems
most likely not be used. designed in this manner have come to untimely ends.
Notice that although the course integration issue is high- Those likely to be successful are those that are based
lighted separately in the literature, it is actually captured on widely accepted and standardized software tools
in the time impediment list as well, as adaptation and in- that themselves have been ported to many platforms,
tegration of a visualization system into a course are rightly such as OpenGL for graphics.
also identified as substantial time sinks.
Capture larger concepts. This will ameliorate the time
2.2 Advice impediments of searching for, downloading, installing,
Visualization systems run a wide gamut. At the two ex- and learning a new tool for each new concept. A vi-
tremes are: sualization system is likely to be more widely used if
it allows for visualizations of an entire “module” of
• Standalone, single purpose, undocumented, platform related concepts, such as all of searching and sorting
dependent visualization systems that do not engage rather than just a single algorithm or two. Further-
the student beyond passive viewing more, the treatment of each visualized concept will
have a similar “look and feel” that allows instructors
• Complete teaching and learning resources that incor- and students to focus on learning concepts instead of
porate visualizations seamlessly and become an inte- learning how to use a tool. For example, systems like
gral resource used for an entire course Alice [9] provide a resource around which a course can
be designed and that can be used for most, if not all,
Many visualizations have fallen into the first category and of a course.
are precisely the ones that lead to instructor frustration as
measured in the survey. They should be avoided by design- Map to existing teaching and learning resources. Ex-
ers of such systems at all costs. Systems like this, while they tending the previous observation, providing a visu-
may be usable by the designer in a local course for a specific alization package that corresponds to an existing re-
purpose, are sure to not gain widespread use. source (for example, a textbook) will make the pack-
Hypertextbooks are an example of the other end of the age more appealing to users of that textbook. If done
spectrum. They are envisioned to be complete teaching and well, the seamless integration of the book with the vi-
learning resources that complement—or indeed supplant— sualization system will also eliminate most of the time
traditional course teaching and learning resources (for ex- required to adapt and integrate a visualization sys-
ample, textbooks). Hypertextbooks may include standard tem into a particular course. It has the drawback, of
text, illustrations, pictures, video clips (for example, Quick- course, of being primarily useful to those who choose
Time movies), audio, and various paths through the mate- to use that textbook. Integrating the system with text
rial based on different learning needs. Most importantly, materials can further improve the adaption of the sys-
they would also include embedded visualization applets of tem. For example, Brown et al. state that a large part
key concepts and models that engage the learner in active of the success of the BALSA system was due to its
learning of the subject. In this case, since the hypertext- tight integration with the teaching materials [?].
book is the course teaching and learning resource, and since
the hypertextbook runs in standard browsers (and likely dis- Design for flexibility. One can make a visualization tool
tributed on CD or DVD), the issues of finding, downloading, so flexible that it can easily be adapted to many differ-
126
ent presentation styles and course resources (for exam- in a repository, visualization tool designers publicize
ple, textbooks). This is a worthy goal but, of course, their work in venues such as the annual SIGCSE sym-
more difficult to achieve. For example, there are many posium (through papers or posters), in general edu-
different implementation nuances that affect how a cational journals such as Inroads and in area-specific
particular algorithm (for example, a Quicksort algo- journals (for example, if the tool visualizes aspects of
rithm) actually runs. Virtually every textbook pro- the theory of computing, in relevant theory journals),
vides a different version. Adapting a given visualiza- and any other appropriate media.
tion of an algorithm to the precise way in which the
algorithm is presented in the textbook may be diffi- The above advice is not meant to be comprehensive, but
cult or time-consuming. The discrepancy in content rather illustrative. Once the impediments are known and
may lead to confusion on the part of students. Making some examples of ways to surmount these impediments have
adaptation easy due to system flexibility can therefore been discussed, we are certain that visualization design-
play a key role in a successful adoption of a visualiza- ers will become more adept at providing the community
tion system. with systems that address the issues that have slowed the
widespread use of the tools and thus promote the use of
Provide comprehensive, integrated support. To elim- visualizations in a positive way throughout the curriculum.
inate the frustrating time impediments of learning a
visualization resource and teaching students to use it, 2.3 Disseminating Visualization Tools
a comprehensive support structure should be part of How a visualization tool designer disseminates a system
the tool. This support should include a very care- plays an important role in how widely the system will be
fully designed GUI that is novice-friendly and easily used. In this section, we present a suggested outline of a
learned. Documentation on the use of the tool as well standard Web site for this purpose. The site should make it
as tutorials illustrating its use should also be part of easy for Web surfers to find the tool, learn about it, down-
the software. The entire support structure should be load it, and install it. The site should further provide a
refined based on feedback from the user community, mechanism for obtaining feedback from those who choose
both learners and instructors. to download the system. Feedback is used to measure the
level of satisfaction of those who use it. In what follows,
Develop a supporting Web site. A carefully designed the word “tool” may mean either software for designing vi-
Web site for a visualization system can do much to sualizations (examples: Animal, Alice, Matrix, . . .), or a
address the time impediments that frustrate instruc- collection of pre-prepared visualizations (examples: Quick-
tors. Choose a clever name for the visualization tool Time movies or Java applets).
that is catchy, informative, and will be easily found
in a Web search. A Web site that provides a place The portal. Acknowledging principles of good Web page
to download the visualization software should include design (see, for example, [36, 37]), we recommend that
many other things as well, such as sample lectures, ex- the entry page, or portal, to the Web site be attrac-
ercises, and PDF documents for hard-copy instructions tively designed and that it provide clear information
that can be used by students who are learning to use describing:
the system. Community forums and FAQs that allow
users of the tool to interact with other instructors who • The name of the tool
are using the system also make adoption of the system • Author contact information
more likely. A good Web site is, in fact, such an im-
– Names
portant aspect of this discussion that we elaborate on
it in sections 2.3 and 2.4. – E-mail addresses
– Institutions
Register the tool in repositories. To help overcome the • A short, clear description of the tool that will let
impediment of time to find visualizations on the Web, visitors know whether the system is of interest
the systems should be registered in relevant reposito- (so that they can decide whether to investigate
ries. Unfortunately, there is currently no single author- further or abandon this particular search)
itative repository for registering visualization tools.
However, there are various competing repositories or • Other pages that provide in-depth details for in-
link collections that can be accessed when searching terested visitors, including:
for visualization tools or content. These include the – A detailed description of the tool
Complete Collection of Algorithm Animations [7], the – Documentation on the use of the tool
forthcoming Algorithm Animation Repository [10], CI- – Supporting materials for the instructor
TIDEL [8], a prototypical repository of visualization – Evaluation instruments and results of prior
resources [42] as well as the “SIGCSE Education Links” evaluations
on the ACM SIGCSE Web site [1].
– Download information
Publicize. It is probably safe to say that most instructors
do not actively seek visualization systems to use in We elaborate further on the links listed above.
their courses. This could be because they are satisfied The description page The detailed description page should
with their way of teaching, they are unaware of visu- provide:
alizations in general, or they have tried visualizations
in their courses before without success. It is manda- • A comprehensive description of the tool and its
tory that, in addition to simply registering their work use
127
• The levels of targeted learners • The background of the person downloading the
• References to the algorithms or concepts being vi- software (instructor, student, or other)
sualized so that instructors can determine whether • The purpose of the download (for use in a course,
the visualization integrates with their way of teach- independent learning, simple inquisitiveness)
ing • The person’s willingness to receive future e-mails
• Further links to any existing publications describ- about the use of the tool (along with a clear state-
ing the tool and its use ment that contact information will not be used for
other purposes)
The documentation page. The documentation page should
provide: The information on this page should point to the eval-
uation instruments (see the next section) to provide
• Documentation on how to use and install the tool the person with a clear idea of the kinds of informa-
tion that might be requested in the future. It is also
• A statement about whether the tool is still main-
always good to provide a free-response box to allow
tained
the person to provide additional comments as well.
• A printable tutorial for students to use when learn-
ing to use the tool 2.4 Evaluation
In this section, we propose sample items to include in eval-
The support page. The support page should provide, where uation instruments intended to measure instructor and stu-
possible and appropriate: dent satisfaction with the tool. The purpose is to provide
feedback to the tool designer that will allow modification
• Suggestions on the use of the tool of the tool to improve instructor and student satisfaction.
• Lecture support material (such as PowerPoint slides) These instruments will be filled out by instructors and stu-
dents after the tool has been used. Thus, the evaluator
• Sample exams and quizzes
needs to maintain contact with instructors after they have
• A set of exercises for use with the tool first downloaded the system.
The evaluation page. The purpose of the evaluation page 2.4.1 Evaluation instruments
is twofold: (1) to provide feedback to the tool designer The evaluation instruments should be extremely easy to
on the level of instructor and student satisfaction with fill out so that as great a return as possible is obtained.
regard to use of the tool, and (2) to provide visitors They should be administered online and be automatically
with results of earlier evaluations. As the tool matures, accumulated in a database belonging to the tool designer.
it may even be possible to include formal studies on Requests for written answers should be carefully thought out
the effects of the tool on student learning, but in this and not used excessively so as not to be too time-consuming
section we just provide suggestions for measuring in- for those filling out the form.
structor and student satisfaction with an eye towards The evaluation instrument for instructors should include
making the tool more enticing. questions that obtain:
Thus, what we suggest is that this page include links
to online statistics-gathering instruments for: • The instructor name and contact information
• The content and level of the course in which the tool
• Obtaining feedback from instructors who down- was used
load and use the system
• Obtaining feedback from students who use the • Course enrollment
system • Assumed prerequisites
This information is so vital to the ongoing success of Scaled questions using the traditional Likert scale with
the tool that we devote Section 2.4 to an elaboration values such as strongly disagree, disagree, neutral, agree,
of possible instruments for these items. strongly agree include:
The download page. This is a crucial page. This page • The tool is easy to obtain
should not only supply an easy way to download the
tool, but it should be designed to elicit information • The tool is easy to install
from those who download. The first part of this page
• The tool is easy for an instructor to use
should be a genuine plea to the person downloading
the tool that asks for help with the project. It should • The tool is easy to show and teach to students
be clearly explained that this is an academic (rather
than commercial) project and that the continued suc- • The tool is easy for students to learn
cess and improvement of the tool depends crucially
• The tool works reliably
on voluntary feedback from the user community. The
download process should thus have a mandatory reg- • The tool contributes to good learning outcomes
istration procedure that requests information about:
Multiple-choice questions which don’t fit on a scale like
• E-mail and other contact information the one above include:
128
• How did you learn about the tool (private communica- Since we cannot directly measure the learning process,
tion, from a conference, from a Web site, from a Web the focus of this section is on measuring learning outcomes
search, in a book, . . .)? and how student attitudes are affected by the use of visu-
alization techniques. We provide suggestions for evaluating
• How often did you use the tool in the course (one or student learning outcomes and attitudes when visualization
two times, regularly for part of the course, regularly tools are employed. We also offer guidance for the visualiza-
throughout the course, . . .)? tion tool designer, the visualization designer, and instructors
• In what context did you use the tool (classroom pre- who desire to study the effects of using visualization tools
sentation, closed lab exercises, open lab assignments, in their courses. Throughout, we attempt to describe ex-
. . .)? periments that support the selection of tools for everyday
teaching situations (hereafter TS) as well as experiments
• In this context, was the use of the tool (required, op- that are designed for more formal education research pur-
tional)? poses (hereafter RP).
• How did students interact with the tool (watched, an- 3.1 Different forms of evaluation
swered questions, provided input or parameters, de- Summative evaluations are those that occur after stu-
signed their own visualizations, gave a presentation in dents’ use of the tool is completed for the study in ques-
which the tool played a role, . . .)? tion. Formative evaluations occur during the study and are
meant to determine whether project-related objectives are
A student evaluation instrument would ask different ques- being met as the study progresses.
tions. It is hoped that the tool designer, upon contacting
an instructor who downloaded the tool, could encourage the 3.1.1 Formative evaluations
instructor to require students to fill out this separate stu-
Formative evaluations typically involve qualitative assess-
dent evaluation instrument on the tool Web site. Scaled
ment. For more details about formative evaluations, see [17,
questions for measuring student satisfaction include:
16]. This section discusses formative evaluations in general,
but focuses on those formative evaluations we believe are
• I enjoy using the tool
particularly well-suited for visualization.
• I feel I understand the concept better when using the
tool 1. Student attention to a visualization
• The tool is easy to use By studying in depth how the student uses the specific
visualization, we can determine if the student is using
• The tool works reliably for me it in ways the instructor intended. Possible implemen-
tation ideas include observations (where the student is
Multiple-choice questions for students include: watched, either directly or through a one-way mirror,
performing a specified task) and eye-tracking cameras
• For any given assignment, how much time did you (where it can be determined on which parts of the vi-
spend with the tool on average (about 5 minutes, about sualization tool the student is focusing).
10 minutes, about 15 minutes, about 30 minutes, about
an hour, more than an hour, . . .)? 2. Time-on-task
• How many exercises or assignments did you do with This evaluation may be thought of as both formative
this tool? and summative. The purpose is to keep track of how
long the student spends working with the visualization
• How did you use the tool (watched in class, used in tool in an assignment. The simplest approach is to
lab, used in university work area, used on my own have the tool record the time at startup and when it is
computer, . . .) shut down. A more detailed implementation involves
the generation of a log by the visualization tool of all
The evaluation instrument, whether for instructor or stu- student interactions with the tool. While generating
dent, should always provide an open field for additional com- a log is more difficult to implement (and analyze), it
ments at the end. Further, since an instructor may use does allow for a more detailed analysis of interaction of
visualizations in different ways in different courses, the in- the student with the tool. This approach may be used
structor should be encouraged to complete an evaluation for in a formative manner. For example, a log would allow
each course. We assume that instructors who were willing the instructor to see if the student is having difficul-
to fill out the instruments would not mind being contacted ties using the tool. Such a formative evaluation allows
again if clarification of responses is needed. the instructor to adjust lecture materials, or perhaps
provide a modified tutorial on the visualization tool’s
3. EVALUATION OF LEARNER OUTCOMES use.
Ultimately, the application of visualization techniques in This is especially true in the case of virtual learning
the teaching and learning process will become widespread environments, where the learner may lack direct feed-
only if instructors have concrete evidence that student per- back. Here monitoring the overall student performance
formance improves and/or that student interest and moti- in a time-on-task sense is a bit more easily adopted.
vation in the subject are enhanced when visualizations are The feature was recently incorporated, for example,
used. into the electronic text book illustrated in [27].
129
Software that allows an instructor to watch the screen subsection, the focus here is on those summative evaluations
of a student at a remote workstation may also be used most applicable to visualization.
when the instructor wants to focus on one particular
student’s interaction with the visualization tool. 1. Analysis of learner understanding using mental models
3. Intermediate student feedback
Evaluation of student outcomes using mental models
Anonymous surveys may be employed during the ac-
is an analysis technique used primarily in formal ed-
tual usage of the visualization tool in a class to get
ucation research studies. Since we have no analysis
students’ impressions of the tool. Students may be
method based on mental models for visualizations, we
asked to solve a problem to gauge their comprehen-
describe an analysis method that has been developed
sion, how much time they are spending using the tool,
for mental models used by programmers. It would be
how much time they are spending on the course with-
important to try to find out what is good comprehen-
out involving the tool, their opinion of the materials,
sion of algorithms (instead of programming in general).
the lecture, the visualization tool itself, and so forth.
By knowing this we would have a solid basis for the
One way to improve the outcome of the survey is to
evaluation of learning outcomes. One possibility for
compare the survey data to actual facts known from
determining the quality of algorithm comprehension is
the tool (from a log automatically generated by the vi-
to use similar study techniques as Pennington [38] did
sualization tool, as discussed in item 2 above). Thus,
with programming.
the impression expressed by the students can be “ver-
ified” by comparing their opinions with actual test re- Pennington [38] studied comprehension strategies in
sults of the performance. These are not absolute values programming. Good [18] continued Pennington’s work
but merely comparable with each other, for example and developed a classification for analyzing novice pro-
to determine the learner’s opinion on the tools versus gram comprehension. Her classification has been ap-
other resources used in the course. plied by Good and Brna [19] and Sajaniemi and Kuit-
Another approach of interest in obtaining intermediate tinen [44]. These studies describe the method and the
student feedback is the use of Classroom Assessment procedure for analyzing the mental models of novice
Techniques (CATs) [45]. While not directly related programmers. Thus, they can be used as examples
to the use of visualization tools in class per se, CATs of how to use mental models for evaluating learning
appear to be a promising mechanism for obtaining on- outcomes.
going student feedback, as well as helping to contex- Pennington had 40 expert programmers that were the
tualize course content for the students. top and bottom quartile comprehenders (20 each) from
her previous study (see the details in [38]). The pro-
4. Peer reviews of curricular materials
grammers had to make a modification to a program
Peer review is a widely accepted technique for exam- normally maintained by another programmer. After
ining the content, construct, and criterion validity of studying the program they were asked to write a sum-
instructional materials [3]. This is particularly valu- mary of the program, respond to a set of comprehen-
able when the instructors’ materials are evaluated by sion questions and explain their answers. Then they
the visualization tool creator, and by other visualiza- were asked to make the modification and, after that, to
tion experts in the field, as well as by “expert teachers” summarize the program and respond to another set of
in the specific content area where the visualization is comprehension questions. Answers to the questions as
being used. The instructor may receive valuable feed- well as the program summaries were the basis for the
back ensuring that proposed use of the visualization measurement of programmers’ comprehension which
tool in class is pedagogically sound. then could be classified as good or poor comprehen-
sion based on the subjects’ quartile.
5. Student interviews
Interviewing a random subset of the students who have The basic idea of Good’s analysis method is that com-
used the specific visualization tool and its associated prehension can be studied using program summaries
materials can provide valuable feedback. By focusing which are analyzed using two different methods: in-
on the students’ experiences with the tool, the instruc- formation type analysis and object description anal-
tor can gain a detailed understanding of the students’ ysis. According to Good and Brna [19], the infor-
comprehension, student attitudes towards the visual- mation types classification is used to code summary
ization tool and the subject being studied, students’ statements on the basis of the information types they
comfort with using the tool, students’ suggestion for contain. The types include eleven categories: func-
tool/lecture improvements, and so forth. Interviewing tion, actions, operations, state-high, state-low, data,
may be done in an individual and/or a small group control, elaborate, meta, unclear, and incomplete. For
format. a definition of these terms, refer to [19]. The object
classification looks at the way in which objects are de-
3.1.2 Summative evaluations scribed. This classification has seven categories: pro-
gram only, program, program – real-world, program –
Summative evaluations summarize the effectiveness and/or
domain, domain, indirect reference, and unclear.
results of the study after students have completed their use
of the tool. Generally quantitative methods are used in sum- The information types described above help to divide
mative evaluations. More details about summative evalua- information into low level information and high level
tions may be found in [16]. Like the formative evaluation information. High level types are function, action,
130
state-high, and data, while low level types include op- To help determine content mastery where a visualiza-
eration, state-low, and control. Elaborate, meta, and tion tool is used, pre- and post-tests are particularly
unclear cannot be classified as either high or low level effective. Typically, the same test is used for both
types. tests. The purpose of the pre-test is to determine the
There are two ways to use these analytical methods: level of prior knowledge the student has. It is of ut-
students can be asked to write program summaries or most importance to ask the “right” questions. If in
they can be asked to answer very carefully designed the case of a study designed for TS it is not possible
questions. Program summaries are difficult to analyze to use identical exams for the pre- and post-tests, the
and therefore these kinds of tasks may not be suitable questions on the pre-test may be a proper subset of
in TS. For RP, the analysis method is very useful for those on the post-test.
finding what kind of effect different teaching methods
may have on students’ mental models. Asking ques- 4. Attitude survey
tions is more suitable for TS because the questions Attitude surveys are generally given before and after
can be designed to make the analysis of the answers students’ use of a visualization tool. They are used to
easy. determine changes in students’ attitudes, confidence,
Next we will discuss the design of the questions. Both motivation, and so forth as a result of their experi-
high and low levels of information types should be ence with the visualization. The two most widely ac-
used when asking questions, although the high level cepted surveys for determining student attitudes to-
questions measure deeper understanding and therefore wards computers and computer science, Francis [15],
should be more valuable. It is possible that a student and Loyd and Gressard [33], are somewhat dated. These
whose comprehension is on a high level cannot cor- survey instruments are most appropriate for use in in-
rectly answer the questions on a low level. This is troductory classes. There is certainly a need for the
not unusual and not necessarily bad, since high level development of a newer, more relevant, survey instru-
comprehension replaces low level comprehension when ment to be tested and validated. Instructors of upper-
learning is deep enough [38]. level courses will need to create their own survey in-
struments. The difficulty is that the survey instrument
Object descriptions enable us to ask three kinds of itself will not be experimentally validated. A good
questions: questions about the program code itself, attitude survey provides valuable feedback about the
questions about the domain of the program, and cross- students’ impression of the visualization tool, and is
reference questions in which the student needs to com- typically administered as a pre- and post-test.
bine both the program code and the domain. Those
students who can answer the cross-reference questions
5. Retention and attraction
have a deep comprehension of the program while the
others have only a surface level comprehension (that Particularly for visualization tools used in lower-level
is, they probably recognize the code and/or the do- courses, it is interesting to monitor the change in stu-
main but may not understand the connection between dent retention in the computer science major and mi-
them). From the evaluation point of view the best sit- nor. For those tools that impact courses for non-
uation is when a student is able to answer all three majors, it is important to examine whether or not the
kinds of questions. The next best is that a student use of the visualization helps to encourage students to
can answer cross-reference questions (deep level) even become computer science majors or minors. Retention
though program or domain questions are not answered. and attraction statistics are often combined with stu-
dent attitude surveys to help gauge student reaction
2. Analysis of learner understanding using levels in Bloom’s to their experiences with the visualization tool.
taxonomy
Bloom’s taxonomy [5] is another common tool used 6. Grades
in formal education research studies. The taxonomy Student grades are a measure of student success in
offers a way to systematically design test tasks for a course. While they are generally not as useful as
analyzing learner understanding. Six comprehension pre- and post-tests to gauge student content mastery,
levels are provided, each of which characterizes an as- grades are easy to collect and analyze.
pect of the learning process. The idea is to reduce the
qualitative improvement in learning to discrete levels 7. Time-on-task
of comprehension and moreover, to measurable scores.
The six comprehension levels are knowledge, compre- While described in the formative subsection, time-on-
hension, application, analysis, synthesis, and evalua- task may also be used as a summative measure. As
tion. The experiment can be designed to measure the an example, consider the following. A student’s use
learner outcome in each of these levels starting from of a visualization tool as part of an assignment can be
the knowledge and ending up with evaluation. timed. For example, if a student does an assignment in
an environment where no clock is available, the student
The 2002 Working Group report [35] provides more
can be asked how much time elapsed while working on
detailed examples of the different Bloom levels. The
the assignment. If the student thinks that a smaller
context used in that report is a course in data struc-
amount of time passed than actually did, this indicates
tures and algorithms.
that student’s interest in using the visualization tool.
3. Pre- and post-content tests See [21] or [4] for more details.
131
3.1.3 Which of these evaluation methods should be amples of such models include Kolb’s experiential learn-
used? ing [23] and the Felder-Silverman learning model [13].
This is, in general, a difficult question to answer. The In the following, we discuss the latter model in some
first question to answer is whether the particular evaluation more detail since it seems particularly suited to science
of a study is primarily for teaching or research purposes. education.
A research study typically requires significantly more as- Felder and Silverman identify four dimensions of stu-
sessment than a study primarily interested in improving a dent behaviors, each of which has two extremes.
teaching situation. Additionally, the specific visualization
Sensory vs. Intuitive – what type of information does
tool and its associated visualizations are key determinants
the student preferentially perceive. Sensing learners
of the specific evaluation strategies to be used. For several
prefer collecting information by observation. They of-
sample studies, see the case studies described in Section 4
ten like facts, working with details, and memorizing
of this paper.
data. They also prefer a practical approach, experi-
3.2 Covariant factors mentation and solving problem using standard meth-
ods. Intuitive learners like conceptual thinking, such
If we are going to set up a study where we compare differ- as theories and principles, and grasping new concepts.
ent tools and assignments, we should recognize that learning
style may affect the results. Consider as a trivial example a Visual vs. Verbal – how is the sensory information
study with only two students, A and B, in which the results most effectively perceived. Visual learners better adopt
of two assignments given to both are compared. Suppose visual information and verbal learners prefer informa-
that in the first assignment, A gets grade of 1 out of 5 and tion in written or spoken form.
B gets grade of 5 out of 5. In the second assignment, sup- Active vs. Reflective – how does the student prefer to
pose A gets 5 out of 5 and B gets 1 out of 5. Obviously, the process information. Active learners learn by trying
average results are the same, and we could claim that we things out and prefer interaction. Reflective learners
found no difference in results for the assignments. Suppose, prefer examining and manipulating information intro-
however, that A is a highly verbal and B a highly visual spectively.
learner and that the assignments were visual and verbal, re-
Sequential vs. Global – how does the student progress
spectively. By considering the learning style as a covariant
towards understanding. Sequential learners like to pro-
factor in the study, we would observe a major change in
ceed linearly with small steps using the given instruc-
students’ performance.
tions. Global learners like to get a holistic view of
Several covariant factors are listed in last year’s working
acquired knowledge.
group report [35]. Here we discuss in detail some factors as
they pertain to evaluation. Initially, Felder and Silverman also had a dimension
inductive vs. deductive learners, but they decided to
1. Learning styles drop it since the model should not promote the exist-
ing conflict between deductive teachers and inductive
Students have different styles of taking in and process- learners [12].
ing information. Thus, some students are more com-
fortable with reading text while others prefer grasp- Obviously, all these axes are continuous, that is, each
ing information from figures and visualizations. Other student lies somewhere between the extremes. Their
students like to process information actively through orientation can be evaluated by exposing the student
experimentation and observations while others prefer to a simple questionnaire, such as presented in [22].
processing information in their minds through intu- However, we note that orientation may change over
ition and reflection. time and may depend on the context in which students
are working.
It is important that we as teachers recognize these dif-
ferences, since they very much affect how our students The general goal is that our students should extend
feel about studying and how well they succeed at var- their skills of adopting and processing information with-
ious activities and tasks we design. A conflict often in all four axes. To accomplish this goal, the teacher
occurring in undergraduate education is that teach- may push the change by setting up activities that train
ers explain the topic deductively, first concentrating the weaker side of student’s behavior.
on principles and theories, and then proceed to ex- Next we present a few examples of assignments that
plain how these theories and principles are applied to demonstrate how to request different types of activities
analyzing phenomena and solving practical problems in this context. The examples are presented in more
[12]. Most undergraduate students, however, seem to detail in [30].
process information inductively, learning details first Consider an assignment that requests the user to trace
and then proceeding to understand principles. As an how various binary tree traversal algorithms work, that
example, in introductory programming courses, most is, list the order in which the nodes are traversed when
students have difficulties with syntax and semantic de- a specific traversal algorithm is applied. A visual form
tails and find it hard to understand the relevant prin- of this exercise could include a picture of a binary tree,
ciples in the background, even if they are explicitly allowing the student to click the nodes on the screen in
explained. the appropriate order. Alternatively, the assignment
Differences in learning styles can be expressed with could be given in verbal order by giving the tree as an
learning models, which are general frameworks for char- adjacency list and asking the student to write the keys
acterizing different aspects of learners’ activities. Ex- in the nodes as a list according to their traversal order.
132
A sensing learner could solve the exercise by applying • Rudolf Fleischer – This study will measure the effec-
a simple mnemonic such that a line is drawn around tiveness of visualizations in the context of a second-
the tree starting from the left side of the root and list- year course on the theory of computation (finite au-
ing each key when the corresponding node is passed tomata, context-free grammars, Turing machines) [14].
from the left (preorder), below (inorder) or right (pos- The study will be done in Spring 2004 at the Hong
torder). An intuitive form of the exercise is that we Kong University of Science and Technology. IP, SO-
give the tree and the pseudo code of the traversal al- TS.
gorithm and ask the student to list the nodes in the
order the given algorithm visits them. • Boris Koldehofe – By using the framework presented in
this working group report and considering the results
2. Student background surveys of a previous study [25], a new study of instructor’s sat-
isfaction with LYDIAN [24], a simulation-visualization
Introductory student surveys are useful for obtaining environment for learning and teaching distributed al-
background data (such as previous classes taken, pre- gorithms, is planned. The purpose is to evaluate which
vious experience, mathematics background, and gen- features of LYDIAN contribute to instructor satisfac-
eral entrance exam scores) as well as information such tion and the teachers’ performance when integrating
as why they registered for this section of the course. LYDIAN in their course. The study also tries to deter-
This data is particularly useful in helping to deter- mine the factors which may prevent instructors from
mine whether the student’s background is a factor in actually using LYDIAN after they download the tool.
the student’s success either in the particular course, or IP, IS.
with the particular visualization tool.
• Ari Korhonen and Lauri Malmi – Intervention study
3. Time-on-task with automatic assessment – the paper [28] presents
the results of the large scale (N=550 students) inter-
Time-on-task has already been discussed in the for- vention study carried out in a virtual learning envi-
mative evaluation subsection. We have also included ronment based on the TRAKLA system [26]. The sys-
it as a covariant factor because it might be expected tem provides individually tailored exercises that the
that increased time spent using visualization may be learner solves on the Web using an algorithm simu-
a factor in performance. lation [29]. The learner receives immediate feedback
around the clock on his or her performance. The sys-
3.3 Testing with human subjects tem has features that traditional instruction cannot
We wish to issue a note of warning for those instructors provide. Thus, the study was pursued to research the
planning to run studies in their classes. In many countries, quality, advantages and limitations of this novel ap-
approval of “human subjects review boards” is required for proach. One of the conclusions was that the learning
all studies involving students. Written consent from stu- environment was as good as if they were solving the
dents involved in studies will also often be required. The same exercises in class room with human tutors giving
case studies in Section 4 provide examples of successful ap- the feedback. CP, SO-RP.
plications that others have used in broaching this sensitive
area within their institutions. Such applications may serve • Ari Korhonen and Lauri Malmi – This study [34] presents
as useful templates for instructors to use for gaining “human some experiences and observations made during 10
subjects” approval at their own schools. years of using the TRAKLA system [26]. It is in-
evitable that learners perform better when they are
allowed to resubmit their work. However, this is not
4. CASE STUDIES the whole story; it is also important to organize the
Many members of the working group have been and are course so that the skills and challenges of the learner
involved in studies that demonstrate the efficacy of the guid- coincide. Moreover, the grading policy seems to have a
ance given in Sections 2 and 3. These are collected online major impact on performance. The study summarizes
and provide concrete realizations of the principles that are the results and points out changes in learner perfor-
described in this report. The letter legends used indicate mance under the several changes the course has gone
whether the study is completed and published (CP), com- through during the time period. CP, SO-TS.
pleted but not yet published (CNP), in progress (IP), ori-
• Ari Korhonen and Lauri Malmi – This study focuses
ented toward measuring instructor satisfaction (IS), oriented
on the effect of different learner engagement levels on
toward measuring student outcomes with respect to every-
learning. The ITiCSE 2002 working group report [35]
day teaching situations (SO-TS) or more formal education
on visualization prepared a taxonomy of engagement
research (SO-RP).
in order to identify and differentiate various types of
Presently, this list includes the following:
actions the learners may perform while using visual-
izations. The plan for this study will appear at the
• Stephen Cooper and Wanda Dann – This study ex- Web site of the Computer Science Education Research
amines the use of program visualization for introduc- Group (http://www.cs.hut.fi/Research/COMPSER/)
ing objects and their behaviors using Alice [11], a 3D at Helsinki University of Technology. IP, SO-TS.
program visualization environment. Statistical data
is collected to show evidence of student performance • Marja Kuittinen and Jorma Sajaniemi – This study
and retention in CS1 for test and control groups. Early [31, 44] evaluates the use of variable roles and role-
summaries may be found in [9]. CNP, SO-TS, SO-RP. based animation in teaching introductory programming.
133
The role concept captures tacit expert knowledge in a Languages and Environments. Available online at
form that can be taught to novices. Details about the http://www.cs.umd.edu/~bederson/talks/
project and evaluation studies can be found at [47]. HCCKeynote-Sept2002.ppt (seen July 14, 2003).
CP, SO-RP. [5] Bloom, B. S., and Krathwohl, D. R. Taxonomy of
Educational Objectives; the Classification of
• Charles Leska – This study examines the impact on Educational Goals, Handbook I: Cognitive Domain.
student performance of using visualization on a unit Addison-Wesley, 1956.
in a computer literacy course. The unit focuses on
[6] Boroni, C. M., Goosey, F. W., Grinder, M. T., and
the binary representation of data and the interplay of
Ross, R. J. Engaging Students with Active Learning
the processor components during the fetch-execution
Resources: Hypertextbooks for the Web. In
cycle. Details about the study can be found at [48].
Proceedings of the 32nd ACM SIGCSE Technical
IP, SO-TS.
Symposium on Computer Science Education (SIGCSE
• Scott Grissom, Myles McNally, and Tom Naps – This 2001), Charlotte, North Carolina (2001), ACM Press,
study compares the effect of three different learner en- New York, pp. 65–69.
gagement levels with visualization in a course mod- [7] Brummund, P. The Complete Collection of Algorithm
ule on introductory sorting algorithms. Results of the Animations, 1998. Available online at
study have been published in [20]. CP, SO-TS. http://cs.hope.edu/~alganim/ccaa/ (see July 14,
2003).
• Jarmo Rantakokko – This study evaluates the effects [8] CITIDEL - Computing and Information Technology
of using algorithm visualization in parallel computing. Interactive Digital Educational Library. Available
The study focuses on student learning outcomes. De- online at www.citidel.org (seen July 14, 2003).
tails about the project and this particular study can [9] Cooper, S., Dann, W., and Pausch, R. Introduction to
be found at [49]. IP, SO-RP. OO: Teaching Objects-First in Introductory
Computer Science. In Proceedings of the 34th ACM
The working group’s URL [46] contains updated informa-
SIGCSE Technical Symposium on Computer Science
tion on these studies and others that have been added since
Education (SIGCSE 2003), Reno, Nevada (2003),
the preparation of this report. ACM Press, New York, pp. 191–195.
[10] Crescenzi, P., Faltin, N., Fleischer, R., Hundhausen,
5. CONCLUSION C., Näher, S., Rößling, G., Stasko, J., and Sutinen, E.
This report has been based on the premise that visual- The Algorithm Animation Repository. Proceedings of
ization can significantly impact CS education only if two the Second International Program Visualization
goals are met – widespread use and positive student out- Workshop, HornstrupCentret, Denmark (2002), 14–16.
comes. Visualization designers have typically not researched Available online at
the factors that would influence instructor satisfaction, and http://www.daimi.au.dk/PB/567/PB-567.pdf.
Section 2 of this report has offered guidance for beginning [11] Dann, W., Cooper, S., and Pausch, R. Making the
this process. Section 3 has provided an overview of student Connection: Programming With Animated Small
outcome measurement techniques. This overview includes World. In Proceedings of the 5th Annual ACM
approaches that can be used for informal everyday teaching SIGCSE/SIGCUE Conference on Innovation and
situations and for more formal education research studies. Technology in Computer Science Education (ITiCSE
We hope this will be of help to both visualization tool devel- 2000), Helsinki, Finland (2000), ACM Press, New
opers and visualization designers who may not be specialists York, pp. 41–44.
in doing such evaluation but who nonetheless are required [12] Felder, R. M. Author’s preface – June 2002.
to validate their product by demonstrating its effectiveness. http://www2.ncsu.edu/unity/lockers/users/f/
felder/public/Papers/LS-1988%.pdf (seen July 14,
6. REFERENCES 2003), 2002.
[1] ACM Special Interest Group on Computer Science [13] Felder, R. M., and Silverman, L. K. Learning Styles
Education. SIGCSE Education Links. Available online and Teaching Styles in Engineering Education.
at http://www.sigcse.org/topics, 2003. Engineering Education 78, 7 (1988), 674–681.
[2] Akingbade, A., Finley, T., Jackson, D., Patel, P., and [14] Fleischer, R. COMP 272: Theory of Computation —
Rodger, S. H. JAWAA: Easy Web-Based Animation A proposed study on the learning effectiveness of
from CS 0 to Advanced CS Courses. In Proceedings of visualizations, 2003. Manuscript.
the 34th ACM SIGCSE Technical Symposium on [15] Francis, L. Attitude towards Computers Scale.
Computer Science Education (SIGCSE 2003), Reno, Computers in Education 20, 3 (1993), 251–255.
Nevada (2003), ACM Press, New York, pp. 162–166. [16] Frechtling, J. The 2002 User Friendly Handbook for
[3] American Educational Research Association, Project Evaluation. National Science Foundation,
American Psychological Association, and National 2002.
Council on Measurement in Education. Standards for [17] Frechtling, J., and Sharp, L. User-Friendly Handbook
Educational and Psychological Testing. American for Project Evaluation. National Science Foundation,
Educational Research Association, Washington, D.C., 1997.
USA, 1999. [18] Good, J. Programming Paradigms, Information Types
[4] Bederson, B. Interfaces for Staying in the Flow. and Graphical Representations: Empirical
Keynote at IEEE Human-Centric Computing Investigations of Novice Program Comprehension.
134
PhD thesis, University of Edinburgh, 1999. [30] Korhonen, A., Malmi, L., Nikander, J., and Tenhunen,
[19] Good, J., and Brna, P. Toward Authentic Measures of P. Interaction and Feedback in Automatically
Program Comprehension. In Proceedings of the Assessed Algorithm Simulation Exercises. Accepted for
Fifteenth Annual Workshop of the Psychology of publication in Journal of Information Technology
Programming Interest Group (PPIG 2003) (2003), Education (2003).
pp. 29–49. [31] Kuittinen, M., and Sajaniemi, J. First Results of An
[20] Grissom, S., McNally, M., and Naps, T. L. Algorithm Experiment on Using Roles of Variables in Teaching.
Visualization in Computer Science Education: In EASE & PPIG 2003, Papers of the Joint
Comparing Levels of Student Engagement. In Conference at Keele University (2003), pp. 347–357.
Proceedings of the First ACM Symposium on Software [32] Levy, R. B.-B., Ben-Ari, M., and Uronen, P. A. An
Visualization, San Diego, California (2003), ACM Extended Experiment with Jeliot 2000. In Proceedings
Press, New York, pp. 87–94. of the First International Program Visualization
[21] Jenkins, J., and Visser, G. Making Learning Fun. Workshop, Porvoo, Finland (July 2001), University of
Available online at Joensuu Press, Finland, pp. 131–140.
http://www.imaginal.nl/learningFun.htm (seen [33] Loyd, B. H., and Gressard, C. P. Computer Attitude
July 14, 2003). Scale. Journal of Computing Research 15, 3 (1996),
[22] Keirsey, D. M. Keirsey Temperament and Character 241–259.
Web Site. Available online at WWW: [34] Malmi, L., Korhonen, A., and Saikkonen, R.
http://www.keirsey.com (seen July 14, 2003), 2002. Experiences in Automatic Assessment on Mass
[23] Kolb, D. A., Ed. Experiential Learning: Experience as Courses and Issues for Designing Virtual Courses. In
the Source of Learning and Development. Proceedings of the 7th Annual ACM SIGCSE/SIGCUE
Prentice-Hall Inc, New Jersey, USA, 1984. Conference on Innovation and Technology in
[24] Koldehofe, B., Papatriantafilou, M., and Tsigas, P. Computer Science Education (ITiCSE 2002) (Århus,
LYDIAN, An Extensible Educational Animation Denmark, 2002), ACM Press, New York, pp. 55–59.
Environment for Distributed Algorithms. In [35] Naps, T. L., Rößling, G., Almstrum, V., Dann, W.,
Proceedings of the 4th Annual ACM Fleischer, R., Hundhausen, C., Korhonen, A., Malmi,
SIGCSE/SIGCUE Conference on Innovation and L., McNally, M., Rodger, S., and Velázquez-Iturbide,
Technology in Computer Science Education (ITiCSE J. Á. Exploring the Role of Visualization and
2000) (July 2000), ACM Press, New York, p. 189. Engagement in Computer Science Education. ACM
[25] Koldehofe, B., Papatriantafilou, M., and Tsigas, P. SIGCSE Bulletin 35, 2 (June 2003), 131–152.
Integrating a Simulation-Visualization Environment in [36] Nielsen, J. Designing Web Usability. The Practice of
a Basic Distributed Systems Course: A Case Study Simplicity. New Riders Publishing, 1999.
Using LYDIAN. In Proceedings of the 8th Annual [37] Norman, D. ”Top Ten Mistakes” Revisited Three
ACM SIGCSE/SIGCUE Conference on Innovation Years Later. Available online at
and Technology in Computer Science Education http://www.useit.com/alertbox/990502.html (seen
(ITiCSE 2003) (June 2002), p. 226. July 14, 2003), May 1999.
[26] Korhonen, A., and Malmi, L. Algorithm Simulation [38] Pennington, N. Comprehension Strategies in
with Automatic Assessment. In Proceedings of the 5th Programming. In Empirical Studies of Programmers:
Annual ACM SIGCSE/SIGCUE Conference on Second Workshop (1987), G. M. Olson, S. Sheppard,
Innovation and Technology in Computer Science and E. Soloway, Eds., Ablex Publishing Company,
Education (ITiCSE 2000) (Helsinki, Finland, 2000), pp. 100–113.
ACM, pp. 160–163. [39] Price, B., Baecker, R., and Small, I. An Introduction
[27] Korhonen, A., Malmi, L., Mård, P., Salonen, H., and to Software Visualization. In Software Visualization,
Silvasti, P. Electronic course material on Data J. Stasko, J. Domingue, M. H. Brown, and B. A.
structures and Algorithms. In Proceedings of the Price, Eds. MIT Press, 1998, ch. 1, pp. 3–27.
Second Annual Finnish / Baltic Sea Conference on [40] Ross, R. J. Hypertextbooks for the Web. In
Computer Science Education (October 2002), Proceedings of the First International Program
pp. 16–20. Visualization Workshop, Porvoo, Finland (July 2001),
[28] Korhonen, A., Malmi, L., Myllyselkä, P., and Scheinin, University of Joensuu Press, Finland, pp. 221–233.
P. Does it Make a Difference if Students Exercise on [41] Ross, R. J., and Grinder, M. T. Hypertextbooks:
the Web or in the Classroom? In Proceedings of the Animated, Active Learning, Comprehensive Teaching
7th Annual ACM SIGCSE/SIGCUE Conference on and Learning Resources for the Web. In Software
Innovation and Technology in Computer Science Visualization (2002), S. Diehl, Ed., no. 2269 in Lecture
Education (ITiCSE 2002) (Århus, Denmark, 2002), Notes in Computer Science, Springer, pp. 269–284.
ACM Press, New York, pp. 121–124. [42] Rößling, G. Algorithm Animation Repository.
[29] Korhonen, A., Malmi, L., Nikander, J., and Silvasti, P. Available online at http://www.animal.ahrgr.de/
Algorithm Simulation – A Novel Way to Specify (seen July 14, 2003), 2001.
Algorithm Animations. In Proceedings of the Second [43] Rößling, G., and Freisleben, B. Animal: A System for
International Program Visualization Workshop, Supporting Multiple Roles in Algorithm Animation.
HornstrupCentret, Denmark (2002), pp. 28–36. Journal of Visual Languages and Computing 13, 2
Available online at (2002), 341–354.
http://www.daimi.au.dk/PB/567/PB-567.pdf.
135
[44] Sajaniemi, J., and Kuittinen, M. An Experiment on online at http://www.algoanim.net, 2003.
Using Roles of Variables in Teaching Introductory [47] Kuittinen, M., and Sajaniemi, J. Home Page of the
Programming. Submitted to Empirical Software Study on Variable Roles and Role-Based Animation.
Engineering (2003). Available online at
[45] Schwarm, S., and VanDeGrift, T. Making http://cs.joensuu.fi/~saja/var_roles/, 2003.
Connections: Using Classroom Assessment to Elicit [48] Leska, C. Homepage of the Visualization Study in
Students’ Prior Knowledge and Construction of Computer Literacy. Available online at
Concepts. In Proceedings of the 8th Annual ACM http://faculty.rmc.edu/cleska/public_html/
SIGCSE/SIGCUE Conference on Innovation and algvizstudy.htm, 2003.
Technology in Computer Science Education (ITiCSE [49] Rantakokko, J. Homepage of the Visualization Study
2003) (2003), ACM Press, New York, pp. 65–69. in Parallel Computing. Available online at
[46] Visualization Working Group Home Page. Available http://user.it.uu.se/~jarmo/hgur.html, 2003.
136