Video Conferencing - User Interface With A Remote Control For TV-sets
Video Conferencing - User Interface With A Remote Control For TV-sets
Master's Thesis
EAT 2015
Abstract
A user interface with a TV remote control solution has been prototyped and
implemented demonstrating a peer-to-peer video conference system. This was
done through the use of on-screen displays, an iterative design process with
multiple usability tests and an evaluation of similar video conference products
as a basis. The test results shows that a prototype with high usability was cre-
ated which the users felt confident using.
We would like to thank Niklas Svensson for his help with the techncial parts of this thesis
and also our two supervisors John Rehn and Joakim Eriksson for all the good advise. The
parallel master thesis by Marcus Carlberg and Christoffer Stengren also in video confer-
ence project has always been supporting and been given relevant advises. Last but not
least we want to thank everyone we have come in contact with that has helped us at Axis
Communication.
iii
iv
Contents
1 Introduction 1
1.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Similar work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Purpose . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.4 Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.4.1 Market overview . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.4.2 Iterative design process . . . . . . . . . . . . . . . . . . . . . . . 2
1.4.3 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.5 Contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.5.1 Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.5.2 Report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.6 Overview of thesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2 Technical background 5
2.1 High Definition Multimedia Interface . . . . . . . . . . . . . . . . . . . 5
2.2 Consumer Electronic Control . . . . . . . . . . . . . . . . . . . . . . . . 6
2.2.1 Communication . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.2.2 Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.2.3 Features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.3 Overlay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.4 PTZ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3 Theoretical background 13
3.1 Interaction Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.2 User-centered Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.3 Cognitive graphical design . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.4 Designing the user interface . . . . . . . . . . . . . . . . . . . . . . . . 14
3.4.1 Card sorting method . . . . . . . . . . . . . . . . . . . . . . . . 15
3.5 Prototypes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
3.5.1 LO-FI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
v
CONTENTS
3.5.2 HI-FI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.6 Usability evaluation methods . . . . . . . . . . . . . . . . . . . . . . . . 16
3.6.1 Use cases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.6.2 Scenarios . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
3.6.3 Usability inspection . . . . . . . . . . . . . . . . . . . . . . . . 17
3.6.4 Usability testing . . . . . . . . . . . . . . . . . . . . . . . . . . 17
5 First iteration :
Requirements & concepts 29
5.1 Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
5.1.1 Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
5.1.2 Exploratory testing . . . . . . . . . . . . . . . . . . . . . . . . . 30
5.1.3 Heuristic evaluation . . . . . . . . . . . . . . . . . . . . . . . . 31
5.2 Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
5.2.1 Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
5.2.2 Exploratory testing . . . . . . . . . . . . . . . . . . . . . . . . . 31
5.2.3 Heuristic evaluation . . . . . . . . . . . . . . . . . . . . . . . . 33
5.3 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
6 Second iteration :
Design & prototyping 37
6.1 Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
6.1.1 Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
6.1.2 Assessment testing . . . . . . . . . . . . . . . . . . . . . . . . . 38
6.2 Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
6.2.1 Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
6.2.2 Assessment testing . . . . . . . . . . . . . . . . . . . . . . . . . 39
6.2.3 Results from open discussion . . . . . . . . . . . . . . . . . . . . 41
6.3 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
6.3.1 Error sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
7 Third iteration :
Graphic design & high level prototypes 45
7.1 Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
7.1.1 Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
vi
CONTENTS
7.2 Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
8 Implementation 49
8.1 Overview of the setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
8.2 CEC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
8.3 Overlays . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
8.4 Technical Limitations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
References 57
vii
CONTENTS
viii
Terminology
Abbreviations
API Application Protocol Interface
HI-FI High-fidelity
LO-FI Low-fidelity
PTZ Pan-Tilt-Zoom
Concepts
Local camera A video conferencing camera that is present and captures the user.
ix
CONTENTS
Vapix VAPIX in an open API created by Axis that uses standard protocols to
enable integration into a wide range of solutions on different platforms. Almost all func-
tionality available in Axis products can be controlled using VAPIX. [2]
Test Person The test person is the person who are performing a usability test.
x
Chapter 1
Introduction
1.1 Background
Axis Communications AB is a company that primary develops network cameras, the com-
pany’s origin is in Sweden with the headquarter in Lund. The company is at the moment
the market leader in network based video surveillance, Axis is also a driving force in the
transition from analog to digital video surveillance. The number of employees at Axis is
around 1600 worldwide and the headquarter hosts around 1000 of them [3].
AXIS cameras contain a processor chip and can be used for much more areas than just
surveillance, without complementing hardware it could be possible to make video confer-
ence calls and therefore entering on a new market with this desirable feature. For example,
connecting the camera directly to a screen and holding a video conference should work
well without the need of a computer which many of the solutions on the market require.
In our thesis we will use a standard TV remote control to a HDMI compliant TV, use on
overlays to configure the cameras through CEC.
1
1. Introduction
1.3 Purpose
The purpose of this masters thesis was to investigate how CEC, TV remote control user
interaction and a high usability GUI can be introduced to and prototyped on an Axis PTZ
camera. The purpose was to make it easier to have video conferencing sessions with no
need for a separate computer to control the system and where all configurations are made
with a standard TV remote control. By using open standards the user should only need an
Axis PTZ camera along with standard television set thus having a plug-and-play experi-
ence. By prototyping this it will be easier to take the next step and in the future make the
prototype into a real product.
The main goal of the thesis is to design and prototype an intuitive interface combined with
a standard TV remote control for a video conferencing system using Axis hardware and
software and standard AV equipment. To achieve this a set of goals has been set up.
• Make a usability evaluation of similar products to identify good and bad video con-
ference interfaces and designs to get a better understanding of how to make a good
user interface for a video conference system.
• Make a working code prototype that shall demonstrate a bidirectional system where
it is possible to install the product, initiate a video conference and remotely control
the other camera’s pan/tilt/zoom functions, without the use of a computer.
• The design of the prototype shall focus on usability and ease of use. The prototype
shall be efficient to use, easy to learn and satisfactory to use. The user shall feel
confident in using the prototype.
1.4 Approach
A literature study was performed to find relevant theories within the area of usability and
also to find relevant technical information regarding HDMI and CEC.
2
1.5 Contribution
Figure 1.1: The design process circle which illustrates one itera-
tion.
1.4.3 Implementation
The implementation of the prototype was done parallel with the iterations. Initially a
set of minor programs was made to get a better understanding of how the camera works
and its different features. In the second and third iteration phase a Hi-Fi prototype was
implemented. Since the camera is Linux based, the implementation was made in the C
programming language. See chapter 8 Implementation for more on this.
1.5 Contribution
Following section describes the writers contribution to the thesis and how the work was
divided.
1.5.1 Development
Both Alexander and Catarina planned and did the exploratory testing and assessment test-
ing together along with most of the design of interfaces for the prototypes. Alexander was
responsible for the technical parts of the thesis.
Alexander
Alexander evaluated set-tops and their remote controls as a part of the evaluation of sim-
ilar products. Also wrote all code for the prototypes and configured them and made the
overlays for the prototype in the second iteration.
Catarina
Catarina evaluated software based systems as a part of the evaluation of similar products.
Also did the heuristic evaluation of the first iteration and made the overlays for the proto-
3
1. Introduction
1.5.2 Report
The report was initially written together by both Alexander and Catarina and later on re-
sponsibilities for the different chapters was divided between the two. Alexander wrote the
abstract and was responsible for chapter 1, 2, 5 and 8 while Catarina was responsible for
chapter 3, 4, 6 and 7. Both were responsible for chapter 9. The chapters were mostly
written by the person who was responsible for them.
4
Chapter 2
Technical background
The Technical background chapter introduces the reader to the protocols and technologies
which were used throughout the thesis. The chapters’ purpose is to give the reader a better
understanding of the thesis’ technical aspects.
HDMI is a digital audio/video interface for transferring video and audio from a HDMI-
compliant source device, known as a HDMI Source to a HDMI-compliant audio/video
receiver, also known as a HDMI Sink. HDMI Sources include DVD players, Blu-ray Disc
players, video game consoles, VCRs and cameras which can be linked together with HDMI
Sinks like TVs, projectors or monitors, or just about any other device capable of sending or
receiving an high-definition (HD) signal to create a HDMI system, such as a home theater
system. HDMI has become a modern replacement for previous standards such as com-
posite video, Separate Video (S-Video), Video Graphics Array (VGA) and Digital Visual
Interface (DVI). HDMI transmits all HDTV standards and supports up to 8 channels un-
compressed audio. HDMI is basically the same as the interface standard DVI for PCs and
displays but with the addition of audio, smaller connector and CEC [4].
The HDMI founders are Hitachi, Panasonic, Philips, Silicon Image, Sony, Thomson and
Toshiba [5]. With over two billion HDMI-enabled products shipped since the start of the
HDMI standard in 2003 it is the industry leading standard for audiovisual equipment and
has extended to a vast array of applications, devices and industries [6].
5
2. Technical background
Figure 2.1: The HDMI connector pinout where the CEC pin (13)
is marked.
Even though CEC has been included in the HDMI standard since its start in 2002, only
the wiring is mandatory, the implementation of CEC in products is optional. This means
that all HDMI-compliant products does not allow for high-level control functions. Many
companies who choose to adapt CEC have their own trade names for CEC, see table 2.1.
The companies use different trade names partly due to vendor specific implementations
which often only guarantees full support of high-level control functions between products
6
2.2 Consumer Electronic Control
from the same company, though in theory products supporting different trade names should
be able to work together and be able communicate entirely using the standardized CEC
commands. See section 2.2.3 Vendor Specific Commands for more on vendor specific
implementations.
2.2.1 Communication
The CEC communication is bidirectional which allows for both the HDMI Source and the
HDMI Sink to send and receive CEC messages. Since the control data is carried over
a single wire bus, the connected HDMI devices can not send messages simultaneously.
The device that receives a CEC message is called a Follower and is required to respond
to the message. The device that has just sent the CEC message is called a Initiator and, if
appropriate, waits for the Follower to respond to the message[7]. Both the HDMI Source
and HDMI Sink can take the roles as Initiator and Follower and will likely switch between
these roles multiple times during the communication session. In figure 2.2 Device 1 is the
Initiator and Device 2 is the Follower.
2.2.2 Messages
Frames and blocks
A CEC message, also called a frame, consist of a number of blocks. A block is always 10
bits long, where the first 8 are information bits followed by an end of message bit (EOM)
and an acknowledge bit (ACK) see table 2.2. The information bits differs depending on
which type of block it is. For header blocks information bits are the destination address,
bit 0 - 3, and the initiator address, bit 4 - 7. For data blocks they are opcodes or operands,
more on this under Data blocks in section 2.2.2. The end of message bit indicate if the
block is the final block of the frame, 1 meaning that it is the final block and 0 meaning that
one or more data blocks follow. The acknowledge bit is used by Follower to acknowledge
the block. It is always set to 1 by the Initiator and set to 0 by the Follower if it reads
its own address in the destination address. The block is then sent back to the Initiator.
7
2. Technical background
Therefor an ACK read 0 by the Initiator indicates a successful transmission of the block.
The control bits, EOM and ACK, are always present and always have the same usage. The
first block of a frame is always a header block and it is also always the only header block
in the frame. The number of data blocks differs depending on the type of message, e.g.
a ping message to ascertain that other devices are turned on contains only a header block
and no data blocks. The maximum frame size is 16 blocks, 16 × 10 bits. [8]
Header/Data Block
7 6 5 4 3 2 1 0 - -
Information bits EOM ACK
Data blocks
There are two different types of data blocks. The first type is opcode (operation code)
blocks that has the purpose of describing what type of message is being sent. The 8 infor-
mation bits in opcode blocks are unique message identifiers, e.g. User Control Pressed,
which indicates that the user has pressed a button on the remote control, has the unique
value of 0x44, see section 2.2.3 Remote Control Pass Through for more on this. The other
type is operand blocks which contains parameters used by the message.
Header Block
0 0 0 0 0 0 0 1 0 1
Initiator Destination EOM ACK
Opcode Block
0 1 0 0 0 1 0 0 0 1
User Control Pressed (0x44) EOM ACK
Operand Block
0 0 0 0 0 0 0 0 1 1
Select (0x00) EOM ACK
Table 2.3 shows an example of a message sent from the TV to a recording device. The
message contains three blocks sent in the order of header, opcode and operand block. The
header block contains the initiator address 0, which is the logical address of the TV, and the
destination address 1 which is the logical address for Recording device 1. Since a message
can only contain one header block and a opcode block must precede operand blocks, the
recording device knows that second block is an opcode block. By reading the unique id of
8
2.2 Consumer Electronic Control
0x44 the recording device knows that the user has pressed a button on the remote control
and awaits a operand block describing which button it is. The information bits for the
operand block is 0x00 which is the unique id for the Select button. The recording device
now knows that the user has pressed the Select button on the remote control. Note that the
EOM of the operand block is set to 1 indicating the end of the message. Also not illustrated
in the table is the acknowledgements for each block sent from the recording device to the
TV.
2.2.3 Features
CEC provides a number of features designed to enhance the interoperability and function-
ality of devices within a HDMI system. Implementation of these features are optional but
if a CEC feature is supported, the feature has to be completely supported with all mes-
sages and their behaviour implemented. Not all feature are covered in this section, only
the features relevant for this theis.
9
2. Technical background
Figure 2.3: Sequence diagram for when a user presses and holds
down the Up key on the remote control for a period of time before
releasing the key.
the feature is to send messages about button presses which are not supported in Remote
Control Pass Through.
It is important to note that if a device supports a CEC feature it is not obliged to use that
feature before Vendor Specific Commands. For example, a device may support Remote
Control Pass Through but choose to send Vendor Specific Commands instead for commu-
nicating about remote control button presses for certain remote control buttons, even if
they are supported in Remote Control Pass Through. [10]
With the use of the open source Pulse-Eight project libCEC [11] along with hardware like
the credit card-sized single-board computer Raspberry Pi [12], it should be possible to
listen to Vendor Specific Commands messages and interpret them.
2.3 Overlay
Overlays is put on the picture from the camera block in the chip module. Overlay images
are converted from bitmap file format to a ovl file format and later on added. The overlays
can be in different sizes and there can be more than one at the same time. It is also possible
to make parts of an overlay transparent.
10
2.4 PTZ
2.4 PTZ
PTZ is an abbreviation for Pan Tilt Zoom and is a term used for cameras. A PTZ camera
is capable of pan, tilt and zoom movements not available to stationary cameras.
11
2. Technical background
12
Chapter 3
Theoretical background
The Theoretical background chapter introduces the reader to some important usability and
design concepts which is important for the coming chapters.
Interaction design (abbreviated IxD) is the knowledge about shaping digital products for
people’s use. IxD fills the design needs of digital products with dynamic behavior and
changing user interfaces.
Designing interactive products to support the way people communicate and interact in
their everyday and working lives [13].
Once we know something, we find it hard to imagine what it was like not to know it. Our
knowledge has "cursed" us [13].
13
3. Theoretical background
measured that the usability can be improved by 38% per iteration. [15]
There are some cognitive laws according to the cognitive graphical design. The laws are
called the law about proximity, similarity, closure, symmetry, common fate, continuity and
past experience. [16]
When creating graphical user interfaces the choose of color is very relevant. The brain
connects different colors to different feelings and therefore the choice of color is impor-
tant.
The brain works faster when it recognises a picture and uses its passive memory instead
of processing letters to a word with the active memory which gives that icons should be
used. Icons shall in general be simple pictures and be based on nouns. [17]
14
3.5 Prototypes
There are some important concepts that are relevant when designing the user interface.
These concepts are learnability, efficiency, attitude, flexibility and relevance. They are
important due to the fact that the concepts together contains all the important characteris-
tics from a smooth user interface.
Relevance How well the product serves the functions needed by the user.
[18]
3.5 Prototypes
A prototype is the manifestation of a design suggestions for you to interact with in any
way. There are two kind of prototypes, horizontal and vertical. The horizontal has many
functions and few details, the vertical in opposite has few functions but is more detailed.
[19]
3.5.1 LO-FI
A LO-FI prototype equals to a low fidelity prototype and is often drawn by hand. LO-FI
prototypes are cheap and quick to make. Another advantage is that it can be easier to give
constructive feedback on a prototype that does not look polished. Some disadvantages are
that usability tests becomes harder and the navigation is quite limited. [19]
15
3. Theoretical background
3.5.2 HI-FI
A HI-FI prototype equals to a high fidelity prototype and is often computer made. It takes
a lot of time to create a HI-FI prototype. Adobe Photoshop or Microsoft Power Point is two
common programs that can be used for HI-FI prototyping. HI-FI prototypes can prefer-
ably be used when usability test must be performed and therefore a click-able prototype is
needed. It can in most cases be more cost effective to first create HI-FI prototypes because
changes are easier done in non implemented prototypes. [19]
3.6.2 Scenarios
Good scenarios are concise and answer the following key questions:
Who is the user ? Why does the user come to the site/view? What goals does he or she
have?
With Goal- or Task-Based Scenarios it is only stated what the user wants to do. Any
information on how the user would complete the scenario should not be included. These
scenarios are useful in helping to define your site architecture and content. This type is
suited for usability test scenarios. It gives them a reason and a goal for going to the site,
but it lets them show you how they would use the site to accomplish that goal.
Elaborated Scenarios give more user story details. These details give the Web team a
deeper understanding of the users and users’ characteristics that may help or hinder site
interaction. Knowing this information, the team is more likely to develop content, func-
tionality, and site behavior that users find comfortable and easy to work with.
When identifying scenarios for usability testing, you should limit your test to 10 to 12
tasks due to time constraints. Additionally, in a usability test, you can ask users for their
own scenarios.
16
3.6 Usability evaluation methods
Debriefing
The main goal of having a written post questionnaire is to clarify and deepen your under-
standing of the test person. Every question in a written debriefing must be asked to the test
person in the same way but in an oral debriefing the questions can be adopted depending
on the test results. The questions shall be subjective, not performance based.
Pilot test
A pilot test is a preliminary study performed to validate the study itself. The pilot test is
partly used to confirm that the questions are correctly formulated, it can also be a time and
cost test and there is a possibility to change the arrangement of the study post pilot test.
[26]
Exploratory testing
This testing technique is cheap and perfect for discovering the interface. The design is not
yet decided and the result from the test will affect the future design work. This method
is used early in the design phase and the goal is to validate how the design works. This
17
3. Theoretical background
method can be applied on paper prototypes which makes it cheap to make radical changes.
The test leader interacts very much with the test person for example ask questions and start
discussions.
Assessment testing
This testing method is in contrast to Exploratory testing a way to test a ready design and
perhaps a completed prototype. This test method can be used either in the end of the design
phase or in a early implementation phase and the main purpose of this is the localise design
issues. [29]
18
Chapter 4
Video conference systems
A market overview of similar products
The market overview is done to take inspiration from similar products and this chapter
explains the method used and the pros and cons with the similar products.
In this chapter different video conference systems has been compared and evaluated. To
introduce the reader to the chapter the first section will cover the video conference camera
this master thesis is based on, followed by an introduction of the existing products on the
market which were chosen for evaluation. The purpose of the evaluation is not to identify
the market or to find and compare possible main competitors but to evaluate the usability
of similar products to identify good and poor design choices. The evaluation and the
conclusions drawn will serve as a helpful basis in the design process of the prototype. The
evaluation is based on four different use cases which are described in section 4.3 Methods
for evaluation. Video conference remote controls are also covered and evaluated in this
chapter.
19
4. Video conference systems
A market overview of similar products
The four most common types of videoconferencing systems are telepresence videocon-
ferencing systems, integrated videoconferencing rooms, set-top or appliance videoconfer-
encing systems and desktop videoconferencing systems. Telepresence systems give the
appearance of being present in an actual meeting even though the participants are geo-
graphically dispersed. They usually consist of higher quality video and audio, often with
several large displays. This is the most expensive type of the four. Integrated room video-
conferencing systems are most often used in conference rooms and board rooms and are de-
signed for larger groups. These systems are usually permanently mounted with customized
configurations and are normally equipped with multiple features which allow the room
to be used for other functions besides videoconferencing. Telepresence and integrated
videoconferencing systems will not be covered in the evaluation, due to their limitations
in availability and high cost. Set-tops and appliance videoconferencing systems are non-
stationary systems that are most useful in small conference rooms and for smaller groups
of participants. UtilityCam falls under this type of videoconferencing system. Desktop
systems are often completely software based and requires a computer and separate camera
to use. These products are often designed for non-business use. [30]
Logitech TV Cam HD
The Logitech TV Cam HD is a set-top video conferencing system that connects to a tele-
vision set with HDMI. The product has its own remote control and is integrated with
Skype. The product is described as a "living room Skype approved all-in-one TV cam"[31].
Though the system is not designed for a business environment there are some similarities
worth looking into, PTZ control is supported along with a 10-foot user interface.
20
4.2 Different products on the market
puter running Google’s Chrome OS operating system. Chromebox for meetings connects
to a display using DisplayPort++ or HDMI with no support for CEC [32]. The interface
is similar to Google+ Hangout but designed for a larger screen. [33]
Blue Jeans
Blue Jeans offers a cloud-based video conferencing service that can connects different
users with no requirements on the hardware which makes Blue Jeans to a very dynamic
video conference client. The main purpose of Blue Jeans is to bring together business
conferencing solutions. [37]
Hookflash
Hookflash is a application for smart phones which offers text messaging, telephoning and
video conferencing over a peer-to-peer network. The system also offer features like call
transfer. [38]
Palbee
Palbee is a Video Conference system that is used mainly for sharing knowledge. The
system offers limited talking hours and participants but unlimited amount of sessions. The
interface is flash based and evry kind of computer with an internet access can use it. [39]
ParkSlide
ParkSlide is a system for video conferencing that from the beginning where designed
around social networking. ParkSlide has since the beginning developed into an enter-
prise direction. ParkSlides vision is to change the common overly complicated systems
and software in video conferencing. [40]
21
4. Video conference systems
A market overview of similar products
The purpose of this use case is partly to capture the first time the user is introduced to
the system and how well it is introduced to the user. It includes the necessary, and possi-
bly unnecessary, steps before the product can be used for video conference call from when
the system is turned on for the first time after all cables has been correctly connected.
22
4.3 Methods for evaluation
a preconfigured contact or manually adding the address. Most of the systems have the call
camera symbol in direct connection to person the user want to call which prevent errors
due to the fact that the user sees the picture at the same time as calling. One thing that
increased the user experience were in Hookflash when calling somebody a spinning circle
and a timer of how long the user have waited occurred on the screen. Many of the initiate
calls is by pressing a button with a camera symbol. This symbol is a bit miss leading, a
telephone might would be better as dialing symbol because the user will not mistake the
dialing button for watch the camera.
23
4. Video conference systems
A market overview of similar products
4.4 Evaluation
The analyse includes to interpret the information given from the use case evaluation of the
video conference systems described above. In this section we will discuss pros and cons
from different graphical interfaces and the effects it will cause for the user as well. The
analytic evaluation will be grounded of what expectations is held on a video conference
system interface which is a subjective form of evaluation and the functionality performance
which is the objective form of a evaluation. Except for the mentioned evaluation methods,
a combination of heuristic evaluation and comparison testing is performed as well. The
video conference cameras compared in this section are described above 4.2 as well as the
testing methods and techniques 3.6.
24
4.4 Evaluation
a lot on the design and had a intuitive tab for installing the camera, the interface were
overall very clean and easy to find in. Parkslides slogan is "Get started in seconds, talk
for hours". Chromebox for meetings had a fast and straight forward setup process, the
user did not get overwhelmed with unnecessary configuration steps since the setup only
consisted of picture, sound and microphone checking. Easy as it might seem there are
some prerequisites, Chromebox has to be setup and running and the user needs a Google
account. Bluejeans is a video conference client that is compatible with a bunch of different
cameras which makes it easier for a company to have meeting without the same type of
hardware. The functionality is also implemented in other VoIP system such as Skype,
Skype is usually not used for video conference calls, although for example Logitech have
a Skype integration for their video conference camera.
Initiate a call
Most of the product that where evaluated by us had tried to make this step intuitive and
logical. A problem with Hookflash were that to initiate a call the user had to manually
type in the number to be able to make a call. Parkslide has a smooth drag and drop system
to add people in the conversation. This is showed for the user with a text and a box where
you can drop the pictures of the one you want to talk with. For example Blue Jeans also
uses the color mapping with red to end call which the human mind connects with stop.
Answer a call
Hookflash has a very simple and good way to answer a call, the buttons are color mapped
to make it more logical for the user. Accept is green, ignore is orange and decline is red.
PTZ Control
In the different PTZ remote controls there are some different approaches of how to control
the pan tilt and zoom. The pan and tilt is often remote controlled with arrows, see figure
4.5, it is logical that the panorama view is navigated by left and right arrow and the tilt
with up and down arrows. The zoom in some remote controls are logical with mapping to
magnifying glass with a plus or a minus. Both 4.5 and 4.6 with a W and T which not are
that intuitive. The W and T refers to a word that does not work cross languages, a symbol
is better. In some cases there are two different speeds for zoom which could be useful in
some cases but in general there could be too many buttons as in 4.5. Depending on what
the remote control should be used for sometimes less is more.
25
4. Video conference systems
A market overview of similar products
26
4.4 Evaluation
separately. The design and complexity of these remote controls can vary a lot depending
on which aspects the manufacturers have decided to focus on.
Figure 4.7: From the left, remote control for the Icon series and
remote control for the 220 series.
Lifesize chose two very different approaches in their remote control designs for their se-
ries, 220 and Icon, see figure 4.7. For their older 220 series the remote control is very
similar to a standard TV remote control with many buttons and a higher learning curve.
For their newer Icon series they took a complete u-turn and made their remote control as
simple as possible with only the most essential buttons: a mute button, navigation buttons
and a select button. That is a total of 6 buttons compared to 33 buttons in the 220 series.
Lifesize says that with their new remote control for the Icon series they "want the users
to never have to look down at the remote"[41], this way the user can focus on the display
instead of trying to figure out which buttons do what. For the user to be able to control
the system with only 6 buttons, down from 33, they also had to revamp the navigation and
menu system, see Menu system in section 4.4.
Google’s remote control for Chromebox for meetings has also tried to adopt the same
simplicity as Lifesize Icon remote control, at least partly. The front look very similar to
the Icon remote control with the only additions of an end call button and a menu button,
see figure 4.8.
27
4. Video conference systems
A market overview of similar products
Figure 4.8: Front side of remote control for Chromebox for meet-
ings.
Figure 4.9: Rare side of remote control for Chromebox for meet-
ings.
28
Chapter 5
First iteration :
Requirements & concepts
In this chapter the early development of the prototype is described. The chapter is divided
into three main parts: Method, Results and Conclusions. Method covers the work process
and the methods used, Results describes the results from the tests and Conclusions consist
of the conclusions drawn from the results.
The goals for the first iteration with a horizontal paper prototype were the following. De-
cide what functions that would be included and decide what steps that should be included.
There was no focus on in-dept details in this iteration, just proof of concepts. The things
that were prioritized in this step were the terminology, the conceptual design and the cat-
egorisation and grouping of features. The goals were fulfilled by creating prototypes that
are based on the conclusion from the evaluation of similar products in the previous chapter.
5.1 Method
29
5. First iteration :
Requirements & concepts
Figure 5.1 illustrates the work process for the first iteration. Initially requirement priori-
tisation and grouping was performed using the card sorting method. Based on the results
from the card sorting and the conclusions from the previous chapter a paper prototype was
made which was tested and evaluated. The conclusions from the evaluations was later
used to improve the prototype in the second iteration.
5.1.1 Design
Prioritizing
Every function from the current function analysis where included in a prioritization session
and where given number with the the numeral assignment technique. The 13 functions
where categorized from 1 to 13 where 1 equals the function that is most prioritized.
Paper prototyping
The prototypes in the first iteration were hand drawn on paper so usability testing and us-
ability inspection could be performed. The prototypes where inspired by conclusions from
the evaluation of similar product in 4.4.
The test was performed on three master students that has specialised in usability and de-
sign at LTH. The reason we chose these test persons, which are not in the target group,
is because we wanted to get a lot of feedback regarding the usability. The students did a
combination of both a form of expert judgment due to their academic background and a
form of personas due to the fact that they getting a scenario they have to adopt to.
Six different scenarios were created which all contained a task for the test person to per-
form. The test leader tried to introduce a discussion for every scenario while the note taker
described the scenarios and recorded what the test person expressed. The test started with
30
5.2 Result
a short description of the master thesis and picture 4.4 where shown for the test person.
For more on the presentation and scenarios, see appendix B.
5.2 Result
5.2.1 Design
The results from the prioritizing and grouping of functions can be seen in figure 5.2.
31
5. First iteration :
Requirements & concepts
Figure 5.3: The view in the first iteration prototype for controlling
the PTZ values of the local camera.
All three test participants had issues with the menu navigation. They expressed uncer-
tainty for what the navigation buttons did in different situations. Most notable was during
scenario 2 where the test participants wanted to pan, tilt and zoom the camera, and sce-
nario 3 and 4 where the test participant wanted to change the settings. During scenario 2,
when the test participants had reached the view shown in figure 5.3, two of the test par-
ticipants expressed that they were uncertain if pressing the arrow buttons on the remote
control would pan and tilt the camera or would continue to navigate through the menu.
Similar comments were made for the settings view, figure 5.4, where the test participants
had issues determining if their action would affect the left or right panel. It was also stated
that not enough feedback was given regarding if tabs or menus were selected or not, though
one test participant thought that it was because the test was made on a paper prototype. The
same test participant suggested an alternative way for changing settings in a list such as
language by using the numpad or the channel up or down buttons. During the RP, one test
participant said that there were some inconsistency for menus layout and how they worked.
One test participant did not understand the term picture-in-picture in the advanced tab un-
der My Settings.
It was also made clear that in the right panel of figure 5.3 it was found intuitive and easy to
understand how to pan, tilt and zoom the camera, none of the participants failed this task
on their first try. During the test one participant was asked about the zoom icon in the upper
right corner and if it would’ve been better to use perspective arrows. The test participant
thought it was much better as it was in the test because that icon was recognized from the
remote control. With different icons for the button on the remote control and for the screen
it would be more difficult to map these together. The purpose of being able to control the
remote camera and not the local camera during a video conference call perplexed a test
participant, as a user he wanted to be in control of his own product, not anyone else. It
was also expressed that it deprives the purpose of configuring the local cameras pan, tilt
and zoom if the user on the other end of the video conference call can change it freely.
Overall the test participant thought that PTZ control was a "luxurious feature" which isn’t
32
5.2 Result
Figure 5.4: The view in the first iteration prototype for changing
system language.
necessary.
• It is enough to only control your own cameras PTZ according to one test participant.
According to Jakob Nielsens second rule, see appendix A, the system should be designed
for its target group with for example terminology which the prototype is. The headlines in
the main menu could have more professional headlines instead of MyContacts, MyCamera
and MySetting but on the other hand the names indicates precisely what the view is aimed
for which is good.
The third rule is fulfilled because the user can always navigate to another state in the main
menu. During a call the user can press the menu button to end the call. If it is to hard to
end call, a mistake call could be embarrassing but on the other hand if it is very easy to
end call the user might end calls by mistake.
33
5. First iteration :
Requirements & concepts
The navigation should always be straight forward and in the MyCamera and MySettings
views there is need for clearer navigation possibles and indications of where the user are.
The interface is indeed flexible for both new and advanced users, by default there is a lot
of help views and the views are design for not that technical users. Even though there is
a help view during the minimalistic call view, the order of the functions must be thought
through.
5.3 Conclusions
Terminology
Picture in picture is a technical term which is hard to understand, a better way could be to
use self explanatory images or icons. My Contacts, My Camera and My Settings should
be changed into something more professional, it can also give the user the false impres-
sion that the user interface is personal, even tough it is room based. In the Language tab,
United States should be changed to English. It is not clear what the Security tab means,
Privacy could be a better term. The zoom indicator in the PTZ view, figure 5.3, is easy to
understand and maps to the remote control button well. A reason for why two of the test
participants failed scenario 1 could be the lack of background information. Inexperienced
video conference users often associate video conference systems with Skype, which is a
personalized product unlike the prototype which were tested. During the setup it should
be made more clear that the user is asked to insert the name of the conference room.
Features
“The hardest part of design, especially consumer electronics, is keeping
features out.” — Donald A Norman
Since a goal of this thesis is to develop an initial prototype which focus on the ease of use,
the scope for included features should be reduced in upcoming iterations. This is espe-
cially relevant for features which can not be fully implemented due to technical limitations
and would only serve as a placeholder and not a proof of concept of how such feature could
34
5.3 Conclusions
or should work. Since the feature can not be fully implemented it would not be correctly
tested in upcoming iterations which is planned to be done on the implemented prototype.
Picture in picture is one such feature that also was found to be a technical term that might
be difficult to understand. The feature picture in picture does not necessarily have to be
removed in upcoming iterations but the option of enabling and disabling picture in picture
should. Advanced settings and features which would primarily be used by advanced users
will thereby be reduced in upcoming iterations.
35
5. First iteration :
Requirements & concepts
36
Chapter 6
Second iteration :
Design & prototyping
This chapter contains information about the method used in iteration 2 and the results
which is half way into the design. The results are presented in form of digital prototypes.
The chapter is formed in the same way as chapter 1.
The goals for the second iteration included creating a HIFI-prototype. In contrast to it-
eration one where it was decided what functions should be in what view, this iteration
focused instead on how to interact with the different functions in a natural and useful way.
The second iteration is more detailed than the previous one. what was most important in
this step, was to get feedback about how well the user understand the hidden functions and
also feedback on the use of color et cetera.
6.1 Method
The goals and test method from the second iteration can be found in appendix C. The iter-
ation started with a complete redesign from iteration 1 that were based on the conclusions
from iteration 1. Digital prototypes were created and usability tests was made. Each test
included a debriefing, the results were used to draw conclusions.
37
6. Second iteration :
Design & prototyping
6.1.1 Design
This was still quite early in the design phase and it was therefore decided to rethink the
design from iteration 1 to be find a better solution. The main menu was completely re-
designed and the main goal was that less is more. Even though the old design in 5.3 kept
the user aware of where they were and how they could navigate to other view, the new
design seen in 6.2 is a better design for TV interfaces.
Techniques used for the assessment testing were concurrent think aloud and retrospective
probing which are technique where the test person talk loud what they think and how they
resonate while retrospective probing means that there is a discussion after the test session.
[42] [43]. The data collected during the test were both quantitative and qualitative data.
6.2 Result
6.2.1 Design
38
6.2 Result
The data presented in table 6.2 shows the results from seven questions, seen in table 6.1,
from the seven test persons and two pilot testers. The debriefing survey can be found in
appendix D. The table shows that question three and question seven got the lowest points
for the questions where high points were preferred, which are all except question two and
four. With the low mean value of Q7 of 3.78, it should be noted that the choice of colors
not optimal and can be improved to next iteration. Figure 6.6 shows a bar diagram for
these values.
39
6. Second iteration :
Design & prototyping
Table 6.2: Debriefing results for question (Q) one to seven with
test person (TP) one to seven and pilot tester (P) one and two.
1 means strongly disagree which scales up to 5 which means
strongly agree.
Q1 Q2 Q3 Q4 Q5 Q6 Q7
TP1 5 3 3 3 4 3 4
TP2 2 4 2 4 3 4 5
TP3 5 2 4 1 5 2 2
TP4 5 1 4 2 4 4 4
TP5 5 1 4 2 4 4 1
TP6 4 4 5 2 5 5 5
TP7 4 2 4 1 5 5 4
P1 5 2 4 2 5 4 4
P2 5 2 4 2 5 4 5
Mean 4.44 2.33 3.78 2.11 4.44 3.89 3.78
Navigation
All users understood right away that they had to use the arrow buttons to navigate through
the menus and used them correctly. Several users commented during the testing that the
navigation through the menus felt fast and smooth. The first time the users came across
that they had to go back a step in the menus was in scenario 1 where they wanted to go back
from the Camera menu to the Main menu. None of the users pressed the correct button
(Exit) on the remote control on their first try. Three users pressed the correct button on
their second try, two on their third, two on their fourth and one on the fifth, see figure 6.5.
In the Camera menu a few test persons expressed that they felt that it would be suitable
with more hints for how to go back to the Main menu. One test person said that the Select
button should send back to Main menu from the Camera menu because the test person felt
that Return or Exit would exit the menu without saving the cameras PTZ configurations
that were made.
After scenario 1, once the users had found that Exit was the correct button, the users
continued the usability test by going back a step in the menus in a correct way. During
the retrospective probing the users were asked why they think they failed to go back to the
main menu from the camera menu in scenario 1.
Information
The information that was not existing in the prototypes was how to return. There were also
missing information about who you are connected to during call. Other things that were
missing were the Axis look n’ feel in the design. When muting the call it was not clear if
it was the local or the remote microphone (the microphone on the other side of the call)
that was muted.
40
6.2 Result
Figure 6.5: The buttons pressed by the users in their first en-
counter of having to go back a step in the menus. Each row repre-
sents the series of buttons pressed by a user in chronological order
from left to right.
Feedback
Many test persons found the prototype to be lacking feedback when changing language, the
only thing that changed were the language symbol in the top left corner which switched
to the selected language. Two test persons said that they were expected to be sent back
to the settings menu when they changed language which the prototype did not do. More
feedback where desired regarding the mute function during a video conference session,
especially when the menu was not present.
Terminology
Generally the users did not find the terminology confusing and felt that the sentences was
of appropriate length. One user found it unclear if the mute function in a video conference
call muted their own microphone or the microphone on the other side of the conference
call.
Colors should be more, exciting, modern and warm. A problem that occurred was that
the help menu was not present when it was felt that it was needed. It was never used for
during the tests. One test person commented on the Camera view that it is not clear if the
changes are saved. The mute/unmute icon and text were contradictory.
41
6. Second iteration :
Design & prototyping
6.3 Conclusions
5
4.44 4.44
3
2.33
2.11
2
1
Q1 Q2 Q3 Q4 Q5 Q6 Q7
Figure 6.6: Mean value of the results from the 1 (strongly dis-
agree) to 5 (strongly agree) scale questionnaire in appendix D,
these are the same results as in table 6.2. Higher value is better
except for Q2 and Q4 where lower is better.
Navigation
One recurring error that was found in the assessment testing regarded the users ability to
go back a step in the menus. As seen in figure 6.5 all users tried the Return button on
their first or second try except one user which pressed the correct button on the second
try. Based on the result it is clear that the correct button should be the Return button. The
reason why the Return button has not been used previously is due to technical limitations.
Return does not map to a CEC UI Command Code like Exit does but is a vendor specific
implementation which falls outside of the scope for this thesis.
Information
The results showed that all users had trouble knowing how to return, see figure 6.5. There
were also missing information about who you are connected to during a call. Other things
that can be improved was to give the prototype a more Axis look n’ feel, by fonts and colors
that are typically used in Axis products.
Terminology
The terminology in the prototype is good and does not need to be changed for the next
iteration.
42
6.3 Conclusions
43
6. Second iteration :
Design & prototyping
44
Chapter 7
Third iteration :
Graphic design & high level prototypes
This chapter contains information about the method used in iteration 3 and the final design
results. The results are presented in form of digital prototypes. The chapter is formed in
the same way as chapter 1 and 2.
The third iteration is the last iteration in the design process for this master thesis. The goal
for this iteration was to create a HI-FI prototype with Axis look and feel. All the results
and conclusions from the usability tests in iteration 2 were considered.
7.1 Method
7.1.1 Design
Important for this iteration was to not change anything that we did not have any base to
change because there is a chance to introduce new usability problems. The changes for
this iteration was based on previous usability test results and conclusions from iteration 1
and 2.
45
7. Third iteration :
Graphic design & high level prototypes
7.2 Result
The main view from iteration 3 seen in figure 7.2 is the interface with Axis look n’ feel,
only three menu selections and a gray text that explains the marked menu choice. The
help menu has been removed. The intuitive icon that shows that the sound is muted can
be seen in figure 7.4, which is a view that is shown during call, gives its information
without being in the way for the user. The indicator in right top side of the view is the
same as in iteration 2 but the colors has been changed. The left indication about how the
user navigates back to the main menu is complimented by a symbol which is the same as
the button that is supposed to be pressed on the remote control. This symbol may differ
depending on TV model but information about the TV brand can be fetched with CEC
and the displayed symbol can be adjusted to fit the remote control. The main difference in
the Settings language view figure 7.5 is that the user is moved to the previous view when
selecting language, this is to give the user feedback. An interaction design advantage we
added to the contact list view figure 7.6 were that there are two items showing down or up
except when the scroll reaches the end of the list. The intuitive scheme in the top of the
view with all the letters shows in what letters there is conference rooms and what letter is
in focus right now. This scheme is a short cut for faster navigating when there is a lot of
conference room.
46
7.2 Result
47
7. Third iteration :
Graphic design & high level prototypes
48
Chapter 8
Implementation
The implementation chapter describes the setup and implementation of the final prototype.
This chapter also covers technical restrictions.
The basic architecture of the program is made up by different states and in them different
views. The program listens to CEC messages and when received changes the states and
views accordingly which alters the GUI. For PTZ control Axis’ own open API, VAPIX,
was used.
49
8. Implementation
Figure 8.1: Overview of the initial setup for one end of the system.
Figure 8.2: Overview of the end setup for one end of the system.
50
8.2 CEC
8.2 CEC
The Remote Control Pass Through feature was used when communicating between the
TV and the camera. The reason for not utilizing the Vendor Specific Commands feature
was that a solution which would work on any HDMI-CEC TV was desired. Initially a LG
TV was used in the setup but was later switched to a Samsung TV since the SimpLink
(LG CEC trade name) seemed to be using Vendor Specific Commands more for button
presses while Anynet+ (Samsung CEC trade name) used Remote Control Pass Through.
Some Samsung remote control buttons were still not supported in Remote Control Pass
Through. The Exit button is one example and was therefor not used which was found to
be the source of the navigation problem in the second iteration, shown in figure 6.5.
8.3 Overlays
The process of generating an image in the Axis camera basically consists of four steps,
seen in figure 8.3. These steps are made in the ARTPEC 5 module, (2) in figure 8.4. In
the first two steps the image is captured and processed to enhance quality. In the third step
the overlays are added to the video stream over the captured image and in the last step the
data is encoded. Since the scaling is made before the encoding, the overlay images can not
be in a compressed file format, instead 24 bit BMP was used.
Along with the captured image, the video stream can be seen as three layers. The other two
layers are the overlay image and the overlay alpha. The overlay alpha determines which
part of the overlay image that will be transparent and show the captured image underneath.
The alpha is determined by a unique color in the overlay image, in figure 7.3 the white color
is the alpha. Not all overlays overlays had transparency, e.g. Main menu, figure 7.2, where
the overlay image completely covered the image captured by the camera.
51
8. Implementation
to other resolutions than the pre-generated overlay image, especially for text which could
become unreadable. An alternative solution for this could of course be to have multiple
overlays for multiple resolutions for each view, but this would require a lot more overlays.
Figure 8.4: The different parts in the camera (oval) and how they
are connected.
As previously mentioned, the communication from ARTPEC 5 to FPGA is not yet fully
implemented at the time of this thesis. It is still possible to send the captured image from
the local camera to the TV via HDMI, (4 → 5) in figure 8.4, but not with the added overlays
since that part is done in ARPTEC 5. By sending the video stream to a separate computer
through the ethernet port (1), this solution circumvents the issue. As the Axis camera was
still in development during this thesis, technical limitations like these should be expected
and should not be considered a bug.
52
Chapter 9
Conclusion and discussion
This chapter presents different aspects of the project, future perspective, alternative solu-
tions and general discussion.
We have in this project created a video conference prototype with an intuitive interface
combined with a standard TV remote control for a video conferencing system using Axis
hardware and software and standard AV equipment. We also did a market overview where
we compared similar products which where included in the first goal. The market overview
could be done in a more effective way but due to the fact that some equipment where hard
to test without buying we read about some systems instead of testing them ourself. The
second goal about creating a bidirectional system where not fulfilled completely. Due to
technical limitations some functionalists where simulated with a computer in the middle
but this problems where outside of our project and will eventually be fixed. Also the instal-
lation of the camera where excluded and put in a fictive cloud solution. After the usability
test we can be quite positive that a huge majority of the target group will feel confident
using the video conference system.
The CEC solution in this thesis has some major benefits, the two biggest being not having
to provide additional equipment, like a separate remote control, and the use of a standard-
ized protocol that, in theory, should make the system interoperable. We found however
some drawbacks to this approach regarding both design and implementation. The CEC
feature does not provide for as much interoperability as we first hoped for. With different
CEC trade names using different vendor specific commands, vendor specific implementa-
tion might be required by Axis as well. How many TV brands must a CEC based video
conference product support to be sufficient? Do all vendors use their own unique com-
mands? The different implementations of CEC from different vendors was only briefly
looked into during this thesis, a more thorough study on this will be required to be able to
answer these questions certainly. Is it possible to design a user interface good enough to
work well with high usability, regardless of the layout of the remote control? With no real
53
9. Conclusion and discussion
influence on the design of the remote control, this is basically the task. Most TV remote
controls are similar and we can of course expect that the companies strive to design remote
controls with high usability. But even when this succeeds, it is still a TV remote control,
not a video conference remote control. Can we expect a product with high usability for
its intended field to still be intuitive to use outside of the field? During our testing we did
not find any correlation between a better result for experienced Samsung users who has
the same type of remote control at home compared to the other testers. With CEC it is
possible to find the brand of the TV and tailor the user interface for that company’s remote
control. Could a dynamic interface which adapts to the TV and its remote control be the
solution? That is possible, in our design we believe we achieved better results by restrict-
ing and adapting to a specific TV remote control as many testers justified their choices of
button presses by the icon mapping between the remote control and the GUI. Note that the
user isn’t actually restricted to the standard TV remote control though, a generic remote
control could be used. Even though higher usability might be achieved for the individual
case, a user of multiple products connected to different TV brands, might find it confusing
if the user interface differs since it is same video conference product.
The amount of testing we did was sufficient according to usability experts, such as J.
Nielsen. We even tested on more test users than suggested (5) [24]. When determining the
suggested number of testers, Nielsen not only looked to find as many usability issues as
possible but also took the test budget regard. The test budget in our case can be seen as the
time we spent on the testing, the preparation and the evaluation of the test results. We felt
that we gained a lot from our elaborated tests with many test participants, but according to
Nielsen it is more favorable to test on fewer users if it leads to more tests. A rule of thumb
here might be to strive for many tests and on a few testers rather than a few tests on many
testers. For our testing we felt that we wanted to be well prepared, and if possible test
on more users to get as many different views and thoughts on our prototype as we could,
which we think we did successfully. In hindsight, testing more scenarios or adding other
types of tests instead of more testers when we had spare time could have given us a better
result. That said, we are still satisfied with our result.
54
9.3 Suggestions for future development
that you do not need one more new unique remote control. There is a possibility to both
skip using the TV remote control and also not have a own unique camera remote control
by having a application in your smart phone where you can sign in and control the camera.
It is also a opportunity to both have a smart phone application and the TV remote control.
The TV remote control limits the video conference user to only use an TV screen while in
fact there often is need for a bigger screen than that.
Usually meetings are setup on/added to the participants digital calendars. There are al-
ready standards for exporting and importing such appointments, this can be an possible
extension of the project. When a user comes to a conference room it could already know
which other conference room the user is going to call and provide the user the option to
call that conference room with one button press.
There is a opportunity to create a personal interface which requires sign in. The guidance
that is should be very easy to get started made us go for a solution without a personal
touch but there is of course pros with personal views. First of all there will be possibility
to configure the settings and save it depending on the users own requirements.
The contact list is design so you first have to choose a country then city and last a confer-
ence room all in alphabetic order. One can imagine a solution where the contact list gets
less steps and the contacts instead are grouped by geographical orientation or by starred
contacts.
55
9. Conclusion and discussion
56
References
57
REFERENCES
[14] Jakob Nielsen. Parallel & iterative design + competitive testing = high usability.
http:// www.nngroup.com/ articles/ parallel-and-iterative-design/ , January 2011.
[Online; accessed 21-Jan-2015].
[17] Donald A Norman. The design of everyday things. Basic books, 2002.
[24] Jakob Nielsen. Why you only need to test with 5 users.
http:// www.nngroup.com/ articles/ why-you-only-need-to-test-with-5-users/ ,
March 2000. [Online; accessed 16-Feb-2015].
58
REFERENCES
59
REFERENCES
60
Appendices
61
Appendix A
10 Usability Heuristics for User Interface
Design
Below is Jakob Nielsen’s 10 general principles for interaction design. He calls them
"heuristics" because they are broad rules of thumb and not specific usability guidelines
[45]:
Error prevention
Even better than good error messages is a careful design which prevents a problem from
occurring in the first place. Either eliminate error-prone conditions or check for them and
present users with a confirmation option before they commit to the action.
63
A. 10 Usability Heuristics for User Interface Design
64
Appendix B
Script for usability testing iteration 1
Presentation
• You are about to use this conference camera for the first time.
• You are in a conference room where the camera will be permanently placed.
• The camera is connected to the TV with a HDMI cable and connected to the internet
with an ethernet cable.
• In your hand you are holding the remote control for the TV, in this test you will be
interacting with the video conference system using this remote control.
Scenario 1 You have recently bought a new conference camera and want to start using it
by getting to the main menu after you have plugged in all the cables needed.
Scenario 2 You have the camera on the shelf and you want to change the cameras PTZ
position using the remote control.
Scenario 3 Your conference system is by default used with English and you want to change
it to Swedish.
Scenario 4 You want to customize your video conference experience by becoming an ad-
vanced user. Remove the nestled window in you view that shows yourself.
Scenario 5 You want to initiate a video conference call with the persons that are located
in the room "Axis Herkules".
65
B. Script for usability testing iteration 1
Scenario 6 You are in your conference room and are waiting for a call from "Axis Hippo-
dromen". Answer the call.
66
Appendix C
Test plan iteration 2
Purpose
The purpose of the usability test in iteration 2 is to receive useful feedback from the real
target group on computer made prototypes. The questions we want to get answers to is
for example: By fixing issues from iteration 1 have we introduced new problems? Is
there a difference now when we are doing a computer made prototypes instead of paper
prototypes? Is it something that the real target group finds good or bad in contrast to the
master students in usability and design from iteration 1? Is the interface easy to use for
the whole intended target group? Is there any part of the interface where the user becomes
uncertain or cannot understand how to follow?
Scope
The testing will be performed on computer made prototypes of the Axis Communications
Utility cameras video conference mode. The prototypes are from iteration 2 of 3 and will
consist of the whole overlay interface system, the cloud service will not be included.
Environment
The test will be executed 2014 week 45 and 46 at Axis Communications buildings at
Emdalavägen in Lund. The number of test participants and sessions that will be held
depends on how many that can volunteer their time for us. The vision is to held the test in
conference rooms in the H building.
Session
The test will be 20 minutes long each and afterwards it will be free discussion for 10
minutes and here is it after that it will be possible for for the test leader to reset the test
67
C. Test plan iteration 2
environment for the next test person. Every test session will be 30 minutes long and the
test persons will come every forty fifth minute which means 15 minutes for resetting the
environment and briefly review the session.
Equipment
The test will contain of table with room for at least four seatings, laptop preferably 15
inches and the operating system should be Linux, camera that can document the test, time-
keeper, paper and pencils for the note taker.
Participants
The test persons will be recruited mainly by our supervisor at Axis Communications and
everybody will be from our target group. Our target group is employees at Axis Commu-
nications who might want to have video conference. For the test we would prefer to have
eight test persons but five is the minimum number we want to test to get the most out of
the usability test.
Scenarios:
1 Ändra kameran så du är i bild och gå därefter tillbaka till huvudmenyn.
4 Stäng av mikrofonen så att personen i andra änden av samtalet inte kan höra vad du säger.
Method
The test will be held in the environment described above, it is the real video conference
environment that is pursued. The test leader brings all the necessary equipment for the test.
First the test person will be greeted welcome and the outline will be described. Afterwards
the test will start and proceed scenario for scenario. Afterwards a debriefing will be held.
Roles
The test will be held by the test leader who will integrate with the testperson but only if
necessary, afterwards a short discussion will occur. The test also contains a note taker that
will write down everything said during the test. The test leader will as well take time if so
needed.
68
Data collection
The primary data collection will be qualitative data that the note taker wrote down and
later analysed. Every test session will be taped which means that the analysis can be done
on that material as well. Another form of data is from the debriefing when the test person
with his or her own words can give suggestions for improvement in either a interview or a
survey.
69
C. Test plan iteration 2
70
Appendix D
Debriefing survey iteration 2
I think that I would need to ask questions to know how to use this system.
1 2 3 4 5
I think that most people would figure out how to use this system very quickly.
1 2 3 4 5
71
D. Debriefing survey iteration 2
Terminology confusing?
72