User testing the CBeebies website: a more child

Transcription

User testing the CBeebies website: a more child
MA Early Childhood Studies
Roehampton University
Olivia Dickinson
User testing the CBeebies website:
a more child-centred approach?
This dissertation is submitted in part fulfilment of the
MA Early Childhood Studies
Roehampton University
September 2006
Abstract
This study investigates the feasibility of a more child-centred approach when user testing
the CBeebies website. The focus of the study has been on the web production team and
how we can work together to improve the user testing we do. In order to find out if a
more child-centred approach is feasible, I investigated how the team understand the
principles and purposes of user testing and how they understand the young children who
use the CBeebies website.
The literature review shows that there is very little literature which focuses on how young
children use the internet or how they use a computer. Observation and assessment
methods in early years’ settings, and literature about children’s minds and empowering
young children are more useful for moving towards a more child-centred approach.
The research methodology was qualitative. I interviewed the web production team and
then conducted two group discussions to find out how to improve user testing and explore
the feasibility of a child-centred approach.
The study finds that the team understand the criteria for a successful user-testing session
but often those criteria are problematic when user testing with young children. Their
understanding of the audience is from a Piagetian child development perspective, which
hinders the use of methods to empower children in the user-testing sessions. While a
child-centred approach is feasible it is not to the exclusion of other approaches.
A direct outcome of the study has been a working group which is establishing a new way
of user testing the CBeebies website. The working group will adapt some child-centred
theories for user testing but still value the existing ‘theories in use’ (Argyris and Schön in
Moon, 1999: 40) of the team. This will contribute to the growing body of research about
user testing with children within the HCI community.
2
Contents
Abstract……………………………………………………………………………………2
Acknowledgements………………………………………………………………………..4
Introduction………………………………………………………………………………..5
Literature Review………………………………………………………………………….9
Research Methodology…………………………………………………………………..25
Discussion: Analysis and Evaluation…………………………………………………….36
Conclusion……………………………………………………………………………….56
Bibliography…………………………………………………………………..................61
Appendix A: final questionnaire for the interviews……………………………………...65
Appendix B: example of bolding key themes in interview write-up…………………….66
Appendix C: write-up of interviews for Liz and Debra………………………………….69
Appendix D: ideas sheet for second group discussion…………………………………...72
Appendix E: list of themes and groups for second group discussion……………………74
Appendix F: second group discussion write-up………………………………………….75
Appendix G: final project definition document for user testing working group……..….79
3
Acknowledgements
Thanks to all those at BBC Children’s who have participated, contributed or supported
me:
Helen Baker
Rachel Bardill
Emma Bullen
Christina Colbeck
Sandrine Dordé
Stephanie Gauld
Nicola George
Eimer Heekin
Sue Howes
Katherine Monk
Aidan O’Brien
Joanne Patterson
Alan Robinson
Rebecca Shallcross
Helen Stephens
Debra Trayner
Cecilia Weiss
Liz Wilton
Many thanks to Gee-Kay Wong and Ana Joly for stimulating and helpful conversations.
Thanks too to Debra Trayner, Liz Wilton, Leo Sen, Philip Buckley, Stephanie Gauld,
Dave Howell and Lewis Ward for conversations over the last year or so that have helped
me to refine my ideas.
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Thank you to those at Roehampton University who have inspired and supported me
throughout the MA: Elise Alexander, Peter Elfer, Hiroko Fumoto and Shirley Maxwell.
Thank you particularly to Elise for her judicious comments at all times.
Thank you to Deborah Susman for her thorough proof-reading.
My greatest thanks to Stephen – thank you for being my 21st century helpmeet.
4
Introduction
I am a producer for the CBeebies website at the BBC. The CBeebies website
(bbc.co.uk/cbeebies) offers interactive games, stories, songs and other online activities
for children aged between eighteen months and six years. I have worked for CBeebies for
over four years and as part of my job, my colleagues and I have done regular user testing
with the target audience for CBeebies’ interactive services (for online and interactive
television). When we go user testing we visit local nurseries and schools, and
occasionally homes, and watch young children as they play on the CBeebies website. We
also sometimes user test with parents and children together, and occasionally parents
alone. The main goal of the testing has always been to ensure that the content we are
creating is of the highest quality and puts the user (the audience of young children) at the
centre of the interactive experience. The positive feedback we receive from the usertesting sessions often confirms the high quality of the content while any negative
feedback or observations are used to improve content. The many users of the CBeebies
site who return again and again to play online with CBeebies show how worthwhile and
useful the user testing is.
However, over the last two years I noticed that the user testing we were doing was no
longer as satisfactory as it had been in the past, both the user-testing sessions themselves
and the results we were getting. It was often hard to quantify what we had found, there
was a sense of frustration in some of the sessions about how we interacted with the
children and some of the team commented that they were no longer learning anything
new when they went user testing.
From March to May 2005 I observed a baby using the Tavistock Clinic’s method of
psychoanalytic infant observation (Rustin, 1989) as part of the Children Under Three
module for my MA in Early Childhood Studies. I observed Mara on four separate
occasions from age three-and-a-half weeks to eleven-and-a-half weeks. I found my
observations of Mara to be rewarding, offering a contrast to the observations of children
under six for user testing that I was finding frustrating. This inspired me to rethink how
5
we observe children when we go user testing. I wanted a user-testing session to be as
rewarding and informative as one of the observations of baby Mara. As user testing was
often the team’s only chance to be in touch with our young users I was keen to ensure
that contact and experience was as positive and rewarding as possible.
Would an observational model like the one from the Tavistock Clinic, which was
designed specifically for observing babies and children under two, be more suitable for
our user-testing sessions than the methods we were currently using, based upon
user-testing methodology for adult users? Moreover, might other methods of observation
and assessment designed specifically for young children and early years’ settings (Abbot
and Gillen, 1997; Sylva, Roy and Painter, 1980; Pascal and Bertram, 1997; Carr, 2001;
Clark and Moss, 2001) be more suitable as well?
In the summer of 2005 I set out to find out how user testing for young children under six
could be changed and maybe improved. I had originally hoped to try out different
methods of observation and assessment for the early years, alongside the team’s more
orthodox user-testing methods. However, user testing cannot be conducted alone, it is
always a team exercise as you need at least one facilitator and one observer. Even though
I wanted to try out new observation methods that might change the roles of the traditional
facilitator and observer, I knew those methods would not result in user testing becoming
a solitary activity. Moreover, whatever I might find out alone it was essential to share it
with the team to ensure that the new method was understood and practiced by everyone. I
therefore realised early on that the main focus needed to shift from adapting different
methods to working with the team to come up with a new way to approach user testing. I
wanted to improve user testing not only as an experience for the children with whom we
user test, but also for the CBeebies web production team themselves. Once I could
establish how the team understood both user testing and the audience of young children, I
could work out with them the best approach for a new way of user testing. While my
reading and ideas from my Early Childhood MA were informing my thinking,
particularly around listening to young children and empowering them, I wanted to work
6
inclusively so that all the team’s opinions were heard; it was not just about listening to
the children but listening to the team too.
At times I have wondered if it would have been simpler to have dictated the new process,
using all the research I had done and knowledge I had gained, rather than democratising
the process so that the team were consulted at every step. This became particularly
pertinent towards the end of the project when Ana, a student completing an MSc in
Digital Television, joined my department at the BBC on work experience. Ana’s
particular focus was on young children’s use and response to interactive television. While
at BBC Children’s she developed some prototypes for the CBeebies interactive TV
service and then tested them at her university’s user-testing lab with children aged four
and five. I have found it inspiring to share ideas and new ways of doing user testing with
young children with Ana, as well as refreshing and reassuring to compare our
experiences. Ana has worked completely alone in developing and testing the prototypes,
which contrasts to my approach. As an outsider to the department who is only with us
temporarily she does not need to consult at every stage and make sure everyone involved
agrees with what she is proposing. She has done a finite piece of research that will
contribute to the next development of the interactive TV service for CBeebies and along
the way has highlighted different methods of user testing with young children.
Denscombe refers to how the participatory aspect of action research democratises it in
some way, showing ‘respect for practitioner knowledge’ (2003: 77). At the very
beginning of the project I expected to carry out action research. That was ultimately not
practicable, but through the paradigms I have worked within I have retained that
aspiration to action research. Those paradigms include my belief that no researcher can
participate in society without adhering to a particular paradigm or world view,
consciously or not; that theory informs practice and practice informs theory; and that
everyone’s experiences are valid and should be valued, through the feminist paradigm of
equality between men and women. Moreover, my approach of listening to the team about
their experiences and trying to see how those experiences can feed into the theory and
practice of the user testing we do with young children reflects the child-centred
7
methodologies that have been so crucial to my overall project. Listening is at the basis of
the work of Donaldson (1978 and 1987), Tizard and Hughes (1986), Paley (1986), Carr
(2001) and Clark and Moss (2001). I have not yet had the chance to apply those listening
methods to the user-testing sessions in order to empower young children but I hope by
having listened to the team I can empower them to go away and listen to the children
themselves.
8
Literature Review
Introduction
The literature review covers four different areas of knowledge and research, all of which
contribute to an understanding of what might be needed to improve user testing with
children aged six and under. I look at definitions of usability and user testing, and how
they are both applied to adult users of different interfaces and websites. I then review the
literature that is specific to children and usability within a web and information and
communication technology (ICT) context. There is a long tradition of observation and
assessment in early years’ settings, some of which originally inspired how CBeebies user
testing could be improved, so I selectively review the relevant literature, comparing and
contrasting different approaches and their long-term goals. Finally, I turn to the children
with whom we user test and so in my fourth area I review literature and findings about
children’s capabilities, their minds, how they learn, and how that has been used to
empower children.
Literature about usability and user testing within web design context
Usability within the context of ergonomic requirements for office work with visual
display terminals has been defined since 1998 under ISO 9241-11 as ‘the extent to which
a product can be used by specified users to achieve specified goals with effectiveness,
efficiency and satisfaction in a specified context of use’ (UsabilityNet, 2006). This is a
definition which has been used by Human-Computer Interaction (HCI) researchers in
usability studies (see Sim, MacFarlane and Read, 2006: 237). However, there is also a
slightly more recent definition within software engineering under ISO 9126-1 that
combines usability as being essential for interface design and to meet users’ needs (ISO,
2001; UsabilityNet, 2006). Usability under the second international standard is defined as
‘the capability of the software product to be understood, learned, used and attractive to
the user, when used under specified conditions’ (UsabilityNet, 2006). This second
definition is most familiar to web developers and designers who practise user-centred
design, always thinking about the needs of the user when creating a website. Both Krug
and Kuniavsky sum this up very simply:
9
‘Usability really just means making sure that something works well: that a person
… can use the thing … for its intended purpose without getting hopelessly
frustrated.’ (Krug, 2006: 5)
‘Ultimately, usability is good design.’ (Kuniavsky, 2003: 20)
Thus usability is designing something (in the context of this study, a website or an online
game or application) that can be used easily by the people for whom it is made (again, for
this study, children aged six and under).
In order to ensure that a website is working well and can be understood, that it can be
used and is also attractive to the user, user testing or usability testing (the two terms are
interchangeable) should be carried out during the development of the site and once it has
been launched for public use (or ‘gone live’). Nielsen identifies just three components of
user testing: representative users, users to perform representative tasks and people from
the website to observe what the users do without helping them in any way (Nielsen,
2003). You can do the user testing in a usability lab with a one-way mirror, cameras and
microphones but it is not essential: it can equally be done in a home or office setting with
a few users and at least one observer. Kuniavsky (2003) covers other ways that you can
get feedback about the usability of a site or product and users’ needs, such as cardsorting, surveys, focus groups and usability diaries. Even with the detail of the different
approaches from Kuniavsky, this all sounds very simple and achievable. However, is it
that simple when user testing with young children?
Both Nielsen and Kuniavsky are writing about user testing with adult users. Kuniavsky
refers fleetingly to teenagers twice (Kuniavksy, 2003: 112, 532) but does not discuss how
user testing might need to be adapted for teenagers, let alone children under thirteen or
much younger children aged six and under. Nielsen has done one usability study on how
children aged six to twelve use websites (Nielsen, 2002) which is discussed below. Krug,
a third example of a ‘usability expert’ in the world of web development and HCI, does
not refer to children or teenagers at all, though he does stress that it is a good idea to get
10
to know your users and test with them, and that it is important not make assumptions
based on your (adult) experiences of being a web user (Krug, 2006: 125-129). I now turn
to specific literature about children’s ICT and internet use in order to find out more about
children and usability.
Literature about children and usability within a web and ICT context
Literature about how children in the early years in the UK use computers is based around
what they learn from their encounters with ICT and how it can benefit or delay their
development. While it is important to review literature about young children and
usability, there are not many written accounts of such research studies; very little
literature focuses on how pre-school children use the internet or how they use a computer.
Therefore I begin in this section by examining the general guidance for early years’
practitioners when planning children’s learning around ICT and then look at how a few
researchers have tried to apply that guidance (from the QCA) to young children’s use of
different forms of ICT at home and at nursery.
‘Curriculum Guidance for the Foundation Stage’ takes an umbrella approach to ICT –
children are expected to develop technological literacy, gain information and
communication skills from ICT and start to understand control technology (SirajBlatchford and Whitebread, 2003: 1; QCA/ DfEE, 2000). Within the general area of
learning ‘Knowledge and Understanding of the World’, specific areas of learning about
ICT range from ‘stimulate all children’s interests in ICT and other technology’ (QCA/
DfEE, 2000: 93) to ‘complete a simple program on the computer and/or perform simple
functions on ICT apparatus’ (ibid: 92). In their review of international research evidence
examining ways in which ICT is used in pre-school settings, Plowman and Stephen
(2003: 150) allude to this ‘broad definition of ICT’ but highlight that ‘most of the
literature currently focuses on ICT defined as computers’ and that there is ‘a scarcity of
good quality research findings on using ICT with pre-school children’. Brooker and SirajBlatchford support this, commenting that ‘there is still a relatively small body of
empirical evidence on the effects of their use by children aged three and four’ (2002:
253).
11
McPake, Stephen, Plowman, Downey and Sime have tried to go some way to correcting
this, looking at the home experiences of pre-school Scottish children using ICT (McPake
et al, 2006), covering interactive TV, DVD players, home internet-use and video games.
However that study is still only looking at how ICT in a general sense can benefit a
child’s learning, not at how a child uses a particular ICT application at a given time (or
even how that application can be made better for the child). Brooker and SirajBlatchford’s study of computer-use in an inner-city nursery (2002) looks at how children
interact with a particular ICT application over a given time (the nursery computer over
four months). However, as with McPake et al, the study is most interested in the effect
the computer and other ICT has on the children’s development and learning, as well as
whether home ICT experiences are an advantage or disadvantage for then using the
nursery computer.
The literature therefore suggests that ICT is viewed only as something to be used with a
child, to benefit the child or to see how it affects a child’s cognitive, linguistic or social
development. As stated above, there is little that goes one step back and focuses on how
pre-school children use the internet or how they use a PC. Even though some of these
researchers may be looking unwittingly at the usability of certain ICT applications and
software, their focus is not on how to make the ICT better for the child or how to conduct
a study to find that out. For literature about how children use the internet or a computer
and how to conduct research around that, we need to look at usability studies of slightly
older children, some over the age of six but most of at least age eight.
Researchers at Microsoft have published guidelines for usability testing with children
(Hanna, Risden and Alexander, 1997 and Hanna, Risden, Czerwinski and Alexander,
1998). In both of these articles the researchers discuss ways to adapt traditional adult
usability testing for children. However while they do address specific issues for children
under six they rarely focus just on that age group: the guidelines cover children from age
two to fourteen. The guidelines are also specific to Microsoft – they assume the provision
of a usability lab and suggest ways to make the environment more appealing to children
12
(at the BBC I do have access to a usability lab but it is more common to user test with
children in their own homes, the office, schools or early years’ settings). In the second
article (Hanna et al, 1998) the focus is on how to make usability central to the design
process of creating computer products for children. Thus while they give age-specific
recommendations and examples (differentiating between three and six-year-olds’
different understanding; reading text alone is not age-appropriate for four-year-olds;
asking parents to rate products as well for the under-fives; card-sorting is only
appropriate for children of eight or over) the emphasis is still on making adjustments to
traditional user-testing methods for all children, not just those under six. This is probably
because one aim of the article is to persuade product development teams of the
importance of having usability engineers working alongside them. So they play down
how hard it might be to user-test with children: ‘children can participate in traditional
laboratory usability tests with only minor adjustments in procedure’ (1998: 7).
Markopoulos and Bekker (2003b) take issue with the advice from Hanna et al (1997) as
‘it is not very clear on what evidence it is based and thus we conclude it describes the
personal experience of the researchers involved’. While personal anecdote can be useful
as a starting point for anyone setting out to do usability testing with children, the main
limitations of their advice is that they give the impression there is no need to delineate
between the different ages of children. The first article (Hanna et al, 1997) promises
much, setting out sections about children from pre-school, elementary school and middle
school but ultimately the authors do not follow this through and give only cursory
examples of different ages, sometimes not being clear to what age they are referring.
Nielsen’s usability study on how children use websites supports how important it is to
delineate between the different ages. His study (Nielsen, 2002) user-tested websites with
children aged six to twelve. This means he has missed the age of CBeebies users,
children aged six and under, but some of the findings around the youngest children in his
group of users should surely be illuminating. However, while he finds that ‘extensive text
was problematic for young children who are just beginning to read’ (2002: 2), when he
then summarises the differences between testing with adult and child users he finds that
‘half of our young users were willing to read instructions’ (2002: 3). There is obviously a
13
problem with grouping all primary school-age children together – presumably the half
who wanted to read instructions came from the older half. Nielsen himself highlights how
grouping children together can be problematic, as there were big differences between
boys and girls in the testing (it was mostly girls who were willing to read instructions).
As with the Microsoft articles, Nielsen’s study is helpful as a starting point but is not
specific enough for doing usability testing with children under six.
Markopoulos and Bekker (2003a) highlight how ‘limited methodological advice has been
published to date regarding the adaptation of usability testing procedures for children of
different ages … considering the importance of the topic and its practical utility there are
surprisingly few research results published to guide the practitioner’ (2003a: 146). The
guidelines published by the Nielsen Norman Group for websites for children following
Nielsen’s usability testing with children aged six to twelve (Nielsen, 2002) are a ‘sign of
maturation for this field’ (Markopoulos and Bekker, 2003a: 149). They also credit Hanna
et al with both the articles already cited (1997 and 1998). However Markopoulos and
Bekker’s emphasis is on how usability testing needs adapting for the particular needs of
children, as well as other special target user groups.
Druin’s research into the role of children in designing new technologies has led to the
setting up of a scheme at her HCI lab at the University of Maryland where children are
design partners with adult research teams (Druin, 2002: 13). Her review of how the role
of children has changed in the HCI community since the 1970s includes a diagram of an
onion representing the four roles children may have in the design of new technologies, as
user, tester, informant and design partner:
14
(Druin, 2002: 3)
The detailed information she gives of each role, from a historical and practical
perspective, seems promising. In the user testing CBeebies does currently, children would
be seen as user, tester and informant, depending on the stage of development of a game or
website. However, as Druin does not really discuss children under the age of five, this
conclusion is only reached from extrapolation. She does allude to adults observing
children and how it is ‘particularly common for research with young children (ages 4 and
younger)’ (2002: 5) but most of her study and motivation seems to be tailored to children
of seven or above. Moreover, the overall feeling of her work is that it is not specifically
looking at usability and user-centred design for children but at the many different ways
children can interact with different technologies. This is shown by the myriad examples
she gives of research methods for children as users (2002: 5-7). When she discusses
children as design partners, she states that ‘the impact that technology has on children
may not be as significant as the impact children can have on the technology design
process’ (2002: 12). Thus her greatest interest lies in developing the design process and
making sure children are equal stakeholders with adults in the design of new
technologies; the emphasis is on the relationship and process, not the end product. This
applies even for the younger children as the emphasis in the article about Swedish
children aged five to seven who were design partners was on how they ‘changed from
being learners and critics to being inventors and design partners’ (2002: 21). No mention
is made of what they invented or designed. For that reason, Druin’s approach needs to be
treated carefully. While it is useful for CBeebies user testing as it may help to make the
15
process of user testing more enjoyable, in that it focuses more on the child as user, the
user testing still needs to feed into the end product, benefiting all users of the website.
In Sim, MacFarlane and Read’s empirical study of fun, usability and learning in
educational software, with children age seven and eight, they highlight how ‘there is no
clearly established methodology for evaluating software for children’ (2006: 236). Their
approach seems to reflect the studies from Brooker and Siraj-Blatchford (2002),
Plowman and Stephen (2003) and McPake et al (2006). Even though they use
observational methods and a ‘fun sorter’ to assess usability and fun of software, they also
measure the learning effect of the software on children. The software was for Key Stage
One Science, which would be studied by children at the very top end of the CBeebies
user profile. However, the children in the study had already taken their SATs in Science
and so ‘were about one year older than the age range for which the products were
intended’ (2006: 240). This study is of some value in suggesting ways for younger
children to evaluate their experience of using software or interactive products (the ‘fun
sorter’ helps children to map how much fun, learning and ease of use they thought the
software had, (2006: 242)) but does not give much insight into how to tailor a usability
study for children under the age of six. The observational methods they use are just free
note-taking with two researchers, with one observing the child and one looking at
usability issues (with some inevitable cross-over, 2006: 241). The children themselves
are that bit older than CBeebies users, so are used to performing ‘tests’ in a school
environment with an adult who is unfamiliar to them.
Markopoulos, Barendregt, Brus and Plakman (2005) suggest a new way for evaluating
interactive products. Their method uses the parent as facilitator and is most suited to
children aged four to six. While the research is in its early days (‘evidence for the
effectiveness of the method is very small’ and ‘it is quite hard to recruit parents and
children to participate’ (2005: 38)) it at least takes into consideration the importance of
children interacting with products in their own habitat (i.e. at home) and allows the
testing to take place over a few days, thus mimicking how they genuinely might use a
product or website if they were new to it. In a similar way, Barendregt (2006) has tried to
16
find ways to tailor user testing to young children. She tried using picture cards from the
Picture Exchange Communication System (PECS) for children aged five to seven to
identify usability problems. She found that ‘the problem identifying picture cards are a
good addition to the thinking-aloud method’ (2006: 132). The picture cards used in
addition to thinking-aloud methods allow children to ‘express more problems than with
standard thinking-aloud’ (ibid). Barendregt gives ‘some new guidelines for usability
testing of computer games with children’ based on experimental studies, rather than on
the personal experience of Hanna et al (2006: 136). However, while these two studies are
helpful to give hints and tips for how to improve user testing for the CBeebies users, they
are both small scale and neither attempt to user test with children under four years. Both
studies are more useful than Hanna et al (1997 and 1998) and Nielsen (2002) as they are
focusing on how methods need to be adapted very specifically for younger children.
Overall, none of the literature about children and usability focuses on a child-centred
approach specifically for children under the age of six. I now review literature about
observing and assessing children in early years’ settings as it more commonly puts
children at the centre.
Literature about observation and assessment of children in early years’ settings
There are many different approaches to observing pre-school children, all of them
designed specifically for understanding children’s behaviour, emotions and experiences.
They also all have different ways of recording what is observed or what is discovered as a
result of the observations. What links all of them is that they put children at the heart of
the observations. They are also linked through the observations being carried out to fulfil
some form of assessment, whether it be assessing the child, the setting, the staff of an
early years’ setting (or parents and other carers) or all three. This is particularly pertinent
for user testing as none of the three of the child, the setting or the staff is ever being
assessed. Instead, the value and usability of the website or online game to the child is
being assessed. The methods of observation discussed below are helpful to offer new
ways of approaching user testing for children under six, as their focus is on the same age
group of children who use the CBeebies website. This is in contrast to traditional
usability methods which are designed around adult users. However as the majority of the
17
observation methods are forms of assessment they need to be treated with some care
when seeing if they can be adapted or adopted for user testing for children under six.
The emphasis in the Tavistock method of psychoanalytic infant observation is on a
synthesis of behaviour and experience, so that there is always a whole picture of the baby
and its developing relationships, emotions and sensations. Unlike ‘empirical research
methods in child development psychology ... [the observations] do not aim to select
specific attributes of behaviour (perceptual or cognitive skills, recognition or memory, for
example) for discrete study...’ (Rustin, 1989: 55, 54). By then encouraging the observers
to record what they see in a ‘literal and factual’ way (ibid: 52), using non-theoretical
language, interpretation, reflection or analysis is left for later. This method of recording
what is seen is similar to user testing, where observers are encouraged to just write down
what users do and say, without any interpretation. Discussion among all stakeholders
once the results are recorded is the time for interpretation and conclusions. Rustin
highlights how psychoanalytic infant observation contrasts to most other forms of
studying children from the fields of child psychology and child development in that it is
holistic and subjective (focusing on the feelings of the observer and the observed). It also
contrasts to other forms of child observation as its focus is on the child at that moment,
rather than how the child changes or develops over time.
The ‘multi-method’ approach of the research project ‘Educare for the Under Threes’ at
Manchester Metropolitan University focused on very young children (Abbott and Gillen,
1997). That focus on children under three could be helpful for CBeebies user testing, but
the project’s primary aim was to assess the contributions from ‘educarers’ in a setting,
thus ensuring good practice in early years’ settings for under threes. So one of the
researchers observed ‘the interactions between staff and children, children and their
peers, and staff with other adults including parents’ (Griffin, 1997: 36). Thus the ‘multimethod’ approach is looking more at adult-child interaction and is therefore not
immediately appropriate to be adapted for user testing.
18
Sylva, Roy and Painter (1980) explain that in the target child technique from the Oxford
Preschool Research Project ‘observational research is undertaken in vivo as children go
about their everyday activities at preschool’ (1980: 8). It therefore does not find out how
they get on when they are older, unlike longitudinal studies such as Head Start in the
USA, nor does it base its findings on anecdote (as per Susan Isaacs). Target child
researchers note key behaviour of one child every 30 seconds over a 20 minute period,
trying to answer the question ‘does preschool experience matter?’ (Sylva, Roy and
Painter, 1980: 9). By using the research findings to answer that question, the target child
researchers are assessing the setting and staff as well as the child. They want to know
how a child adapts his or her behaviour to being in a pre-school setting and thus the
observations for their study can only take place in that setting. The target child method is
similar to the multi-method approach in that both want to get an evaluation of the child’s
experience at pre-school out of the observations, though the observation techniques
differ. While that means that both methods could be useful for user testing as they look at
a child’s experience ‘in the moment’, they are less useful once it is clear that they are
assessing the setting and staff as well.
The Effective Early Learning (EEL) Project explicitly aims ‘to create a methodology for
evaluating and developing the quality of early learning ... for 3 and 4 year olds in the UK’
(Pascal and Bertram, 1997: 1). The EEL Project evaluated the quality of early learning
within a setting using Laevers’ Child Involvement Scale and Adult Engagement Scale
(ibid: 11), together with Vygotsky’s theories on the social context of children’s learning.
In the EEL Project ‘process’, participants for each setting are expected to ‘improve the
quality of provision’ once they have evaluated their setting (ibid: 11). The Child
Involvement Scale is in the spirit of psychoanalytic infant observation and the target child
approach: ‘it attempts to measure the processes of learning, rather than to concentrate on
outcomes’ (Pascal and Bertram, 1997: 11). So even though the scale is measuring how
well a child learns, it does not then test what the child has learnt. This is similar to what
user testing tries to achieve – observing how well a child interacts with a website or
online game but not necessarily then testing them on what they have experienced. By
using the Adult Engagement Scale alongside the Child Involvement Scale the researchers
19
make clear that they are assessing the staff and thus the setting as well, as the staff’s
interactions with the children contribute to determining the quality of the setting. Thus a
common purpose of at least three of these methods for observing young children is the
assessment of quality.
As Dahlberg, Moss and Pence (1999: 5) point out, ‘in the 1990s, the concept of quality in
the early childhood field ... has been questioned or problematized.’ Both Vernon and
Smith (1994) and Dahlberg, Moss and Pence give potted histories of the concept of
quality, particularly since 1945 and within the realm of early childhood. There is a
growing body of literature that addresses this problem (e.g. Elfer and Wedge, 1996;
Selleck and Griffin, 1996) and it seems intrinsic to any discussion of the value of
assessment and observation. Moreover there is plenty of literature about the value itself
of observing and assessing young children (Drummond and Nutbrown, 1996; Carr 2001;
Drummond, 2003). Most of these discussions focus on the assessment of the child which
brings us back to the question – what or who is being assessed through the observations?
Ackers (1997) highlights how observation and assessment became interchangeable in the
multi-method project: ‘if we are to see observation and assessment as a priority in the
educare of young children then perhaps as educarers we need to consider ... making time
to record, analyse, discuss and share our observations’ (1997: 85). Both Drummond and
Nutbrown (1996) and Carr (2001) highlight that ‘we could not proceed with the “how” of
assessment until we confronted “what” was to be assessed’ (Carr, 2001: xiii). This is not
quite the same as the difference between what or who is being assessed. While Carr looks
what children can do in their day-to-day lives as they become learners and then how to
assess that learning, some of the observation and assessment methods discussed look
more at the difference between assessing a setting, a child and an adult carer or educator.
In user testing, the child is never supposed to be ‘what’ is being assessed, it is always the
website or the online game. This leaves a potential gap around the child – will we ever
find out what children think, know and can do when they use the CBeebies website if we
persist in the usability model of testing the website and not the user? I am therefore
20
interested in exploring children’s capabilities and so now review literature about what
children can do and how we can empower children to use those capabilities.
Literature about children’s capabilities, their minds and how they learn, and how
that has been used to empower children
The main themes in literature that explores how children can often do much more than
we (society, educators, institutions, even parents) think they can, are listening and
language. From Donaldson in 1978 to Carr and Clark and Moss in 2001, researchers have
listened to young children, recorded what they say and then found out much more about
how children think than is sometimes supposed, how children perceive the world around
them and how they make sense of it. Carr and Clark and Moss have taken that one step
further and enabled children to record for themselves what and how they think.
Donaldson rejects ‘certain features of Jean Piaget’s theory of intellectual development’
(1987: 9) when she shows that young children can use language for thinking in ‘real-life’
contexts. Thus she establishes that children under seven are actually quite good at
appreciating the point of view of another person in a literal sense, through Hughes’
adaptation of Piaget’s three mountains task (Donaldson, 1987: 20-23). If the situation
makes sense to the child or uses familiar situations, characters and language (policemen,
naughty boy, hiding) then they find it easier to understand the task in hand and the
alternate point of view. This therefore challenges the long-held view that children are
unable to ‘decentre’ and are highly ‘egocentric’ (Donaldson, 1987: 18-19). Donaldson
goes on to show that children respond when language is changed, for example when a
statement is modified with an adjective in the case of the ‘steps’ between teddy, the chair
and the table (1987: 45-46). Notwithstanding that even when tested on a group of five
graduates, two of whom also have PhDs, I found that Piaget’s test of sub-classes was
only properly understood when McGarrigle’s adaptations for young children were used
(‘more black cows or more sleeping cows’, not ‘more black cows or more cows’,
Donaldson, 1987: 44), Donaldson shows that young children are capable of reasoning
when they can understand what they are being asked. Donaldson looks in some detail at
the demands of school on young children, and the effect on their thinking and reasoning.
21
However the majority of CBeebies users are not yet at formal school and so for the
purposes of this study it is significant enough that she finds ‘by the time they come to
school, all normal children can show skills as thinkers and language-users to a degree
which must compel our respect’ (1987: 121). Therefore, young children’s intellectual
development must not be underestimated when we are observing children as users of the
website.
Donaldson’s findings have been backed up by Tizard and Hughes’ study of 30 four-yearold girls from middle-class and working-class backgrounds at home and school (Tizard
and Hughes, 1986). By studying the conversations the girls had with their mothers they
found that the girls used language to understand new ideas and were persistent and
logical thinkers. These ‘passages of intellectual search’ showed that the young children
were ‘powerful and determined thinkers in their own right’ (1986: 9). Both Donaldson
and Tizard and Hughes are helpful when thinking about the environments in which we
user test (home, nursery, school) and the language we use with the children.
Paley (1986) recounts how she changed her approach as a kindergarten teacher in the
early 1980s to become a listener; she learnt that discussion blossomed when she kept
quiet and did not express her own point of view. From this, she built up a technique for
children to tell their own stories, culminating in the children she taught taking it in turns
to tell stories each day. Paley found a way to empower the young children in her care,
enabling them to find their own voice. She showed too that each child had their own
opinions and fears, something which can often be ignored. Carr (2001) also uses stories
to empower children. Like Paley, Carr has her own story about listening to the children in
her classroom when she was a kindergarten teacher and how that prompted her to develop
an alternative model of assessment to the one usually accepted by students, teachers and
parents (Carr, 2001: 2).
The ‘Learning Stories’ approach Carr has developed as a way to assess children’s
learning enables children to participate in their own learning and assessment. Children
can be fully involved in discussing and documenting what they have learnt, while early
22
years’ practitioners describe children’s learning and make decisions about where to go
next in the child’s learning. The learning dispositions that form the framework for the
Learning Stories are all about children as active learners, as subjects in society: ‘learning
dispositions are about responsive and reciprocal relationships between the individual and
the environment’ (Carr, 2001: 22). A child is a ‘learner-in-action’ (2001: 5) – being
ready, being willing and being able [to learn]. Their learning story is then told through
the five domains of the learning dispositions: taking an interest, being involved,
persisting with difficulty or uncertainty, communicating with others and taking
responsibility (2001: 23).
The Mosaic approach (Clark and Moss, 2001) is similar to Carr’s Learning Stories in that
it finds ways to listen to young children and then document children’s thoughts and
experiences. The approach includes the idea of ‘children as experts in their own lives’
(2001: 6) and advocates a way of listening that allows children and adults to be coconstructors of meaning (2001:1). This approach reflects Dahlberg, Moss and Pence’s
argument that early childhood institutions need to look beyond providing pre-specified
outcomes with a view of the child as an empty vessel and instead look towards a
‘discourse of meaning making’ (Dahlberg, Moss and Pence, 1999). Central to all three
approaches (Carr’s Learning Stories, the Mosaic approach and Dahlberg, Moss and
Pence’s ideas) is the idea that children are capable of being the architect of their own
lives, as long as they are given the means to communicate and record what they do. Thus
for user testing, where the children are sometimes seen as objects to observe, perhaps
even ‘empty vessels’ absorbing the online content, new approaches are possible where
the child’s perspective and opinion can be recorded, documented and discussed (with the
child and the adults).
Now that I have evaluated the literature that contributes to an understanding of what
might be needed to improve user testing with children aged six and under I will conclude
this chapter with the three research questions that I am considering at every stage of the
project. The three questions are:
23
1. What is the web production team’s understanding of the principles and purposes
of user testing?
2. What do the web production team understand about the audience when they go
user testing?
3. How feasible is a more child-centred approach when the web production team go
user testing?
The first research question refers back to the literature about usability and user testing
within a web design context, as well as sometimes to the literature that is specific to
children and usability. The second research question refers to the fourth area of literature,
about children’s capabilities, their minds and how they learn, and how that has been used
to empower children. The second research question also refers at time to the third area of
literature, about observation and assessment of children in early years’ settings. For the
third research question, all of the literature is relevant as it is the key question of my
research project, how adult usability testing methods can be combined with child-centred
methods that empower children to create a new and improved way of user testing for
children aged six and under. The first and second research questions build up to the third
– it is essential I know how the team understand the audience and give them a good basis
of the principles and purposes of user testing before I can even broach the subject of
making the user testing we do more child-centred.
24
Research Methodology
The parameters of my research methodology: qualitative research, the feminist
paradigm and the child-centred view
‘Social theory informs our thinking which, in turn, assists us in making research
decisions and sense of the world around us. Our experiences of doing research
and its findings, in its turn, influences our theorizing; there is a constant
relationship that exists between social research and social theory.’
(May, 1997: 28)
This description from May of the ‘mutual interdependence’ between theory and research
is the starting point for the research methodology I have used. Within May’s explanation
of how ‘we need theory and theory needs research’ (1997: 30) there is the assertion that
no researcher can participate in society without adhering to a particular paradigm or
world view, consciously or not. Thus when I am at work at the BBC I am not overtly
applying a paradigm to the research I do with my team. Yet when I come to analyse it
and decide upon the different methods I am going to use in order to find out different
things, a particular paradigm – the way I view society and the way I live my life – comes
into play.
The overall approach of this study is qualitative; the reasons why that is the most
appropriate approach for the research will become apparent as I explain each step of the
methodology. However, I am also applying a feminist paradigm throughout the research.
The feminist paradigm I am applying is more accurately described as a variety of feminist
perspectives. I will give some examples of different feminist perspectives, such as
standpoint feminism, the critique of disengagement and not separating reason from
emotion, and show how they apply to the research I have carried out. Some of them have
their parallels in child-centred research. All of those perspectives though are not only
linked by feminism but also by the mutual dependence of theory informing practice and
practice informing theory. When I, as a researcher, participate in society I am adhering to
25
a feminist paradigm or world view, consciously or not. Thus those feminist theories
inform my practice.
While in social research the distinction is often made between qualitative and quantitative
methods, Pring’s assertion of the false dualism between those two schools of thought is
useful to understand how and why I have chosen an overarching feminist approach.
Pring’s false dualism argument is that a researcher can be working within the political
arithmetic tradition, which uses a quantitative approach to produce hard data, but that
they do so ‘without commitment to a thoroughly determinist position or to a rejection of
the distinctively personal way in which the social world and experience is interpreted by
each individual’ (Pring, 2004: 97). Thus a quantitative researcher can take a feminist
approach when interpreting results by relating to the subjects of the research and not
insisting upon separating reason and emotion (May, 1997: 20). I am guided by the belief
that theory informs practice and practice informs theory and so I cannot reject theory and
let the facts speak for themselves (May, 1997: 11). Nor can I assume that facts exist out
there about the social world, independent of interpretation, as empiricists might. In that
sense, a qualitative approach may be the only option. But it is important to remember that
within feminist perspectives both quantitative and qualitative methodologies are valid.
While some feminist researchers may reject any form of science as it does not take into
account the many different versions of social reality, there are still feminist scientists and
feminist researchers who apply hard quantitative methods.
The basics of usability are to put the user at the centre of the design of a website or
application. By involving the user in that way, with the emphasis on the user’s
experience, usability methodology tentatively alludes to some of the aims of standpoint
feminism. Standpoint feminism takes women’s experiences as a starting point but the
paradigm is only complete once women are actively involved in the theorising about
those experiences (May, 1997: 22). Nielsen (2002 and 2003), Krug (2006) and others
would be unlikely to see themselves as working within a feminist paradigm; indeed many
feminist researchers would be likely to label usability researchers as working within
‘male-science’, reflecting hierarchical society and methods, with the usability researcher
26
becoming expert in people’s computer and internet use. But by always involving the user
and placing primacy on the user’s real-life experience, usability researchers have taken
one step towards standpoint feminism. Druin’s (2002: 13) emphasis on the user as
design-partner goes one step further, as she favours a democratic and participatory way
for applications to be designed and developed.
Of course, both Druin and I focus on the child user in our research, not on women’s
experiences of websites. However I would argue that by putting the child at the centre of
the research, many child-centred researchers are unconsciously applying different
feminisms. Within feminist criticisms of research, women are seen as marginalised in
society; their experiences and world-view are not the dominant one. Child-centred
researchers similarly tackle how children are marginalised and silenced in society. While
some elements of feminism prioritise women’s experiences within society, sometimes
above men’s, others focus on the equality of everyone within society, irrespective of
gender, class, age or race. From that perspective, everyone’s experiences are valid and
there should not be one dominant view. Pring’s assertion that the distinctions within
paradigms are as important as the distinctions between them (Pring, 2004: 48) is pertinent
here. The child-centred theories and methodologies of researchers like Carr (2001) and
Clark and Moss (2001) may not seem feminist at first glance but their aims and outcomes
can be seen as part of one overarching feminist paradigm. In the research I have done at
work with my team I have referred to and discussed child-centred theories with them, but
for myself, I have applied a feminist paradigm when analysing the interviews, discussions
and interpretation of child-centred theory.
In my research project my team at work are central to my methodology. As I am a
member of the team, I am involved with them and identify with them; just by virtue of
who I am in relation to them, I take a feminist stance of the critique of disengagement
(May, 1997: 20). While the team have been involved in the process of finding new ways
to user test the CBeebies website with children under six they have not been involved
every step of the way in critically evaluating what works and what doesn’t. They gave
their feedback about the current process when I interviewed each of them individually
27
and when I conducted the group discussions with them. The smaller ‘working group’ I set
up as a result of those group discussions is more involved in actually formulating a new
process. Therefore, I am not doing action research with the whole team, though the
potential outcomes (new or improved practice for young children) may be similar. The
methods I use with the smaller working group are more akin to action research. The
emphasis with the team is on practical outcomes and their input into the research will
affect the theories and methodologies I then apply at later stages of research.
Before I started my research my team and I thought we were already applying a particular
methodology when we went user testing. We based our user testing on the usability
methods advocated by Nielsen (2002), Kuniavsky (2003), Krug (2006) and others. We
also adapted those methods as we were aware that most usability methods were designed
for adults. So we used expertise from colleagues in the department who had done user
testing with children and took advice from usability managers and behavioural scientists
across the BBC. We referred to the article from Microsoft about user testing with
children (Hanna et al, 1997) and the 70 design guidelines for websites for children
published by the Nielsen Norman Group (Nielsen and Gilutz, 2002). Thus we had a
methodology of sorts, even to the extent that we had a departmental document to refer to
about usability, user testing and child development (Ward, Howell, Wilton, Winter,
Dickinson and Gauld, 2004). So from the very beginning of my research, the principles
of usability gave me a foundation and starting point. These principles are based on
qualitative methods and analysis. Some quantitative methods can be used to analyse
usability test results (see Nielsen, 2004) but overall usability studies and user testing are
qualitative in both methods and analysis.
The other methods I have chosen for my research project reflect the evolution of my
thinking as well as the evolution of the project itself. I will shortly explain in detail why I
have chosen each method. When explaining these choices, I am keeping in mind how the
methods fit within the parameters I have explained above: qualitative research, the
feminist paradigm and the child-centred view.
28
My methods of research
I began the project with the intention of trying out different observation methods at the
different user-testing sessions we do each month, motivated originally by the Tavistock
method of psychoanalytic infant observation (see Rustin, 1989). I eliminated this as an
effective methodology early on but the actual process of attending the sessions shaped the
subsequent stages of the project. As we take it in turns to go user testing each month in
pairs, it was unusual for the same person (me) to go to every user-testing session over
three or four months. At first I was there as a third person observing the children’s
interaction with the website but I was also able to observe the differing approaches to
user testing from my colleagues. I then compared the different reports written about each
session. These reports and the discussions I had with my colleagues after each session
gave me insights into subjective interpretations of the findings. More significantly,
attending the sessions showed me that the methodology advocated within the department
for user testing was not always being followed and that sometimes my colleagues were
justified in not following it. Therefore, I realised after just two or three sessions that my
original plan to try out different methods of observation while my colleagues carried on
with more traditional user testing was just not possible. I had been aware that our
methodology was flawed in some ways because we were having to adapt adult usability
methods for testing sites with children under six but the continuity of attending every
session highlighted how we as a team needed to rethink what user testing means for
young children.
I therefore decided to interview everyone individually at the BBC who does or has done
user testing for the CBeebies website. I wanted to find out their exact thoughts on why
they thought we did user testing, how they saw the whole user testing process and how
they thought user testing could be changed. While I had support from my manager and
the lead designer to make changes to the user testing process, I wanted to be sure that
everyone was consulted, felt involved with those potential changes and had a chance to
have their say.
29
The final list of twelve questions (see appendix A) was carefully ordered to progress from
general thoughts on user testing to the user-testing process itself and then to how user
testing could be ‘made better’. I interviewed fifteen of my colleagues. Four were
designers, nine were from the editorial team (one editor, three producers, six researchers)
and one was from the TV production side of CBeebies, who conducts focus groups to test
new CBeebies television programmes. O’Connell Davidson and Layder refer to how
difficult it can be to tell whether an interviewee is telling the truth and being completely
open (2004: 116). By giving my colleagues the option to use pseudonyms I gave them the
opportunity to be as open as they liked within a working environment, in case they had
any concerns about making comments about other colleagues or current working
practices. Equally I stressed that the final report I would write was designated mainly for
the university rather than our department. As they all chose to use their own names they
appeared to have very few concerns, although one was indecisive and only decided to
definitely use her own name near the end of the interview. If the department as a whole
wants to use my findings I will need to ask permission again from my colleagues to allow
others access to the interview findings.
The interviews themselves were semi-structured: there was a script of twelve open
questions but I probed further if need be and a dialogue built up between me and the
interviewee. Therefore my interview method was qualitative. May refers to the three
necessary conditions for the successful completion of interviews, from Moser and Kalton:
accessibility, cognition and motivation (May, 1997: 115). All fifteen interviews were
conducted successfully, with only a few problems around accessibility and cognition; all
the interviewees understood what was expected of them and as they had all been user
testing were able to provide the information I was seeking. Three of the questions were
sometimes taken more literally than I expected and did not yield as much information as I
was seeking. The answers to the fourth and fifth questions (‘Think back to the last usertesting session you went on. What did you think worked well?’ and ‘What didn’t work
well?’) sometimes focused on the game or website that had been tested, rather than on
how the session itself was conducted. So, a game ‘worked well’ if it had been positively
received by the children. I tried to seek clarification by asking questions about the session
30
itself but it was interesting how closely a child’s response to a game was seen to affect
the success of a session by at least four of the interviewees. Question eight, ‘Do the
children learn anything new?’ had the shortest set of answers. I did not probe further with
this question and will explain why in my analysis of the interviews.
The third necessary condition, motivation, was sometimes problematic. It was difficult to
phrase any probing questions in an open a way as possible. It was sometimes hard to ask
for clarification without a closed question. Moreover, in my desire to make each
interviewee’s participation as valued as possible I found myself agreeing with comments
more than I expected and making approving comments (‘that’s really good’, ‘that must
have been rewarding’). I also clarified what they said by repeating comments back but in
my own words. While the interviews were semi-structured and worked as a conversation,
if I were to do them again I would try to be more neutral in my responses and repeat back
exactly what they were saying to me if I needed clarification, rather than framing their
comments within my own belief system.
I recorded and transcribed every interview. I found that as I knew all the interviewees
very well, both the interviewees and I were sometimes making assumptions about what
we each understood from the other. Thus points that seemed clear during the interview
were actually incomplete thoughts once I listened again to the audio. I had sometimes let
the interviews progress too much as conversations, (see Denscombe , 2003: 164) so I did
not always find out (or remember) as much detail as I had wanted. This approach was
consistent across all fifteen interviews, irrespective of my working relationship with them
or their gender (just two were men). Of the fifteen, the designers and producers (seven
people), I saw as my peers, in the sense that I work alongside them but do not manage
them or report to them. Of the six researchers, I have been responsible for managing five
of them at one time or another so I was aware of my more senior status to them. The
editor of the CBeebies website is my manager. For both the researchers and the editor I
felt that the interviews were open and honest but of course as we are not on a totally
equal footing at work some opinions may have been more muted than with my colleagues
who are peers.
31
When writing up the interviews I ordered the document by the questions rather than the
interviewees so I could compare what each person said for each question. I then went
through and put into bold text the points that seemed most important (see appendix B as
an example). I then extracted those phrases in bold text and put them all into a separate
document, so that each question had clear themes beneath it. I summarised those points in
another document, which I used to brief Liz, the lead designer and Debra, the senior
designer (both usability experts and both of whom I had interviewed) on the findings so
far (see appendix C).
I chose to have a meeting with Liz and Debra before scheduling a group discussion with
all the interviewees primarily to find out how we could begin to return to the principles of
usability in the user-testing sessions (i.e. a return to the original methodology we were
supposed to be following). The meeting clarified two things. One was that the existing
methodology did need rethinking for children under six – while it was being followed
successfully for user testing for the CBBC website (for children aged seven to twelve) we
needed to have our own methodology for user testing the CBeebies website. The other
was that to apply child-centred theories to the new methodology may be harder than I had
expected with my colleagues. When we discussed question six, ‘What do you learn when
you go user testing? (Do you learn anything new?)’ we talked about having high and low
expectations of the audience and its capabilities, partly in the context of HumanComputer Interaction and partly in the context of child development. Within HCI we
discussed how the emphasis tends to be from a marketing perspective (what can children
do that will ensure the success of this site and therefore make money for advertisers and
commercial companies). Within child development, the two designers talked mostly
about the mental capabilities of children, referring to the usability document that contains
basic child development milestones (Ward et al, 2004) from birth to six years. As they
did not seem to think beyond that document in our discussion, that informed the structure
and plan of the first group discussion and the expectations I had of my colleagues.
32
I originally planned just one group discussion but we ran out of time to fit in everything I
wanted to discuss so I then scheduled a follow-up. In that first group discussion I decided
to focus on our audience of young children and on how we are successful as a team
making online content for children for one of the most popular UK websites for young
children. I did that to ensure that the overall feeling was positive and also to try to get
them to look at the children who use the website in a different way than they might have
done before. I was therefore sharing for the first time how child-centred theories were
guiding me and could help us in our user testing. When talking to the two designers there
had been a feeling that the three of us could implement changes which the team would
have to go along with. But I was keen to make sure, as with the interviews, that everyone
felt involved and that they were initiating any changes. In that way I was applying my
overall paradigm of feminism as I saw everyone’s perspectives as valid, there was no
‘right’ way to change the user testing. I was also applying standpoint feminism as I
wanted to involve all of the team in the changes, using their experiences of user testing as
a starting point.
In the discussion, after I had asked everyone to think about what is really important about
CBeebies and what is really important about what we do as a team, I divided them into
three smaller groups, chosen by job role and known team dynamics: two groups of three
and one group of four (three people were absent and I did not join any of the smaller
groups). In these smaller groups I asked them to think about ‘what is a child’ and write
down their thoughts on post-it notes of a particular colour. I asked them to particularly
think about real children they knew and about what children can do overall, not just
referring to computer use. I then showed a video of children of five and six-years-old
making fences and chairs out of wood, using adult carpentry tools (Carr, 1998). I also
handed out photocopies of ‘The amusement park for birds’ from Reggio Emilia
(Malaguzzi et al, 1987). I then asked the groups to carry on writing down ‘what is a
child’ and ‘what can a child do’, but on different coloured post-it notes from before.
When I asked each group to feed back to the group as a whole I was interested to see how
much difference there was between the two sets of post-it notes. I also asked them to talk
33
about what they were thinking before and after seeing the video. The final step in the first
group discussion was to write down, in their three small groups, their thoughts about the
children they meet when they go user testing, on a third set of differently coloured post-it
notes. After they fed back those thoughts we had already run over time so I had to
postpone the last two steps I had planned for another week and another group discussion.
The group had already started to talk about the shortcomings of the user-testing process,
once they started thinking about the children they user tested with. I asked them to save
those thoughts for next time.
The second group discussion took place six weeks later, much later than I had planned,
due to illness and the difficulty of getting everyone in one place at one time. I asked them
in the interim to look at the user testing and child development document we used in the
department (Ward et al, 2004), in order to think about what a user-testing methodology
‘ought’ to be but also, more importantly, to refresh their minds about child development
and what children can do. I had had a good response to the Learning Stories video (Carr,
1998) and some of the group were interested in how children could be involved in talking
about and planning their learning activities (even though I had offered the video as an
example of what children can do, rather than as an example of Carr’s ‘four Ds’ of
Learning Stories). Therefore I decided to be more overt than before about the childcentred theories. I put together a document outlining different approaches to user-testing
with children, together with some assessment techniques (target child and Learning
Stories) and a reference to how young children’s thinking and reasoning skills are
sometimes underestimated (see appendix D). I also chose to include ‘the role of the child’
in the user-testing sessions in the group discussion.
In the second group discussion I gave them the list of themes from the interviews (see
appendix E). I again divided the group into three smaller groups. As before, I chose
carefully who was in each group, dividing up different job roles. I asked each group to
come up with solutions to each theme and then share those ideas with the larger group. I
gave out the document I had prepared about different approaches (appendix D) for them
to refer to while coming up with solutions. I also gave them the Reggio Emilia photocopy
34
again (Malaguzzi et al, 1987) and wanted them to keep in mind what a child can do
throughout their discussions. I had wanted to show the video again (Carr, 1998) but
unfortunately did not have it available that day.
When I listened to each group as they discussed their assigned themes and tried to come
up with solutions, I had a general feeling of having heard it all before. I had chosen to
give them the themes from the interviews in order to avoid repetition but discussing it in
groups just meant the same comments and complaints came up. I had asked them to come
up with action points of what we could actually do for each theme and was hoping the
document about different approaches would help them with that. There were some very
good points when they all fed back but overall the document had been ignored and the
solutions were more for general guidance than specific ideas (with some notable
exceptions which I document in the next chapter). I was disappointed as I had hoped for
more specific action points.
I finished the discussion by suggesting we set up a working group of a few of us in order
to actually put their ideas into practice. It had become evident that to keep discussing in a
large group was impractical. I asked for volunteers for the group and four people put
themselves forward that day. The group is now led by me and made up of two designers,
two researchers and one producer from interactive TV (who has not been involved in any
of the discussions so far but needs to establish a way to user test interactive TV content
with children under six). There is also one client-side developer as I requested someone
from the technical team. On the periphery is the senior designer, who will advise about
usability best practice, and another researcher who has expertise in usability for children
with special needs. The other outcome of the group discussion was that we cancelled the
regular user-testing sessions for the time being, or at least over the summer, in order to
provide us with the space to develop a process that works for us and the children.
35
Discussion: Analysis and Evaluation
Research questions
1. What is the web production team’s understanding of the principles and purposes
of user testing?
2. What do the web production team understand about the audience when they go
user testing?
3. How feasible is a more child-centred approach when the web production team go
user testing?
Introduction
Throughout this chapter in which I discuss and interpret the findings from the interviews
and group discussions, I am keeping in mind the metaphor of two threads entwining,
coming undone and then entwining again, perhaps less tightly – imagine it as a double
helix. The two threads represent themes that have run throughout the project as well as
those that inspired it at the beginning. The first is respect for usability (reflecting the
principles and purposes of user testing from the first research question) and the second is
putting the child at the centre (reflecting the child-centred approach of the third research
question). The second research question, to find out how the web production team
understand the audience when they go user testing, is represented by the rungs of the
double helix, while the first and third research questions are the twisted sides. Finding out
how the team understand young children when they go user testing is crucial to joining
up and keeping entwined the two dominant themes of the whole project. At various times
the threads have become separated, harking back to their ostensible original lack of
connection, reflecting how usability is not always compatible with empowering children
and putting them at the centre.
A key question at each stage of the project has been ‘where is the child?’. While that
question reflects both the third research question and the second thread, it is most
relevant when the child is ‘missing’ from thoughts, discussions or theories. If the child
disappears then the second thread, of putting the child at the centre, completely unravels
36
from the first thread and the rungs joining the two together. Disappointingly for me, the
main themes or concepts from the findings of the project are often focused around the
first thread and thus the child has got lost. The child was lost at different times and for
different reasons, but sometimes because of my chosen approach. I therefore look very
closely not only at what I said to the team but also what I did not say. Was I always
giving the team a chance to entwine those two threads together themselves? Equally, I am
aware that what the team have not said is sometimes as important as what they have said.
As discussed in the previous chapter, underlying the whole project is the principle that
theory informs research or practice and research or practice informs theory. As I interpret
the findings of the project, this principle becomes more significant, coming to the fore
more than I had expected. It reflects the double helix metaphor too, in that you cannot
have one thread without the other. Of course, overarching the whole study is a feminist
paradigm. Feminism in terms of prioritising women’s experiences above men’s in society
is not applicable here, instead it is the feminist paradigm of equality of everyone within
society, irrespective of gender, class, age or race. Reflecting standpoint feminism,
everyone’s experiences are valid and there should not be one dominant view. This
reflects the child-centred approach but also acknowledges how the experiences of the
team themselves are just as important as the experiences of the children. Those
experiences of the team have fed into the theory and practice of the end project, showing
the mutual dependence of theory informing practice and practice informing theory.
Denscombe (2003: 272) refers to ‘early coding’ and ‘new insights’ in the process of
analysing qualitative data such as interview transcripts. In my discussion I am aware of
how sometimes the early coding and analysis has needed rethinking in light of new
insights. I have a good ‘audit trail’ (Denscombe, 2003: 274), recording my thoughts and
decisions around the interviews and the two group discussions. That trail and its decisions
are dependent to some extent upon the document I prepared for Liz, the lead designer,
and Debra, the senior designer. The document is the culmination of the initial stage of
coding and analysis of the interviews, summarising the key themes I had found,
organised question by question (see appendix C). New insights around the interviews
37
only came about once I could compare my colleagues’ responses to what was discussed
in the two group discussions. Those insights are less well-coded than the original bolding
and extracting of themes that I did for the interviews. Also, by organising the analysis of
the interviews by each interview question rather than by another method, such as by
interviewee, the three research questions or any other overarching theme, it is not always
clear from the original analysis how the different questions relate to each other and have
connecting themes. There are obviously connecting themes between the answers to the
different interview questions and those themes informed how I structured the group
discussions. Thus some of the initial coding contributed to the evolution of some of the
later insights but also had to be rethought as new themes or connections emerged.
The interviews
When I asked the second, third and ninth questions in the interviews I was mainly trying
to find answers to the first research question. I was trying to improve my understanding
of what user testing meant to the team. I asked:
•
What do you think is the most important part of user testing?
•
What do you see as your role in user testing?
•
How do you use your user-testing findings?
When I analysed the answers to those questions my response was not a child-centred one.
When I summarised the main points for Liz and Debra I judged what the team had said in
light of the methodology they ‘ought’ to be applying. When I had decided to interview
everyone I had done so on the premise that the methodology advocated within the
department for user testing with young children was not always being followed and that
sometimes my colleagues were justified in not following it. However, I did not look at
their answers in those terms. I let the first thread, of respect for usability, dominate my
analysis, rather than acknowledging how the overall analysis needed the two threads
intertwined in order to contribute to how the user testing could be changed. One reason
for this may have been my own confusion about what was a valid methodology for user
testing with children under six. I had obviously been influenced by usability experts such
as Krug (2006), Kuniavsky (2003) and Nielsen (2003), despite the fact that none of them
addresses user testing with young children. Part of my original plan was also to return to
38
the basics of usability and user-centred design with the team and then apply child-centred
theories and practices from that point.
As I show below with examples from the second, third and ninth questions, I kept that
aim too far at the front of my mind once the interviews had been conducted, but once I
gained new insights in later analysis the two threads sometimes remained entwined. The
examples also show how I sometimes missed what my colleagues were really saying by
focusing too much on the first thread in my analysis for Liz and Debra. Moreover,
sometimes I needed to look at answers to other interview questions to find the answer to
one particular interview question. Cecilia herself alerts me to this during her response to
the very first question, ‘Why do you think we do user testing?’:
“I think – no, that’s answering a different question! [I say she can carry on
nevertheless] I think one of the drawbacks of the way we do our user testing right
now is the fact that a lot of it is done when something is finished. So we learn
what we did wrong or, not what we did wrong, but what we could do better for
next time and take that information with us. But I would like to do more user
testing at the halfway stage of any content – and that’s my plan from now on.”
She sets up very early on what her belief is about user testing and that runs throughout
the interview. However, in her later answers she does not repeat her plan to always user
test halfway, presumably as she has already told me about it.
When I summarised the team’s answers to question two, ‘what do you think is the most
important part of user testing’, I said there was a split in opinion between listening or
observing being most important and being prepared or organised. I also acknowledged
there was a significant minority who favoured talking to the children but I wondered if
that minority were the least experienced in user testing. However, evaluating the answers
from the perspective of the second research question, seven or eight of my colleagues
think that the most important part of user testing is paying full attention to the child,
listening to what they say and observing what they do. Indeed, some of them favour
39
talking to the children and asking them what they like, which shows a connection to the
child and putting the child at the very centre.
I grouped the minority who favoured talking to the children with those who had
mentioned guiding the children in the sessions. I was aware that within traditional user
testing the facilitator in no way guides what the user does, returning every ‘what do I do
here’ question with a question of their own: ‘I don’t know, why not have a go and see?’
or ‘What do you think you do?’. Therefore I was alert to any suggestion from the team
that they might be guiding the children and somehow adversely affecting the results we
were getting. In fact, only two people, Eimer a researcher and Aidan a designer,
mentioned guiding the children in their answers to the second question and both of them
made clear that it was of secondary importance:
Eimer: “I think it’s actually getting kids to sit down in front of the PC and seeing
firstly how … their reaction to it. It is generally what they do when they’re in
front of that game or that story. And then secondly, as the adult… just kind of
guiding them through it but I think the initial, what their understanding and
competency is, is very important.”
Aidan: “I always quite like just sitting back and trying not to help kids too much,
and just seeing how they find their way about and if it’s really… if it’s simple to
use, just observe them first of all, maybe give them like a hint to … ‘start there’,
maybe but then just really see how they go from there… just observe and maybe
give guidance after that.”
I then asked him how much the session would be split into in terms of observation and
guidance.
Aidan: “Probably because of our age group I mean… I would think a tiny bit of
guidance to kick things off and then if they get stuck, suggestions, what do you
think you would do here… but leave things quite open.”
40
In my analysis of the second interview question I was partly influenced by the answers
for the next question, ‘what do you see as your role in user testing?’. Eimer and Aidan
again mentioned guiding the children, as did Rachel, the editor of the CBeebies website.
Others talked about “talking them through it”, “directing them” and “dictating the
session”. Some of the team therefore did see their role as being a guide when they were
with the children and I summarised it in that way for Liz and Debra.
The other comment I pull out for analysis of question two is: ‘Also emphasis on
environment and being at ease – no distractions.’ This was mentioned by three of the
respondents. I had been expecting some answers along those lines as we had often
discussed at work how a busy computer class or a cramped nursery were not ideal
environments in which to user test. The comments about environment reflected both how
the team were thinking about the children and how they understood some of the
principles and purposes of user testing. Not all usability experts recommend using a usertesting lab – Nielsen is keen to show how it can be done in a home or office setting with a
few users and at least one observer. However he does stress that in those settings you
must ‘close the door to keep out distractions’ (Nielsen, 2003). Nikki and Helen, both
producers, really focused on how the child should be feeling:
Nikki: “It’s essential to have children in a safe, relaxed environment, to try and
get a good rapport with them to start with, which is often difficult with time
limitations.”
Helen: “The most difficult thing I find is when a child isn’t relaxed – so I suppose
the answer to that would be… making sure the child is relaxed, you know, be
comfortable… putting them at ease, either in an environment they feel
comfortable in, or… you know.”
Their comments show how they are identifying with the children when they go user
testing and thinking about how the sessions affect them.
41
For question three I was concerned that some of the team were confusing being a
facilitator and being an observer, using a comment from one of the producers, Cecilia, as
an example:
Cecilia: “But the observing is not done in a stand-back watching way, it is done
very proactively, which is why it’s good to have two of you, because you both
take notes.”
I then highlighted how only the designers (four of the interviewees) and Rachel, the
editor of the CBeebies website, suggested how editorial and design could take it in turns
to be facilitator and observer. I was concerned that the editorial team were not
understanding the principles and purposes of user testing as they were happy for more
than one person to take notes, for the adults to be proactive and involved with the
children and for the difference between facilitator and observer to become muddled. Yet
those very responses were verbalising the concerns that had prompted me to interview the
team. Their comments confirmed that the methodology we thought we were applying was
not working because of the age of the children we were user testing. The answers to
question three also tell me many things about how the production team understand the
audience of children aged six and under when they go user testing, and thus answers the
second research question.
Looking at the answers to question nine in light of the second and third research
questions gives one crucial insight. The child has been lost by most of the team. I
summed up the majority response to question nine as: ‘Overwhelmingly about feeding
into next or future projects (eight-nine respondents)’. I then expressed surprise at this
response as I had always believed user testing was about the specific site or game being
tested in one particular session, and how the user experience of that particular site or
game can be improved. I was thinking more of the principles of usability and user testing,
not the reality of how the team really do use their findings. While some of my colleagues
may aspire to make changes at the halfway stage of a game’s creation, often there is not
42
enough time within the production process or they do not have the authority or resources
to instigate that change. Liz, Debra and I discussed how the researchers particularly are
not empowered to make changes. However, there was doubt from Liz and Debra in our
discussion that findings from user testing actually did feed into future projects – in their
experience, it usually only happened by chance. The three of us were thinking too much
about how findings ‘ought’ to be documented and interpreted, and how the timing of user
testing in a project is vital, following the user-testing methodologies of Krug (2006),
Kuniavsky (2003) and Nielsen (2003). Along with most of the team we had also lost the
child-centred focus.
A few members of the team are thinking of the child though, and the benefit to the whole
audience of what is gained from a user-testing session. Therefore the rungs of the double
helix are still precariously in place.
Katherine: “But I like to think I would use them [her findings], especially from
the last thing I learnt, that they like to be interactive, they like to hear noises, and
see things react to something they do. Even like a really simple thing like
listening to those animal noises which, you know, is so basic, and seeing the pig
put his face in a load of mud and be covered in mud… they just really seemed to
love that. I think that’s… I like to think that I would use what I’ve learnt.”
While this comment is partly about feeding into future projects, Katherine, a researcher,
is also very much thinking about the children she met at her last user-testing session. She
is telling me how she understands the audience and how she can think of what the child
might want from both the content we produce and the user-testing sessions. She is also
telling me what she learns when she goes user testing, which I had asked in question six.
In this response she gives me more detail than before, as for question six she talked most
about learning how competent a child is and how their age and ability can affect their
response to a game. Helen also talks about the role of the child in her response to
question nine, suggesting an ideal approach would be to have children feeding into a
43
project from the very beginning, telling us what they like about a particular television
programme or character and how we can reflect that on the website.
I was disappointed with the responses from the whole team to questions seven and eight.
The questions were:
•
How do you think the children find user testing?
•
Do the children learn anything new?
Question eight had the shortest set of answers and question seven was not far behind (the
first and seventh questions had about the same length of responses and were the second
shortest set of answers). For question seven, I summarised that most of the team think the
children enjoy user testing with a significant minority emphasising how the children were
sometimes shy and could find it intimidating and tiring. Question eight made them think
most about the child’s role, more than for question seven, but they mostly focused on the
children just finding out that there are new games to play on the CBeebies website and
learning computer skills. I had been hoping for answers that showed how the team put the
child at the centre in user testing. I was thinking mostly of the second thread, putting the
child at the centre, and the second and third research questions, both when composing
and when analysing the answers to these two questions.
If I had asked ‘What do the children learn’, without the emphasis on learning anything
‘new’ I may have got more detailed and in-depth answers. By using the word ‘new’ the
team seemed to think about what new content we had produced recently, as shown by
comments by Sandrine, a researcher, and Rachel, the editor:
Sandrine: “No. Only thing I can think of they learn is when we show them ‘check,
have you played that game before? That’s our new game…’. The only thing they
can learn is what’s available…. broaden a bit their horizons.”
Rachel: “Yeah, learn about new pieces of content, we’re not really there to teach
them, there to observe. They might learn about a new piece of content but they
don’t develop a skill, they don’t develop anything in that respect.”
44
The use of the word ‘learn’ made some of the respondents think about teaching (as with
Rachel’s comment above, three others spoke about teaching and educational value).
While a lot of our content is produced with the six early learning goals in mind (see
QCA/ DfEE, 2000) and thus offers ‘soft learning’ this did not seem to be at the forefront
of any of the interviewees’ minds, though to be fair many made the point that not much
can be learnt in a half hour session. I had not phrased the question well enough to find out
what I actually wanted to know, namely how much the team think of children as active
learners throughout their daily lives.
When composing question eight I had had the different observation and assessment
methods for young children at the back of my mind, particularly Carr’s Learning Stories.
As I had not yet introduced any of the team to her methods and ideas at the time of the
interviews I did not feel it was appropriate to start expounding upon learning dispositions
to them when I asked that question. However, Cecilia’s answers to questions six and
eight show that she sees us, the team, as constantly learning something new, whereas if
user testing is ‘perfect’ then the children won’t necessarily learn anything:
Question six: ‘What do you learn when you go user testing? (Do you learn
anything new?)’
“I think we always learn something new, we really do ... To go in and say ‘I knew
what was going to happen and it was a waste of my time’ you should go and get
another job.”
Question eight: ‘Do the children learn anything new?’
“If our user testing was really, really good and perfect and in the home … in
many ways then they themselves wouldn’t because they may be doing something
that they do a lot or they do fairly frequently, unless we’re testing a game at the
halfway stage.”
45
Cecilia then goes on to say that children do learn things to help them count, encourage
them to clap and play a tune, but she thinks it is a moot point whether they learn that
during our testing or playing the same game after we’ve gone.
In our discussion about the main findings from the interviews, Liz and Debra suggested
that maybe the majority of the team saw themselves in a nurturing role when user testing
and thus made assumptions about how the children are not capable and need help. I partly
agreed that the team take on a nurturing role, as some of them do, and therefore
underestimate the audience, not giving them as much independence as they might like.
The lack of information and insight I had got for questions seven and eight from the team
supported that angle to some extent. I had hoped to discuss with Liz and Debra how
children’s capabilities are more far-reaching than the basic child-development milestones
outlined in our internal document (Ward et al, 2004), as well as looking at the role of
expectation of users within user testing generally. But they focused on the accepted
knowledge we already had and I found it hard to move the discussion beyond that.
The first group discussion
In my plan for the group discussion I wanted to build up to thinking about the children
with whom we go user testing. I wanted the team to focus on the second thread, putting
the child at the centre, while my focus was on the second and third research questions.
Moreover, while I was trying to challenge the belief that user testing is only testing the
website and not the user, putting the child at the centre is actually one of the principles
and purposes of user testing. As user testing is crucial for user-centred design, the first
research question was also present throughout that first group discussion.
I chose not to feed back what had been said in the interviews and instead took them down
a different path for a while to see what the response then was. We first explored the value
and strength of both the CBeebies brand and the CBeebies web production team before
talking about the children whom we meet. This was important because they were all
acknowledging that there were problems with the current user-testing process: I wanted
to reassure them that these were not problems per se with the team and the brand. When I
46
then tried to make sure the team were really thinking about the children they meet when
they go user testing, I was trying to challenge any preconceptions they might have about
how competent young children are. I took that approach because of the disappointing
responses I had received from the team for questions seven and eight and the apparent
lack of understanding from Debra and Liz about the high and low expectations of the
young audience. But of course, thinking about early coding versus new insights
(Denscombe, 2003: 272), the team had given me more than I realised about how they
understand the audience and how feasible it would be to take a more child-centred
approach to user testing. It was just they hadn’t given me all of that information in the
answers to questions seven and eight. I also made the assumption that if Debra and Liz
were unable to look beyond basic child development information then the rest of the team
were not doing so either.
In my original planning for the group discussion I was keen to share with the team the
four Ds of assessment from Carr: Describing, Discussing, Documenting and Deciding
(Carr, 2001: 101). It offered a good process to suit what we do in user testing, particularly
the writing up and following through of the user-testing findings, plus it was designed
originally for assessing young children. I wanted the main focus of the discussion to be
about children – general vignettes of children and specific examples of those with whom
we user test. This seemed important not only for the reasons above about preconceptions
of the audience but also in order to wholly understand Carr’s Learning Stories the team
needed to think from a more child-centred position. So while the discussion was building
up to finding solutions for which the four Ds might be appropriate, I did not want
methods of documenting to dominate the discussion. Thus when I showed part of the
video about Learning Stories I chose the extract carefully. I was using the video as an
example of what young children can do (make gates, saw up wood), not as a lesson in the
four Ds of assessment.
There were two significant findings from the first group discussion. The first was around
the difference between the responses when I asked the team to think about ‘what is a
child’ twice during the group discussion and then about the children they meet when they
47
go user testing. Liz, the lead designer, commented that the second set of post-its, those
written after they had watched the video, were ‘more advanced’ in terms of what children
could do than the first set of post-its. This was certainly true as comments included
‘creating and making things’, ‘decisive’ and ‘capable/ intelligent’. The first set had a
mixture of active and passive comments, and positive and negative impressions of
children. There was mention of them being creative, curious, always learning, unselfconscious, liking to build things. There was also mention of them not being dextrous, of
being cautious, of getting easily frustrated, not understanding things, being self-absorbed.
There was nothing overtly negative in the second set of post-its, although while they were
watching the video itself one comment was ‘scary!’. However, there was an increased
reference to the role of adults in a child’s world. Out of 22 comments, eight were
somehow related to adults. These were those eight responses:
Role-play
Mimic adult behaviour
Inspired by their parents
Social/ cultural perceptions
Social pressure by adults on what kids can/ can’t do
Perception from adults what kids can do
Copying adults
Pick up on adult attitudes
While ‘social/ cultural perceptions’ may not seem to mention adults in a child’s world I
have included it in this list as once everyone had fed back about their second set of postits, the discussion was mainly about what is imposed on a child by an adult and what is
socially acceptable. All three groups mentioned how the children were copying or
mimicking adults and I challenged the group on this as in the video we had not seen any
adults making gates or sawing up wood – it was all being done by the children. They said
they viewed the children as aspiring to be like adults. They then acknowledged how they
had low expectations of children but that that is often to do with society’s expectations
and portrayals of children.
48
When the group told each other about the children they meet when they go user testing
and wrote examples down on a third set of post-its, they told me how they did not refer
back to the other sets of post-its. To them, the third activity was a specific request and
task but the other two activities were not tasks in the same way. This explains to some
extent why most of the comments about the real children they encounter when they go
user testing did not reflect any of their previous thinking. After watching the video they
said children are ‘proud to show you what they’ve done’, but in referring to a user-testing
session they said ‘they like encouragement’. From the video, they commented that
children are ‘not hindered by reality’ and are ‘capable/ intelligent’ but the comments
about user testing emphasised how children need boundaries and have limited
capabilities:
Apprehensive and scared
Not very dextrous
Need structure
Very short attention span
Easily distracted unless they really like what they’re doing
They like encouragement
Need more positive reinforcement
Can get demoralised
Unpredictable
Computer literate yet like security/ praise of easy games (e.g. Teletubbies)
One of the group summed up their response to the first group discussion by saying that
‘the [CBeebies] website breaks cultural perceptions but the user brings you back to
reality’. In the same way, it seemed the video had broken their cultural perceptions but
thinking about real children had brought them back to reality.
I therefore received mixed messages in answer to the second research question, how the
team understand the audience when they go user testing. Before we had talked about the
49
children they meet when they go user testing, they all seemed inspired and motivated by
the video and really enjoyed the discussion about challenging society’s expectations
about children. Once we did talk about the real children they meet, that possibility
became more muted. If they were not able to see the young children they meet as capable
and intelligent then I was not sure they would be able to apply a child-centred approach
to user testing. While they were thinking deeply about the audience and trying to put the
child at the centre their inability to make that leap to empower children meant I was not
convinced that a child-centred approach along the lines of Learning Stories (Carr, 2001),
the Mosaic approach (Clark and Moss, 2001) or even Paley (1986) was feasible.
Moreover, even when they were identifying the abstract child, they still saw that child
from a Piagetian view. They regarded young children as unable to decentre and generally
‘egocentric’, so only thinking of their own point of view. The first set of post-its included
these comments:
‘Selfish (meaning self-absorbed, not learnt to think outside self)’
‘Difficult to see others’ point of view’
Donaldson (1987: 20-23) has shown that this is not true of children of four, five and six
years old, while Astington (1993) asserts that between the ages of two and five, children
begin to have insights into their own mental life and that of others, so they are able to
infer other people's thoughts, feelings and perceptions from words and actions.
The team’s reaction to the video, the second significant finding from the first group
discussion, showed that a new user-testing model was possible, though not necessarily a
child-centred one. This informed my plan for the second group discussion. Even though I
had carefully chosen the extract so that there was very little about the four Ds of
assessment, there was enough within the voice-over about the gate-making activity to get
the team interested in it. They were all intrigued by the talking and planning and wanted
to see the talking in process (the ‘Discussing’) with the children. In the first group
discussion we ran out of time to reach the point at which we were trying to find solutions
to the problems of user testing. I had hoped that by the time we reached that point that I
50
could bring in the four Ds as an option to help us, or even find that the group were
moving towards them without realising (talking about the importance of documenting
what we find, interpreting it as a group, making decisions etc.).
The second group discussion
As documented in the methodology chapter, there was a much longer time gap between
the first and second group discussions than I had originally planned, giving me time to
reflect on the first discussion and make changes accordingly. I tried to harness the team’s
interest in the Learning Stories assessment process, particularly about how the
practitioners involved the children in their own assessment, to feed into thinking about a
new model of user testing. The positive reaction to the video (Carr, 1998) seemed to
show that the team needed new ideas to inspire them for finding a new way to approach
user-testing. Those ideas could enable them to put the child back at the centre and enable
me to establish the feasibility of a more child-centred approach when we go user testing.
However, both those ideas (see appendix D) and my plan for the discussion show how the
double helix of respect for usability, putting the child at the centre and how the team
understand the young audience is delicately entwined.
The sheet I prepared for the second group discussion (appendix D) pulled out key ideas
from the literature I had been reading. The ideas were not all explicitly about putting the
child at the centre but they were acknowledging the role of the child in different ways.
Most importantly though, the ideas were all new to the team and offering new ways of
looking at the user testing we do: I wanted to enable them to find alternative ways of
doing user testing. I was trying to combine the two threads so that even though I saw the
second thread, putting the child at the centre, at the fore, I was still maintaining respect
for usability (the first thread). I did this in two ways. The first was to give neutral
explanations of each idea or alternative method, so that I was not judging what we
currently did. The second was to show how the Learning Stories could be adapted for our
needs around user testing, not just suggesting we lift the assessment method wholesale.
Thus while one of the key parts of Carr’s Learning Stories is the involvement of the
51
child, in my explanation of it I highlighted the documenting side and how we could adapt
it for our needs.
By offering the sheet of ideas I was trying to follow the principle that theory informs
research or practice and research or practice informs theory. I wanted to see if the
theories I had found useful could inform the actual practice of user testing with the
CBeebies team. But even though I tried to portray the ideas neutrally I was still
unconsciously applying my own world view by which ideas I chose to include and how I
described them. Indeed, by hoping that others’ theories and research could inform our
practice and research I was already applying my own particular paradigm to the
discussion, my belief that no researcher can participate in society without adhering to a
particular paradigm or world view, consciously or not.
While I had intended the second group discussion to be a chance for the team to entwine
the two threads together themselves, combining respect for usability with putting the
child at the centre, the structure and focus of the group discussion eventually denied the
presence of the child. As I said in my introduction to this chapter, if the child disappears
then the second thread, of putting the child at the centre, completely unravels from the
first thread and the rungs joining the two together in the double helix. I asked the group to
keep in mind what young children can do throughout the brainstorming and discussion
and therefore handed out the Reggio Emilia photocopy again (Malaguzzi et al, 1987),
though I was unable to re-show the video of children making gates. I also tried to focus
on the second thread and the third research question by recapping what we had discussed
in the first group discussion. However from the point at which I introduced the seven
main themes from the interviews, the discussion focused on the first research question
and the first thread. Even though I added the role of the child to those seven themes (see
appendix E) I was still asking predominantly for solutions to problems that were about
the basic principles and purposes of user testing.
When I introduced the eight themes for discussion I asked the group if they had anything
to add. Sandrine suggested stress on the child during the sessions. I said I felt that fell
52
into a few of the areas – role of the child, environment, timing, our role and
communication – and so I thought it would be covered when we looked at each of those
areas. In retrospect, I wonder if it should have been included as part of the role of the
child. Sandrine was obviously thinking solely about the child and was therefore putting
the child at the centre, something I was trying to achieve through the inclusion of the role
of the child within the discussion.
In the write-up for each group (see appendix F) there is no evidence that anyone had
incorporated any of the ideas I had given them, nor the thoughts from the previous group
discussion, the Learning Stories video (Carr, 1998) or the Reggio Emilia photocopy
(Malaguzzi et al, 1987). This may have been because I did not really give the team a
chance to read over the sheet in detail before they divided into their three groups to
discuss the eight themes, though they could refer to the sheet throughout the discussion.
On that basis, I felt I had misjudged how ready the team were to use theory to inform
practice. However, as I show below, using examples from Group C’s discussion, that
misjudgement arose for two reasons. The first was emphasising respect for usability
above putting the child at the centre in the plan of the group discussion and the second
was focusing on initial analysis of the second group discussion above new insights. This
second reason reflects the analysis of the interviews – sometimes I needed to look in the
less obvious places for answers to the second and third research questions.
Group C discussed ‘your role’ (i.e. the team’s) and the role of the child, as well as
documenting of results. The theme of ‘your role’ produced more answers to research
question three and the second thread than the theme of the role of the child. For example,
points around ‘your role’ included:
We have to ask lots of questions as they don’t tend to give up how they are
feeling or what the experience is like for them freely.
Also need to encourage them so they can have the best experience possible.
Children cannot always read instructions or hear audio on laptop.
53
These points show how the team would apply a child-centred approach to user testing,
and not necessarily follow the ‘traditional’ principles and purposes of user testing. In
contrast, the points for the role of the child are not as active and do not put the child at the
centre in the same way. They go some way to answering the second research question
about how the team understand the audience when they go user testing:
To help us find out how well a piece of content had been made or how it could be
improved.
To express their honest opinion about how they feel when using the piece of
content.
They are the truth and evidence of what we do.
They are the window into our audience, their lives, what they like and dislike.
They need to have some degree of computer literacy to get the most out of the
content.
I had included the role of the child in order to make sure a child-centred approach could
be considered within the discussion, alongside some of the child-centred ideas and
theories I had offered (appendix D). While I had then focused more on the principles and
purposes of user testing through the seven themes from the interviews, Group C had
thought about the child, just again, as with the interviews, not where I was necessarily
looking for the child.
Argyris and Schön define two forms of theory of practice – espoused theories and
theories in use (Moon, 1999: 40). Espoused theories are part of the public perception of a
particular profession, seen to formally guide action, while theories in use are the patterns
learned and developed in day-to-day work. This division reflects the user-testing
methodologies that the web production team apply to their work. The espoused theories
are both the methodology of adult usability testing and the methodology advocated in the
internal document about user testing with children (Ward et al, 2004). The theories in use
are the actual methods the team use when in a user-testing session with children under
six, adapting the methodology that we have within the department to suit each child or
54
each session. In their three small groups the team talked about their own experiences of
user testing, using those experiences to come up with solutions. When they shared their
experiences of user testing, whether with each other or with me in the interviews, they
were showing how their own theories in use guide them day-to-day and inform their
practice. This is similar to the principle that theory informs research or practice and vice
versa. The practice of user testing (the experiences of the team) was informing the theory
(the adapted methodology the team were trying to apply to each session without
necessarily realising, their theories in use). The team’s theories in use were more
informed by child-centred practices and theories than it initially appeared. By applying
those theories in use the team showed how they were ready to use theory to inform
practice, they were just not necessarily ready to apply new theories to inform their
practice.
While I had felt at the end of the second group discussion that the second thread had
become completely unravelled from the first thread and wanted to go some way to
repairing that, the actual methods that the team applied and their theories in use showed
that the two threads were still loosely entwined. I suggested a working group be set up in
order to take the action points from the second group discussion forward and to make the
process of applying change to the user testing more manageable. The specific ideas that
had come out of the discussion needed a process behind them (how to build relationships
with schools and nurseries; how to engage teachers and practitioners; create example
scripts for user-testing sessions; decide the generic form for documenting the findings)
and specific individuals to work on them. I also wondered if with fewer people it would
be easier to investigate the team’s attitude to a more child-centred approach, as well as
applying new theories to their practice. I was keen to move on from what I saw as a
‘talking shop’ of repeated ideas and complaints and actually do something that would
make a difference. The team had shown me how their experiences informed their theories
in use when they went user testing. I wanted to build upon that, particularly within the
paradigm of standpoint feminism, where everyone’s experiences are valid and there is not
one dominant view.
55
Conclusion
I started this project over a year ago, in summer 2005. My emphasis then was on trying
out different observational methods for young children in the user-testing sessions. The
focus of the thesis and the methodology applied has obviously changed since then but at
the heart still stands the original motivation, to find a different and more effective way to
conduct user-testing sessions with children under six. In my original plan I had hoped to
have a new way of user testing up and running by the spring of 2006; I now hope to have
a new way of user testing in place in the autumn of 2006. The user-testing working group
I set up after the second group discussion in May is currently on track to deliver a new
approach by October. The amount of time it has taken to reach different points
throughout the project has always been longer than I anticipated. Therefore one learning
outcome for me has been realising that taking an inclusive approach to everyone in a
team requires a flexible schedule and attitude. I have particularly reflected on that
inclusive approach at the end of the project after comparing my approach to the work
done by Ana, the Masters student on work experience in my department at the BBC.
When I saw the apparent ease with which Ana could produce one piece of finite research,
I was tempted to be more didactic than I have been. However, I have made sure I have
kept true to my original paradigms and methodologies. Those paradigms include my
belief that no researcher can participate in society without adhering to a particular
paradigm or world view, consciously or not; that theory informs practice and practice
informs theory; and that everyone’s experiences are valid and should be valued, through
the feminist paradigm of equality between men and women. In Dewey’s definition of
reflective action, the first necessary attitude is open-mindedness: ‘an active desire to
listen to more sides than one, to give full attention to alternative possibilities, and to
recognize the possibility of error even in beliefs that are dearest to us’ (Zeichner and
Liston, 1996: 10). I have sometimes wondered if I have been too open-minded, too
willing to listen to everyone and consider alternatives, not actually making any decisions
or producing concrete outcomes. Equally, as shown with my analysis of what I did and
what I found out, I have at times been less open-minded than is ideal for reflective action.
56
I have perhaps been more of a sophisticated believer than a critical believer (Wright Mills
in Zeichner and Liston, 1996: 10), wanting to find out opposing points of view in order to
refute them rather than engaging with those opposing points of view to show up the flaws
in my own belief system.
Another aspect to consider is my own status within the team. I have given reasons above
for not taking a didactic approach and for being led by particular paradigms. However,
even if I had decided to wholeheartedly embrace either didacticism or action research for
the whole project I would have had to gain approval from my managers and other more
senior people within the department to take either approach. As the majority of the people
I interviewed are junior to me or are my peers I have felt able to lead them in the group
discussions and in the working group. Working in a small team has also contributed to
the democratisation of the process – we have a strong tradition of our research feeding
into each others’ ideas and evolving into processes. But, ultimately, we are part of a
bigger department and the final decision on exactly how CBeebies user testing might
change needs to be considered within that. Though I hope that the new process from the
working group can be rigorously applied and tested to prove what works best, and then
disseminated to the team and department, we have already identified that we have at least
five project stakeholders (teams and individuals) and three sponsors (see appendix G,
project definition document for the working group). The department is also undergoing
management and structural changes as I write so those stakeholders and sponsors are
changing, as is the make-up of the web production team itself. Therefore there are a
number of unknown factors contributing to the risks of the project, though my status and
role will remain the same (leading the working group but not having ultimate say on the
new process).
To conclude now with a return to the three research questions:
1. What is the web production team’s understanding of the principles and purposes
of user testing?
2. What do the web production team understand about the audience when they go
user testing?
57
3. How feasible is a more child-centred approach when the web production team go
user testing?
The three research questions are obviously intrinsically linked, whether through the first
two questions building up to the third or through the metaphor of the double helix, where
the second research question is the rungs of the double helix and the first and third
research questions are the twisted sides. In the previous chapter I have evaluated my
methods and my findings in detail, showing when and how the two threads of the double
helix sometimes became separated. I have also shown when and how the two threads
have been entwined or at least loosely twisted together. However, I have sometimes
focused on the second and third research questions to the disadvantage of the first
research question. I have most often focused on it when evaluating my own analysis of
the team’s responses in the interviews or the group discussions, becoming frustrated
when the first thread of respect for usability seems to take over the research, analysis or
discussion.
The findings from the second group discussion showed most clearly what the web
production team understand about the principles and purposes of user testing (appendix
F). The seven themes from the interviews shaped the format of the second group
discussion and those seven themes were decided upon after discussion with Liz and
Debra. They reflected the frustrations the team had expressed to me – they were
effectively seven different areas that everyone saw as problematic across user testing and
which needed solutions. While I had some concerns that the focus upon them meant that
we lost the focus on the child in the second group discussion, their origination from the
individual interviews shows how the team understood the criteria for a user testing
process that adheres to ‘traditional’ adult user testing as advocated by Krug (2006),
Kuniavsky (2003) and Nielsen (2003).
Throughout my evaluation the second research question has been crucial to keeping the
two dominant themes of the whole project entwined. The web production team
understand different things about the audience when they go user testing, shown from the
58
varied responses in the interviews and the contrasting comments in the first group
discussion. When in the first group discussion they listed capabilities of real children that
were hindered by society, they showed that they did not necessarily believe the
empowering descriptions they had used for the second set of post-its could exist in
reality. As one of the team actually said, ‘the user brings you back to reality’. Indeed,
while they may have suggested ‘more advanced’ capabilities after watching the Learning
Stories video (Carr, 1998), ultimately they understand the audience from a Piagetian child
development perspective. In the second group discussion this did not change – although
there was evidence that a more child-centred approach in user testing was feasible, this
was based around willingness to listen to the children (‘To express their honest opinion
about how they feel when using the piece of content’; ‘They are the truth and evidence of
what we do’) rather than re-evaluating their capabilities.
The third research question has been the main focus and point of the project. The
feasibility of a more child-centred approach has decreased and increased throughout the
project, depending on what information the team gave me about what they understood of
the audience and how they understood some of the child-centred ideas I was presenting to
them. Through the basic principle of user-centred design the web production team do
think of the child user throughout their working lives. In order for the user testing itself to
be more child-centred and therefore empowering the children within the sessions, the
team do need to rethink some of their conceptions of real children. Overall, the team are
often thinking of the child when they go user testing and thinking from the child’s
perspective. However, because of the constraints of what user testing is ultimately for and
what it is trying to achieve (good usability and well-designed websites for all users),
CBeebies user testing cannot completely embrace such child-centred approaches as Carr
(2001) and Clark and Moss (2001). Therefore a more child-centred approach is definitely
feasible but not to the exclusion of other approaches.
I have set up the working group in direct response to the third research question, giving a
small group of people a chance to combine all their ideas and knowledge about user
testing and the audience to produce a new way of doing user testing for CBeebies
59
Interactive. The documented audit trail I have of the interviews, the group discussions
and now the working group meetings will feed into that new process. Moreover, by
learning to value the methodology that the team were already applying, their ‘theories in
use’ (Argyris and Schön in Moon, 1999: 40), the new process can be a continuation of
what some of them were already doing. If the working group succeeds in its objectives
then the two threads will remain entwined, ensuring that new team members can
confidently climb the rungs of the double helix, wholly understanding the audience and
delivering an enjoyable user-testing experience for both the CBeebies team and the
children with whom we user test.
60
Bibliography
Abbott, L. and Gillen, J. (eds.) (1997) Educare for the Under Threes - identifying need
and opportunity, Manchester: Manchester Metropolitan University
Ackers, J. (1997) ‘The ways in which children’s learning and development are assessed
and the extent to which this influences provision’ in Abbott, L. and J. Gillen,
(eds.) Educare for the Under Threes - identifying need and opportunity,
Manchester: Manchester Metropolitan University
Astington, J. W. (1993) The Child’s Discovery of the Mind, Cambridge: Harvard
University Press
Barendregt, W. (2006) ‘Evaluating fun and usability in computer games with children’
PhD thesis, University of Eindhoven
Brooker, L. and J. Siraj-Blatchford (2002) ‘ “Click on Miaow!”: how children of three
and four years experience the nursery computer’ in Contemporary Issues in Early
Childhood Vol. 3, no. 2 pp.251-272
Carr, M. (1998) Assessing Children’s Learning in Early Childhood Settings: What to
assess? – Why assess? - How to assess? (video) Wellington: NZCER
Carr, M. (2001) Assessment in Early Childhood Settings: Learning Stories, London: Paul
Chapman
Clark, A. and P. Moss (2001) Listening to Young Children: The Mosaic approach,
London: National Children’s Bureau and Joseph Rowntree Foundation
Dahlberg, G., P. Moss and A. Pence (1999) Beyond Quality in Early Childhood
Education and Care: Postmodern Perspectives, London: RoutledgeFalmer
Denscombe, M. (2003) The Good Research Guide for small-scale social research
projects, Maidenhead: Open University Press and McGraw-Hill Education
(second edition)
Donaldson, M. (1987) Children’s Minds, London: Fontana (first published 1978)
Druin, A. (2002) ‘The role of children in the design of new technology’ in Behaviour and
Information Technology Vol. 21, no.1, pp.1-25
Drummond, M. J. (2003) Assessing Children’s Learning, London: David Fulton (first
published 1993)
Drummond, M. J. and C. Nutbrown (1996) ‘Observing and assessing young children’ in
61
Pugh, G., (ed.). Contemporary Issues in the Early Years, London: Paul Chapman
and the National Children’s Bureau (second edition)
Elfer, P. and D. Wedge (1996) ‘Defining, measuring and supporting quality’ in Pugh, G.,
(ed.) Contemporary Issues in the Early Years, London: Paul Chapman and the
National Children’s Bureau (second edition)
Griffin, B. (1997) ‘The contribution made by educarers to the social construction of
identity’ in Abbott, L. and J. Gillen, (eds.) Educare for the Under Threes –
identifying need and opportunity, Manchester: Manchester Metropolitan
University
Hanna, L., K. Risden and K. Alexander (1997) ‘Guidelines for Usability Testing with
Children’ in interactions, Vol. 4 issue 5, pp9-14
Available online http://www.microsoft.com/usability/UEPostings/p9-hanna.pdf
Hanna, L., K. Risden,, M. Czerwinski, and K. Alexander (1998) ‘The role of usability
research in designing children’s computer products’ in Druin, A. (ed.) The Design
of Children’s Technology, San Francisco: Morgan Kaufmann
Available online http://research.microsoft.com/users/marycz/druin98.htm
Harms, T. and R. M. Clifford (1980) Early Childhood Environment Rating Scale, New
York: Teachers College Press
ISO (1998) Ergonomic requirements for office work with visual display terminals
(VDTs) - Part 11: Guidance on usability (ISO 9241-11)
ISO (2001) Software engineering - Product quality - Part 1: Quality model (ISO 9126-1)
Kuniavsky, M. (2003) Observing the User Experience: A Practitioner’s Guide to User
Research, San Francisco: Morgan Kaufmann
Krug, S. (2006) Don’t Make Me Think! A Common Sense Approach to Web Usability,
Berkeley: New Riders (second edition)
Malaguzzi, L. et al. (1987) The Hundred Languages of Children, Reggio Emilia,
Italy: Department of Education.
Markopoulos, P. and M. Bekker (2003a) Editorial to ‘Interaction design and children’ in
Interacting with Computers, Vol. 15, issue 2, pp141-149
Markopoulos, P. and M. Bekker (2003b) ‘On the assessment of usability testing methods
for children’ in Interacting with Computers, Vol. 15, issue 2, pp.227-243
Markopoulos, P., W. Barendregt, M. Brus and I. Plakman (2005) ‘The Parent
62
Evaluator Method’ Proceedings Workshop on Children Computer Interaction:
Methodological Research, p. 36-38, at Interact 2005, Rome, September 2005.
May, T. (1997) Social Research: issues, methods and process, Buckingham: Open
University (second edition)
McPake, J., S. Downey, L. Plowman, D. Sime and C. Stephen (under review, 2006).
‘Already at a disadvantage? The impact of socio-economic disadvantage on
young children’s developing ICT competences in the home’ in Contemporary
Issues in Early Childhood
Moon, J. (1999) Reflection in Learning and Professional Development, London: Kogan
Page
Nielsen, J. (2002) ‘Kids’ Corner: Website Usability for Children’
http://www.useit.com/alertbox/20020414.html
Nielsen, J. (2003) ‘Usability 101: Introduction to Usability’
http://www.useit.com/alertbox/20030825.html
Nielsen, J. (2004) ‘Risks of Quantitative Studies’
http://www.useit.com/alertbox/20040301.html
Nielsen, J. and S. Gilutz (2002) ‘Usability of Websites for Children: 70 Design
Guidelines’, Fremont: Nielsen Norman Group
Available online http://www.nngroup.com/reports/kids
O’Connell Davidson, J. and D. Layder (1994) Methods, Sex and Madness, London:
Routledge
Paley, V.G. (1986) ‘On listening to what the children say’ in Harvard Educational
Review Vol. 56 no. 2 pp.122-131
Pascal, C. and T. Bertram, (eds.) (1997) Effective Early Learning: Case Studies in
Improvement, London: Paul Chapman
Plowman, L. and C. Stephen (2003) ‘A “benign addition”? Research on ICT and preschool children’ in Journal of Computer Assisted Learning 19 pp.149-164
Pring, R. (2004) Philosophy of Educational Research, London: Continuum (first
published 2000)
QCA/ DfEE (2000) ‘Curriculum Guidance for the Foundation Stage’ London: QCA/
DfEE
Rustin, M. (1989) ‘Observing Infants: Reflections on Methods’ in Miller L., M. Rustin,
63
M. Rustin and J. Shuttleworth, (eds.) Closely Observed Infants, London:
Duckworth
Selleck, D. and S. Griffin (1996) ‘Quality for the under threes’ in Pugh, G., (ed.).
Contemporary Issues in the Early Years, London: Paul Chapman and the National
Children’s Bureau (second edition)
Sim, G., S. MacFarlane and J. Read (2006) ‘All work and no play: measuring fun,
usability and learning in software for children’ in Computers & Education 46
pp.235-248
Siraj-Blatchford, J. and D. Whitebread (2003) Supporting Information and
Communications Technology in the Early Years, Maidenhead: Open University
Press and McGraw-Hill Education
Sylva, K., C. Roy and M. Painter (1980) Childwatching at Playgroup and Nursery
School, London: Grant McIntyre
Tizard, B. and M. Hughes (1986) Young Children Learning: Talking and Thinking at
Home and School, London: Fontana (first published 1984)
UsabilityNet, (2006) http://www.usabilitynet.org/tools/r_international.htm#9241-11
Vernon, J. and C. Smith (1994) Day Nurseries at a Crossroads, London: National
Children’s Bureau
Ward, L., D. Howell, L. Wilton, G. Winter, O. Dickinson and S. Gauld (2004) Usability:
CBeebies and CBBC Interactive, Unpublished internal document, BBC
Zeichner, K. and D. Liston (1996) Reflective Teaching: An Introduction, New Jersey:
Lawrence Erlbaum Associates
64
Appendix A
Questions about user testing for the team
1. Why do you think we do user testing?
2. What do you think is the most important part of user testing?
3. What do you see as your role in user testing?
4. Think back to the last user-testing session you went on. What did you think
worked well?
5. What didn’t work well?
6. What do you learn when you go user testing? (Do you learn anything new?)
7. How do you think the children find user testing?
8. Do the children learn anything new?
9. How do you use your user-testing findings?
10. If we didn’t go user testing, how do you think it would affect your work, the
CBeebies site and our output?
11. What one thing would make user testing better, for you?
12. Is there anything you would like to add that we haven’t discussed about user
testing?
65
Appendix B
Interviews with the team about user testing –
notes from their answers for question one
Why do you think we do user testing?
Emma
Research into how children use games, how the games are working, whether the
children enjoy them. Also look at gender and navigation.
Jo
We can’t assume that we know everything and just because we think something’s good
they’ll think it’s good.
You might think it’s easy to remember what you were like as a child and what you liked
as a child and what would appeal but if you have to do it to make sure that the people
who it’s intended for get out of it what they’re supposed to get out of it. Make it
worthwhile.
Katherine
Make sure content is accessible for children and find out what they like and can and
can’t do and might need help with.
Also to know that what we produce is doing what it’s supposed to do.
Christina
To get an accurate view on how correct our content is and aimed at the right level.
What works well and doesn’t work so well. Trying to find gaps for the content we have.
Nikki
To ensure that we can provide most accurate content for our users, our audience.
Accurate: suitable for age range, capabilities, style of learning, enjoyment, requirements.
Eimer
To gauge how children cope with new content that we’re producing. How young
children use the computer from various different backgrounds and if they’re able to
do so on their on own or with parent or carer alongside them.
And also to see what’s popular – what works and what doesn’t. What sort of things? If a
child struggles with something or lacks interest or really captures their imagination, so
pull something more into frame than something else.
Helen
Particularly for our age group essential to see if they can actually do it as can’t make
assumptions.
Also has to be quite instinctive as they can’t read instructions. Games and content need
to be instinctive and can’t tell until you test.
Last three years – skills increased enormously, bigger change than any other age group.
66
First started user testing reception class can’t do it, now most can.
Sandrine
To see if we’re going the right way when we create either new sites or new pieces of
content. Does it make sense for the child, do we make sense for the child.
Navigation particularly – arrows at bottom, does that make sense for them.
To keep in touch with a child’s ability in terms of technical skills – mouse, keyboard.
And very important with children with disabilities as not expert.
Cecilia
In order to see how well our content is received by the audience. To pick up any small
technical points such as ease of playing a game, using the mouse, scrolling, looking
around.
One of the drawbacks right now is that a lot of it is done when something is finished. We
learn what we did wrong or what we could do better next time and take that information
with us.
Like to do at halfway stage of content – and that’s my plan from now on.
At the moment apply to next project and not to current project.
Other facts: very fact of being with audience on regular basis puts you on their
wavelength and also with adults who work with children (i.e. in school or nursery).
Really important to watch how they behave in nursery before even doing user
testing.
Your own children – can be spot on and can be a drawback.
User testing always always valid.
Regularly so always in contact with audience.
Aidan
I haven’t done a lot. But from my point of view from when have done it, usability of a
game, children can use it, age it’s aimed at, functionality, if it works, easy to get
story across, that it works basically.
User testing done – couple of away days, couple of time at Canberra.
Alan
Basically to keep in touch with our audience. Downstairs don’t do any user testing
from what I’ve seen – packed schedules. Now doing a bit more.
New from word go built it into our system.
Really important to keep in constant contact with the audience and see how they respond
to things. Pick up a lot. We’re so competent with the computer we assume that a lot of
kids are and they’re not e.g. kids in White City area don’t have computers at home.
Quite shocking some of their computer skills. Have kids in from some of our producers.
Important to do it across the country – Newcastle, Bristol, Wales. Welsh sites and
content.
What we do, though we don’t get a lot done, is a start and that’s important.
Rachel (recording not great)
67
Age group we work with. How they think and feel, and opinions. Remove ourselves from
cocoon. How they operate and how innovative they can be as well – and what opinions
they bring up. Involved at every stage of developing a game or application or tool.
Liz
Ascertain that everything we’re building and designing is from a user-centred
approach. In terms of checking we’re always putting the user first and thinking about
their needs and what they’re (we’re?) trying to achieve, trying to do when they come to
our websites.
Trusting our own experience or judgement about websites isn’t necessarily a good idea
(though have a lot of internal knowledge) because we’re ‘super users’ and our audience
are normal, every day people. Situation quite different to sitting in the office – they’re at
home, with children, children in school or upstairs in bedroom using it depending on age
group so in order to get a ‘real life’ experience then need to do user testing.
Usability point of view – obviously strive to do our best, that everything works, but only
true test is to put it in front of people. See if they can do it, can find things, without that
prior knowledge we all have.
Debra
Us personally? – questions designed for team? Rephrase to ‘why do you think CBeebies
and/ or CBBC do user testing’
Mainly to see if what we are doing… difference between should and why people think.
In theory we do it to make sure that what we’re doing is what the users want and
that they can actually navigate – two key points.
Should and..? Whether results are applied in a consistent, analytical way (from the user
testing).
In theory to check what users want and can navigate around it.
Sue:
Q: Why do you have the focus groups?
To check that the content we’re putting out and the programmes are working really,
that the children enjoy them, giving them something they want. Also seeing where
the gaps are, seeing what else they’re interested in, that gap with the older ones, what
they’re interested in.
68
Appendix C
Questions about user testing for the team
Brief write-up for Liz and Debra on Thursday
1. Why do you think we do user testing?
Mostly agree with each other. Focus on child alone on PC or using it with an adult –
brought up by K and Ei. i.e. ‘find out what they might need help with’ – is this a valid
part of our user testing? What role do we play as adult partner? Is there any point doing
testing with children on their own?
New: important to watch children away from PC as well.
2. What do you think is the most important part of user testing?
Split between listening/ observing and being prepared and organised.
Also significant minority (mostly newer people, less experienced?) who favour talking to
the children, asking them what they like, guiding them to some extent.
Also emphasis on environment and being at ease – no distractions.
3. What do you see as your role in user testing?
Split again? Emphasis from editorial team on being proactive – being there as a guide,
talking to them, ensuring they have a good experience. Confusing being facilitator and
being observer? (i.e. Cecilia and ‘Observing not done in a stand back watching way but
done in a proactively way so good to have two of you.’)
Only designers and Rachel suggesting having similar roles, taking it in turns, gathering
data, feeding back etc.
4. Think back to the last user-testing session you went on. What did you think
worked well?
5. What didn’t work well?
Interesting focus on content – i.e. answering question ‘what game did the children
respond well to’. But then backed up by ‘if they respond well you feel much better…’
Equipment – technical bit etc.
Environment – distractions, like home, at ease
Ratios of adults and children
Planning – only coming from design team and Rachel (consult experts)
New:
Talk to the adults before we see the children
Need children with certain competence with the mouse/ computer – agree?
69
Communication with teachers/ staff
6. What do you learn when you go user testing? (Do you learn anything new?)
No real themes?
Significant number of answers about what you learn about running a session/ how to user
test/ interacting with children
A few about needing a purpose to learn anything worthwhile? (Sandrine and Aidan)
How to improve it for next time/ feed into future projects
High and low expectations of the audience and its capabilities – look at for child-centred
approach. But also see if any literature on having certain expectations of users as part of
usability?
Not all agree always learn something new.
7. How do you think the children find user testing?
Most think they enjoy it with significant minority emphasising being shy/ it being
intimidating and tiring.
Is this something we ever ask about adults and user testing?
8. Do the children learn anything new?
Shortest set of answers.
This made them think most about the child’s role, more than the previous question.
General agreement that only learning things around what’s out there to do and play, plus
some computer skills but not really.
Maybe as said new they couldn’t focus on the learning process – otherwise an indictment
of our content?
9. How do you use your user-testing findings?
Overwhelmingly about feeding into next or future projects (eight - nine respondents).
Surprised by this as not how I see it but am I in minority? Views on what user testing is
‘for’? i.e. isn’t it about that piece of content and how I can improve that particular
experience?
Scheduling and time – how reports are written.
Not always make immediate changes but if important do try.
10. If we didn’t go user testing, how do you think it would affect your work, the
CBeebies site and our output?
Slight split. Most agree it’s vital but Sandrine, Liz and Debra emphasise how it needs to
be done properly to be worthwhile (and Aidan?).
Thoughts on Sandrine’s comment about evolution of user testing – OK as it is at mo for
‘new researchers’ but not for ‘advanced researchers’. Think is getting at point that it is
70
mostly audience connection and learning basics of our age groups’ competence. Thoughts
from L and D?
11. What one thing would make user testing better, for you?
Time and scheduling
Planning and preparation
Following through – completion
New idea:
User testing not for me, it’s for the kids. Therefore make it more free and easy. Again,
who is it for?
12. Is there anything you would like to add that we haven’t discussed about user
testing?
Timing, scheduling, planning
New ideas:
New independent projects stemming from findings (K, plus H and talk to children about
ideas at planning stage)
Discussion (part of follow through?)
Evolution (see above)
Communication between editorial and design
Putting it actually into practice
71
Appendix D
Ideas that may be useful while discussing what and how you want to change about user
testing (or for later once have action points)
Margaret Carr and Learning Stories (video just seen as well)
Inspired by Te Whairiki (pronounced Te Fareeki) curriculum in New Zealand
4 D’s:
Describing – defining learning but also can be observations of an activity
Discussing – talking with staff (us and early years’ staff, perhaps?), the children and
families (i.e. parents in user testing?) in order to develop, confirm or question an
interpretation
Documenting – recording an assessment but equally, could we use it to record an
observation or user testing experience? You can use text, pictures, photos or a
combination
Deciding – what to do next? Informal or formal planning
‘Learning Story’ observation sheets are used as part of Documenting – something to be
adapted for us? Uses five ‘learning dispositions’ from NZ:
• Finding something of interest
• Being involved
• Engaging with challenge and persisting with difficulty
• Communicating with others/ expressing a point of view
• Taking responsibility
Usability diaries (Kuniavsky and other usability experts)
Ask users to keep a dairy over several days, a week or a month of how they use a website
or the internet
Parent as evaluator (Markopoulos et al)
Tried out with four to six year olds. Parents learn first what to do with a game/ website/
CD-Rom etc. and then follow instructions to become the facilitator with the child. Can be
combined with usability diaries and can be filmed. It takes a few days or weeks but may
be more ‘natural’ to a child – i.e. more as they would normally use something at home
with a parent.
Picture cards (Barendregt)
PECS images used for children aged five to seven to communicate how hard or easy they
found tasks on the computer, how much they liked something etc. Early days and only
one piece of research but quite useful to help children think aloud and articulate what
they thought.
Using webcams of children’s faces together with stationary camera of whole room?
Eye-tracking software/ video
72
Target child observation method (Sylva, Roy and Painter)
Record key behaviour every 30 seconds for 20 minutes in different columns, with no
interpretation:
Activity
|
Language
|
Task |
Social
Adapt for our needs while user testing?
Children’s ability to think and reason at home and school (Donaldson and Tizard and
Hughes)
Donaldson (1978): ‘by the time they come to school, all normal children can show skills
as thinkers and language-users to a degree which must compel our respect’. Donaldson
then finds children have to find new ways of thinking and reasoning once they are at
school.
Tizard and Hughes (1984): four-year-olds learn most and think most at home, away from
school and nursery (irrespective of parental job/ background).
Fun sorter (Sim, MacFarlane and Read) with seven and eight year olds
73
Appendix E
Group A
•
•
Time, timing, scheduling (incl. what user testing is for – when and for how long)
Communication with children and staff/ parents/ carers
Group B
•
•
•
Planning and preparation
Equipment/ technical glitches
Environment
Group C
•
•
•
Sharing findings, documenting, writing up (incl. communication in team and
between editorial and design)
Your role – in parental role, as facilitator, as observer
The role of the child
74
Appendix F
Group discussion part two.
Group A
•
Time, timing, scheduling (incl. what user testing is for – when and for how long)
Problems
Unplanned
Teacher not engaged with process – engaging teachers in objectives is hard
Using only London schools
Budget to allocate from outset
Anarchy
10-11.15 – break time
Fit too much into testing
Solutions
Value in unprompted testing
Aims of testing established
Planning in part way through a game or application
Testing early in design process then test again
Plan reasonable testing milestones into project plan
Build in to budget funds to go around country to focus test
Two types of testing – getting to know audience and product testing
Build relationship with schools
•
Communication with children and staff/ parents/ carers
Very good briefing of objectives of testing
External ‘best’ contacts
Usability – external agencies good for the range of children you use
Database of best teachers
Have sufficient people to help
Time to build rapport
Research – visit/ watch them play
Warm up sessions with schools
Home visits in their own environment
Take care about ‘feeding’ them (or teaching? Can’t read writing, sorry!)
iCan school good example of communication
Prepare your technology
Brief teacher about type of child we need
Presents and fun important
Other points
75
Scotland can do testing in Scotland for us
Timing and role – exhausting as in active role?
Engage the child – stickers and presents
Two types of testing – could combine?
What about communication in the session itself? i.e. working out what tasks we do etc.
Need to prepare more than we do currently
Need time to write up
Maintaining all results, updating
Make changes for external and internal projects, part of bigger, fewer, better
Share results in team meeting (NB therefore design and tech need to be invited)
Group B
•
•
•
Planning and preparation
Equipment/ technical glitches
Environment
Planning and Preparation
Set objectives: what do we want to get out of this?
Break it down into realistic chunks – organise more than one session if necessary.
Have a script or planned format to refer during session.
Build prototypes/scenarios (print/flash).
Do a practice run through.
Ensure you have a good mix and balance of skills in the team.
Plan what roles each person will take on (note taking/facilitator etc.).
Equipment & Tech glitches
Choose locations that have computers – this will relax the kids as they will be familiar
with using them. Move away from going to places locally just for convenience.
Phone the school or nursery first and ensure they have correct software i.e. Flash. Ask
them to double check if uncertain.
Liaise with tech team to ensure someone ‘back at base’ is on hand in case of any live tech
problems.
Environments
Real spaces, nursery and home, with familiar , real set-up
Reduce numbers and distractions: 30 kids is not productive!
Be more prescriptive about what we need, take control of the sessions i.e. we would like
6 children, 3 boys, 3 girls, for 45mins, need 3 computers with flash 5 etc.…
Create a friendly environment, let the children know how helpful they are being and that
there is no right or wrong.
76
Group C
Sharing findings, documenting, writing up (incl. communication in team and between
editorial and design)
Your role – in parental role, as facilitator, as observer
The role of the child
1. Our Role
We have to gain users’ trust but we also have to get them to do what we want them to.
We often have to act as negotiators between groups of children who may be taking turns
or playing together.
We have to ask lots of questions as they don’t tend to give up how they are feeling or
what the experience is like for them freely.
Also need to encourage them so they can have the best experience possible.
We are the outsiders in a different environment so can take us time to settle into our
surroundings – get organised, make sure everything is working.
Don’t always see children in a natural place where they are uninfluenced from their peers
or the familiar people around them.
We are the active observers. Children cannot always read instructions or hear audio on
laptop.
It’s important we also have a non-active observer, facilitator, who can take notes while
the other tester supports the children.
We must make sure we test with all types of children, from different abilities,
backgrounds, gender, ages and from different parts of the country. The advantage of
testing with older children is that they can often be observed better as they don’t need as
much help.
The broader the testee, the more we know how the content is used, how well it is used
and how it could be improved.
At the same time it is important to encourage the child by not giving someone of their
ability a piece of content that will be too hard for them to test or understand – although
this is hard to know in advance if the child is a new tester.
2. Role of Child
To help us find out how well a piece of content had been made or how it could be
improved.
To express their honest opinion about how they feel when using the piece of content.
They are the truth and evidence of what we do.
Must test similar content produced by our competitors at the same time as testing our
own content, to see how it compares.
They are the window into our audience, their lives, what they like and dislike.
They can help us improve our content for all.
They need to have some degree of computer literacy to get the most out of the content.
77
3. Documentation
We could film testing sessions to keep a record – not always practical though.
We could create a generic form that gets filled in during or upon return of testing – this
could include all relevant details about the child – age, background, ability, gender, what
was tested and how they got on.
This would help in subsequent sessions – if the same or a similar game was being tested
we would know what things we may need to look out for.
It would also help in the production of new games – we could see what elements children
found difficult or what they really enjoyed.
78
Appendix G
Project Definition Document
Project name: CBeebies User Testing
Deliverables: a new user testing process in place for the CBeebies interactive services
with supporting documentation on the Children's Interactive intranet (ciki)
Objectives:
Primary:
To look at the current problems of CBeebies user testing and find a way to fix them
To improve the user testing experience for both the CBeebies team and the children with
whom we go user testing
Secondary:
To evaluate whether child-centred theories can form the basis for CBeebies user testing
To evaluate how useful ‘traditional’ usability theories are for children under six – can we
maintain or improve upon user-centred design for our audience?
To maintain the best online content for children under six in the UK through effective
user testing
Scope:
1. Inside: user testing of the CBeebies website (broken down into testing of Flash
games and navigation), CBeebies interactive TV, CBeebies WAP service and
PDA, Video on demand and IPTV. User testing with children and parents together
and children on their own.
2. Outside: user testing of CBBC services, user testing of CBeebies Grown-ups site
and WAP service.
Assumptions:
User testing of CBeebies needs improving - agreed by design team, editor and editorial
team
CBBC has a user testing process in place that works for older children (6+)
Risks:
Ideas to improve the process do not actually work for the audience
Not enough time is set aside to concentrate on the project by managers
Design, editorial and tech do not agree how to set up new process
Project manager has particular aims that the group do not buy into
External influences from marketing, scheduling within BBC Children's
Take too academic approach, need to be really hands-on
Think only within BBC - need to learn from competitors and toy manufacturers
Key stakeholders:
Head of Children's Interactive and On Demand
Editor of CBeebies Interactive
CBeebies interactive team
79
Design team lead and design team
Interactive TV team
Tech team
Project Sponsors:
Design Lead, BBC Children’s Interactive (Liz)
Editor of CBeebies website (Rachel)
Senior Producer, Children’s iTV (Anthony)
Project Manager:
Olivia Dickinson
Team Members:
Leo Sen
Joanne Patterson
Aidan O'Brien
Katherine Monk
Helen Stephens
Dave Howell
Debra Trayner - usability expert
Sandrine Dordé - special needs expert
Key Milestones:
Identify key roles of each team member
Identify key problems of user testing – already done in group discussion, need to go over
Identify solutions to those problems – again, already done to some extent in group
discussion
Set up panel of ten children with different abilities, background and ages - need to
identify what personas we want
Make contact with competitors and toy manufacturers?
Try out solutions - practical, hands-on - go user testing
Approve or discard solutions
Agree new process
Try out new process
Finalise process and share with stakeholders
Write guide to new process
Apply new process, with stakeholders
Budget, resources and constraints:
Currently no budget but two options:
Is money available from Marketing, Communication and Presentation?
could use some of redesign user testing budget?
Limited time - need to negotiate time with line managers?
Project approval date: October 2006? (do want to set a date as point of project is short
term to put new process in place)
80