Toward a robotic companion design

Transcription

Toward a robotic companion design
Toward a robotic
companion design
WRUT LIREC Group
Institute of Computer Engineering, Control and Robotics,
Wrocław University of Technology,
ul. Janiszewskiego 11/17,
50–372 Wrocław, Poland
email: [email protected]
Wrocław 2008
Authors:
Krzysztof Arent
Mariusz Janiak
Jan Kędzierski
Bogdan Kreczmer
Łukasz Małek
Robert Muszyński
Adam Oleksy
Krzysztof Tchoń
Marek Wnuk
Contents
Chapter 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
5
Chapter 2. Overview of social robot designs . . . . . . . . . . . . . . . . . . . . . .
9
2.1.
Mechanoid robots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9
2.1.1.
SPUTNIK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9
2.1.2.
SR4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
9
2.1.3.
PeopleBot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.1.4.
Bryn Mawr . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
2.1.5.
BIRON . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.1.6.
Eldi . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
2.1.7.
GRACE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.1.8.
RoboX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.1.9.
HERMES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.1.10. uBot5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.11. PaPeRo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.12. Pearl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.2.
2.3.
2.4.
Humanoid robots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.2.1.
Kaspar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.2.2.
SIG2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.2.3.
Kismet/MERTZ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
2.2.4.
WE-4RII
2.2.5.
Toyota Rolling Robot
2.2.6.
Wakamaru . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
2.2.7.
EMIEW2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
2.2.8.
MDS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2.2.9.
Monty . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
Android robots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
2.3.1.
BARTHOC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
2.3.2.
ROMAN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
2.3.3.
Albert . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
Zoomorphic robots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2.4.1.
iCat . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2.4.2.
Pleo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2.4.3.
Leonardo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
Chapter 3. Overview of social robot components . . . . . . . . . . . . . . . . . . . 35
3.1.
Mobile bases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
Contents
3.2.
3.3.
3.4.
3.1.1.
Balancing platforms overview . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.1.2.
Concluding remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
Arms and hands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
3.2.1.
Arms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
3.2.2.
Hands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
3.2.3.
Concluding remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
Vision acquisition system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
3.3.1.
Commercial vision systems
. . . . . . . . . . . . . . . . . . . . . . . . . . 45
3.3.2.
Pan-Tilt-Zoom . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
3.3.3.
Camera interfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
3.3.4.
Concluding remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
Sound acquisition and emission systems . . . . . . . . . . . . . . . . . . . . . . . . 60
3.4.1.
Microphones . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
3.4.2.
Analog signal conditioning . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
3.4.3.
A/D converters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
3.4.4.
Testbed for sound acquisition system . . . . . . . . . . . . . . . . . . . . . 63
3.4.5.
Loudspeakers and power amplifiers . . . . . . . . . . . . . . . . . . . . . . 63
3.4.6.
Concluding remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
Chapter 4. Body specification of FLASH . . . . . . . . . . . . . . . . . . . . . . . . 67
4
4.1.
Basic technical features . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
4.2.
Balancing platform Cosmos . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
4.3.
Arms and hands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
4.4.
Concluding remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
Chapter 1
Introduction
This report contains an overview of representative, modern designs of social robots and
their components, that has motivated the specification of a robotic companion to be
built within the LIREC project. Initially, the LIREC robotic companion should serve
as a research platform for conducting the LIREC research and experimental work. Ultimately, the robot should be converted into a long term robotic companion equipped with
memory, learning, cognitive, social, and migration capabilities. The robot specification
has been formulated by taking into consideration both the LIREC’s objectives and the
state of the art in the area of social robotics. This specification, presented in Chapter
4 of this report, is concerned mainly with the ”body” of the new robot. The robot’s
tentative name could be FLASH (Flexible LIREC Autonomous Social Helper).
The robot is to be able to perform autonomously, in stable and believable way, in selected
scenarios, e.g.:
1. Fetch and carry, the robot can accomplish the following functions:
• indoor navigation (mobile platform, bumpers, infrared and ultrasonic sensors,
laser range finders, cameras),
• reception and understanding of voice instructions (auditory system, neck, torso,
a software from INESC-ID, a software from HW),
• localising, approaching, following or leading the user (vision system, ultrasonic
and proximity sensors),
• object recognition and grasping (vision system, hands)
• expression of emotions (vision system, auditory system, face, torso, neck).
A possible experiment: the user and the robot stay in the same room, the user asks
the robot to approach, the robot turns towards the user, localises him/her, expresses
a satisfaction that it can be helpful, finds its way to the user, approaches the user,
stops (bows asking: what I can do for you?), then the user makes the robot to grasp
an object and instructs the robot to deliver the object to a given destination inside
the room, the robot follows the instruction, places the object, confirms that the task
has been done and returns to its initial localisation. A variation of this may consist
in asking the robot to go and bring a given object lying on a table.
2. Agent as cognitive prosthetic, the following activities can be done autonomously:
• indoor navigation (mobile platform, bumpers, infrared and ultrasonic sensors,
laser range finders, cameras),
• human localisation and recognition (vision system, auditory system, a software
from QM),
• searching for objects, objects recognition and grasping (vision system, hands, a
software from QM),
• voice and/or gesture communication (loudspeakers, speech synthesiser, hands,
neck, torso, a software from INESC-ID),
• expression of emotions (head, hands),
• reminding the user of some actions (Internet connection, speech synthesiser, a
software from INESC-ID).
A possible experiment: the user asks the robot of which medicines he/she should
take, the robot mentions 3 types of pills, the user asks where he/she has left these
pills, then the robot points toward the pills’ location with its hand, the user asks the
robot to bring him/her a specific box with pills, the robot moves toward the box and
Chapter 1. Introduction
brings it to the user, the user thanks the robot, the robot expresses a satisfaction of
well done service.
3. Teaching proxemic preferences, the robot should do autonomously the following:
• voice/gesture/facial expression communication (loudspeakers, speech synthesiser,
hands, head, a software from INESC-ID),
• sound direction recognition (auditory system, a software from INESC-ID),
• user’s recognition (vision system, auditory system, a software from QM),
• emotion recognition (vision system, auditory system, a software from QM, a software from HW),
• turning head (eyes) toward the user (head, vision system),
• localisation and approaching the user (mobile platform, vision system, auditory
system, ultrasonic and proximity sensors),
• instruction following (auditory system, mobile platform, hands, head, a software
from INESC-ID, a software from HW),
• migration to other embodiments (a software from UH, HW, WRUT),
• learning (all sensor systems, memory, a software from HW).
A possible experiment: 3 users define their proxemic preferences, the robot memorises
them, then one of the users asks the robot to approach him/her, the robot approaches
in accordance with the definition, but the user get frightened by the robot and
communicates using words/gestures/face expressions that the robot has approached
too closely, the robot adjusts its distance, the user praises the robot, the robot shows
its content. A version of this may consists in shouting at the robot that gets too close
to the user, the robot shows a shame, apologises, and adjusts its position. Another
version involves a migration: the users communicate their preferences to robot1,
then robot1 migrates to robot2, much bigger or uglier whose close presence cannot
be tolerated anymore, so an adjustment of proxemic data must be made.
4. Traveller’s companion This scenario includes many elements of the previous scenarios accomplishable in an autonomous way.
As we have already said, our prospective objective is to design a socially interactive
robot whose behaviour (e.g. cues provided by gestures, face expression, and body language) is stable and highly believable. In order to achieve this objective, the mechanical
and control system design of FLASH will make use of the research results obtained
at WRUT. In the design process a contribution from other partners, concerned with
psychological, ethological, and embodiment issues will be of fundamental significance.
Taking into account the scenario requirements, and referring to a feasibility study based
on an overview of social robotic designs presented in this report, we have derived the
following characterisation of FLASH:
FLASH is a mechanoid robot, designed to accompany a human. This is a mobile robot
mounted on a balancing wheeled mobile platform. The robot’s body consists of a torso,
a pair of hands, a neck, and a head. FLASH is able to navigate autonomously through
the home space, and safely approach people. It has advanced interaction capabilities
with the external world, such as object detection, people recognition as well as auditory,
visual, and gesture communication skills. The robot will also be able to accomplish some
limited grasping/manipulation. The FLASH’s control system is based on a network of
microcontrollers, DSPs, and PCs running Linux. Its control algorithms unify classical
and intelligent control strategies. A fast Internet connection will provide the robot with
a constant link to the electronic world. The robot’s design complies with its role in
robotic performance scenarios realised in LIREC’s showcases. As a result of the research
collaboration within the LIREC consortium, the robot will meet the requirements of a
long term robotic companion. The robot’s design will have remarkable flexibility and
openness. Its software is intended to be totally open source. FLASH is going to be a
low-cost, completely EU design.
The robot architecture admits agents distributed among several computers which can
communicate with each other wirelessly by means of YARP (Yet Another Robotic Plat-
6
form). This will make the robot a testing platform, open for contributions from other
partners.
This report is composed in the following way. The next chapter presents a social robots
panorama. Chapter 3 makes a review of social robot components. Special attention is
paid to mobile platforms and hands as well as a vision and auditory equipment. The
content of chapters 2 and 3 fosters a kind of feasibility study of a robotic companion
hardware. This results in the specification of the robot FLASH provided in chapter 4.
7
Chapter 2
Overview of social robot designs
This chapter makes an overview of 27 contemporary robot designs, manifesting social
capabilities. Below we present not only purely social robot constructions, but also the
robots whose behaviour contains social elements or may serve as an inspiration for social
robot designers. Among these 28 robots 6 come from Japan, 12 from US, and 7 have
been elaborated in EU.
The robots have been divided into 4 categories, based on their appearance: mechanoid
(12), humanoid (9), android (3), and zoomorphic (3). In the overview we have purposely
omitted walking robots (like Asimo, Qrio, Feelix, iCub), actroids (like Geminoid HI-1,
DER2, Saya). Also, we have not included the classical, ancestor designs like Shakey,
Flakey and Cog.
2.1. MECHANOID ROBOTS
2.1.1. SPUTNIK
General description
SPUTNIK is a small, portable mobile platform designed for research applications in
human-robot interactions, navigation, image processing and recognition, teleoperation,
remote sensing, etc. The robot may also serve to develop robotic assistants or service
units. The robot is presented in figure 2.1.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Dr Robot Inc., Canada,
Web: http://www.drrobot.com/products$\-$item.asp?itemNumber=SPUTNIK,
Purpose: research platform,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 0.47m, diameter 0.40m, weight 6.1kg, speed 1m/s,
Components: mobile platform, neck, head,
Scene detection: vision camera, infrared and ultrasonic sensors, pyroelectric human
motion sensor, audio modules,
Robot-human communication: audio signals, movable head, eyes, and lips,
Autonomy: indoor navigation,
Control system: onboard DSP, external PC running Windows or Linux,
Openness of design: yes,
Availability, price: on sale, EUR 2987.
2.1.2. SR4
General description
The robot SR4 is a commercial mobile robot that may serve as an educational or an
application development platform. SR4 can be customised with user’s hardware and
software. The robot is presented in figure 2.2.
Chapter 2. Overview of social robot designs
2.1. Sputnik
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Smartrobots Inc., US,
Web: http://www.smartrobots.com,
Purpose: education/application development platform,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 0.9m, diameter 0.45m, max speed 0.3m/s,
Components: mobile platform, body,
Scene detection: ultrasonic sonar, infrared sensors, vision cameras, proximity sensors, position triangulation system,
Robot-human communication: microphone, loudspeakers, touchscreen, speech
synthesiser,
Autonomy: point-to-point motion, collision avoidance, navigation,
Control system: a RISC machine running Linux,
Openness of design: yes,
Availability, price: on sale, SR4-P costs USD 4495-5395.
2.1.3. PEOPLEBOT
General description
This robot provides a mobile platform for service or performance robots in many application areas. The robot has a capability of indoor navigation, is equipped with a number
of sensors, and a gripper. Employed as a mobile platform, a.o. in BIRON and Bryn
Mawr robots described below. The robot is presented in figure 2.3.
Specific features
•
•
•
•
•
•
•
•
•
10
Manufacturer: MobileRobots, US,
Web: http://www.activrobots.com/ROBOTS/peoplebot.html,
Purpose: service and performance mobile platform,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 1.1m, width 0.47m, depth 0.38m, safe speed 0.8m/s,
Components: mobile base (lower platform), body, upper platform,
Scene detection: ultrasonic sonars, infrared sensors, microphones, laser range
finder, cameras,
Robot-human communication: microphones, loudspeakers, LEDs, voice recognition system, stereovision system, touchscreen, gripper,
2.1. Mechanoid robots
2.2. Robots from SR4 family
2.3. PeopleBot
11
Chapter 2. Overview of social robot designs
2.4. Bryn Mawr Tour Guide
•
•
•
•
Autonomy: indoor navigation, collision avoidance,
Control system: onboard RISC controller, optional onboard computer running
Linux or Windows,
Openness of design: limited,
Availability, price: on sale EUR 24000.
2.1.4. BRYN MAWR
General description
The Bryn Mawr Tour Guide is another robot built on the basis of the PeopleBot. An
extra equipment includes a sensitive microphone and loudspeakers. The robot’s architecture consists of 3 basic systems: navigation, localisation, and voice communication.
The robot is presented in figure 2.4.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
•
12
Manufacturer: Bryn Mawr College, US,
Web: http://cs.brynmawr.edu/TourGuide/,
References: C. Chiu, The Bryn Mawr tour guide robot, PhD Thesis, http://cs.
brynmawr.edu/TourGuide/finished.html,
Purpose: research platform,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 1.2m, width 0.47m, depth 0.38m, safe speed 0.8m/s,
Components: PeopleBot, camera,
Scene detection: microphones, laser range finder, sonars, cameras,
Robot-human communication: microphones, loudspeakers, speech recognition
and synthesis, dialoguing, carrying out voice requests,
Autonomy: indoor navigation, collision avoidance,
Control system: as PeopleBot,
Openness of design: none,
Availability, price: unavailable.
2.1. Mechanoid robots
2.5. BIRON
2.1.5. BIRON
General description
BIRON has been dedicated to the investigation of human-robot interactions. The interaction begins with the identification by the robot of a potential interlocutor. Then the
robot focuses attention on him/her, and starts a dialogue. The robot’s appearance is not
of primary importance; BIRON has been built onboard of the PeopleBot mobile platform, and uses the PeopleBot’s navigation system. The robot is presented in figure 2.5.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: University of Bielefeld, Germany,
Web: http://www.techfak.net/ags/ai/projects/BIRON/welcome.html,
References: A. Haasch, et al., BIRON – The Bielefeld robot companion, http:
//www.techfak.net/ags/ai/publications/papers/Haasch2004-BTB.ps.gz,
Purpose: robot companion,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 1.4m, width 0.47m, depth 0.38m, safe speed 0.8m/s,
Components: PeopleBot, touchscreen, camera,
Scene detection: stereo microphones, laser range finder, sonars, cameras, touchscreen,
Robot-human communication: microphones, loudspeakers, speech recognition
and understanding, dialoguing, people tracking, carrying out voice requests,
Autonomy: indoor navigation, collision avoidance, people tracking,
Control system: as PeopleBot,
Openness of design: none,
Availability, price: unavailable.
13
Chapter 2. Overview of social robot designs
2.6. Eldi
2.1.6. ELDI
General description
The robot Eldi was built at the University of Las Palmas, Gran Canaria, Spain. Since
1999 the robot has been a main attraction of the Elder Museum of Science and Technology at Las Palmas. The robot’s body consists of two components: the commercial mobile
platform Nomadics XR4000 that provides Eldi with basic mobility and sensoric capabilities, and the torso that contains a control system, a radio communication system, a
vision system, a system of sound and picture transmission, a touchscreen, a microphone,
and loudspeakers. The torso is equipped with a head mounted on a 2 DOF neck. The
robot is presented in figure 2.6.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
•
14
Manufacturer: University of Las Palmas, Gran Canaria, Spain,
Web: http://mozart.dis.ulpgc.es/eldi/home2.html,
References: M. Castrillon Santana, et al., Eldi’s activities in a museum, http:
//citeseerx.ist.psu.edu/viewdoc/summary?doi10.1.1.26.9290,
Purpose: museum guide,
Appearance: mechanoid,
Mobility: 4 wheel mobile platform, holonomic,
Physical parameters: height 1.6m,
Components: mobile platform, torso, head,
Scene detection: vision system, ultrasonic sonars, infrared and tactile sensors,
radiophares, laser detectors,
Robot-human communication: touchscreen, microphone, loudspeakers, speech
recognition and synthesis, face recognition, movable head,
Autonomy: localisation, point-to-point motion, collision avoidance,
Control system: a network of PCs running Windows and Linux,
Openness of design: none,
Availability, price: unavailable.
2.1. Mechanoid robots
2.7. Grace
2.1.7. GRACE
General description
GRACE (Graduate Robot Attending ConferencE) is a robot whose main task consists
in attending conferences, registering the conference events, and interacting with humans
using a voice communication system. The robot has been mounted on the mobile platform B21 built by RWI. GRACE distinguishes itself by a graceful, animated female
face that shows up on an LCD screen, and serves for expressing robot’s emotions. A
coordination of the speech with movements of the lips increases the impressiveness of
GRACE. A male companion called GEORGE has also been designed. The robot is
presented in figure 2.7.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Carnegie Melon University, Naval Research Laboratory, Northwestern University, Metrica Inc., Swarthmore College, US,
Web: http://www.ri.cmu.edu/projects/project_522.html,
References: R. Simmons, et. al., GRACE: An autonomous robot for the AAAI
robot challenge, http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.
61.1464,
Purpose: conference robot,
Appearance: mechanoid,
Mobility: 4 wheel mobile platform,
Physical parameters: height 1.6m, weight 150kg, max speed 0.9m/s,
Components: mobile platform, torso, LCD screen,
Scene detection: vision system, ultrasonic sonars, infrared and tactile sensors, laser
range finder,
Robot-human communication: microphone, loudspeakers, LCD screen, animated face, emotion expression, speech recognition and synthesis, voice-lips coordination,
Autonomy: localisation, point-to-point motion, collision avoidance, carrying out
voice requests,
Control system: a network of processors running Linux, a laptop running Windows
for speech recognition,
Openness of design: none,
Availability, price: unavailable.
15
Chapter 2. Overview of social robot designs
2.8. RoboX
2.1.8. ROBOX
General description
RoboX is a typical example of a guide robot whose function consists in a voice communication with humans. The main task of RoboX has been guiding exhibition visitors in
accordance with their expectations and an established visit programme. The robot can
speak English, German, French and Italian. The design of RoboX has been adapted to
the navigation in crowded places. The robot is presented in figure 2.8.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Federal Institute of Technology, BlueBoticsSA, Switzerland,
Web: http://www.bluebotics.com/entertainment/RoboX,
Purpose: exhibition robot,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 1.65m,
Components: mobile platform, slim torso, head with eyes,
Scene detection: vision cameras, laser scanner, tactile plates, foam bumpers,
Robot-human communication: microphone, loudspeakers, LED matrix, keyboard, movable eyeballs and eyebrows, face tracking and recognition,
Autonomy: point-to-point motion, collision avoidance, navigation, object localisation, people tracking,
Openness of design: none,
Availability, price: on sale EUR 48500, also for hire EUR 6000/day.
2.1.9. HERMES
General description
HERMES is a combination of the mobile platform Athene II and manipulators not
requiring calibration. Similarly to humans, HERMES receives information from the
16
2.1. Mechanoid robots
2.9. Hermes
external world through 4 sensoric channels: vision, auditory, kinesthetic and tactile.
The robot is able to explore the environment or to carry out human’s requests, like
bringing a glass of water from the kitchen or delivering a book to the library. An
Internet connection enables HERMES to inform the human on the current date, time,
or weather conditions. The questions and requests are received by the robot via a voice
interface accepting English, German or French. The robot is presented in figure 2.9.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Munich University, Germany,
Web: http://www.uniwbw.de/robotics/robots/hermes/,
References: R. Bischoff, V. Graefe, HERMES - a versatile personal robotic assistant, Proc. IEEE, a special issue on Human Interactive Robots for Psychological
Enrichment, pp. 1759-1779, 2004,
Purpose: service robot,
Appearance: mechanoid,
Mobility: 4 wheel mobile platform, holonomic,
Physical parameters: height 1.85m, width 0.7m, depth 0.7m, weight 250kg,
Components: mobile platform, torso, hands, head,
Scene detection: stereovision and tactile sensors,
Robot-human communication: microphone, loudspeakers, speech recognition
and synthesis, gestures, head movements, acceptable as a friendly partner,
Autonomy: point-to-point motion, collision avoidance, navigation, map building,
indoor exploration,
Control system: a multiprocessor system including PCs and DSPs,
Openness of design: none,
Availability, price: unavailable.
17
Chapter 2. Overview of social robot designs
2.10. uBot5
2.1.10. UBOT5
General description
This is the latest member of the uBot family of small research platforms designed for
mobile manipulation. uBot5 moves on a balancing wheeled platform. The robot has
been equipped with an LCD touchscreen and a webcam, and used for telepresence applications. The robot is presented in figure 2.10.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: University of Massachusetts, Amherst, US,
Web: http://www-robotics.cs.edu/Robots/UBot-5,
References: P. Deegan, et al., Designing a self-stabilising robot for dynamic mobile manipulation, Robotics: Science and Systems - Workshop on Manipulation for
Human Environments, Philadelphia, Pennsylvania, 2006,
Purpose: mobile manipulation,
Appearance: mechanoid,
Mobility: self-balancing mobile platform, nonholonomic,
Physical parameters:
Components: mobile platform, body, hands, LCD screen,
Scene detection: webcam,
Robot-human communication: touchscreen,
Autonomy: indoor navigation, manipulation,
Openness of design: none,
Availability, price: unavailable.
2.1.11. PAPERO
General description
An amiable jabbering robot intended to be a partner and a human domestic companion.
When not talking, the robot starts wandering around the room. The robot is capable of
interacting with humans, learning, and shaping its character. If greatly praised, PaPeRo
may begin to dance, it can also express its feelings by dancing. PaPeRo can transmit
the information found on the Internet, switch on and off household appliances, and be
a partner in several games. The robot is presented in figure 2.11.
18
2.1. Mechanoid robots
2.11. PaPeRo
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: NEC Corporation, Japan,
Web: http://nec.co.jp/robot/english/intro/intro01.html,
Purpose: robotic companion,
Appearance: mechanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 0.385m, width 0.248m, depth 0.245m, weight 5kg,
max speed 0.2m/s,
Components: mobile platform, torso, head,
Scene detection: microphone, CCD cameras, ultrasonic sonars, infrared sensor,
Robot-human communication: microphone, loudspeaker, LEDs, sound detection and recognition, face detection, identification and tracking, speech recognition,
expressive dance,
Autonomy: indoor wandering, people following, various spontaneous activities,
Openness of design: none,
Availability, price: on sale USD 41000.
2.1.12. PEARL
General description
Pearl is a robot developed within the project “PERSONAL ROBOTIC ASSISTANTS
FOR THE ELDERLY” to provide a social robot capable to perform home-care of elderly
people. The main tasks of robot are: intelligent reminding, tele-presence, social interactions. The robot performances were tested while interacting with elderly residents of
retirement community centres. The robot is presented in figure 2.12.
Specific features
•
•
•
•
•
Manufacturer: Carnegie Mellon University, US,
Web: http://www.cs.cmu.edu/~nursebot,
References: M. Bennewitz, et al., Learning Motion Patterns of Persons for Mobile
Service Robots, Proc. Int. Conf. on Robotics and Automation (ICRA), Washington,
DC, pp. 3601-3606, 2002, M. Montemerlo, et al., Experiences with a mobile robotic
guide for the elderly, Proc. 18th National Conf. on Artificial Intelligence (AAAI-02),
pp. 587-592, 2002,
Purpose: social robot,
Appearance: mechanoid,
19
Chapter 2. Overview of social robot designs
2.12. Pearl and Flo
•
•
•
•
•
•
•
•
•
Mobility: wheeled mobile platform, nonholonomic
Physical parameters: height 1.30m,
Components: mobile platform, body, head, touchscreen,
Scene detection: proximity sensors, microphones, vision system,
Robot-human communication: speech recognition and synthesis, motion detection,
Autonomy: point-to-point motion, collision avoidance, navigation, object localisation, people tracking,
Control system: 2 single board computers (PCM-5864) plus 2 PCs,
Openness of design: none,
Availability, price: unavailable.
2.2. HUMANOID ROBOTS
2.2.1. KASPAR
General description
Kaspar (Kinesics and Synchronisation in Personal Assistant Robotics) is a humanoid
robot developed at the University of Hertfordshire. The robot has an 8 DOF head and
neck, and is able to move its hands. Its face is a silicon-rubber mask, equipped with 2
DOF eyes fitted with video cameras, and a mouth capable of opening and smiling. The
robot is presented in figure 2.13.
Specific features
•
•
•
•
•
•
•
•
•
•
20
Manufacturer: University of Hertfordshire, UK,
Web: http://kaspar.feis.herts.ac.uk/,
Purpose: human-robot interactions, autism therapy,
Appearance: humanoid,
Mobility: none,
Physical parameters: child-sized,
Components: torso, legs, hands, neck, head,
Scene detection: microphone, cameras,
Robot-human communication: facial expressions,
Autonomy: none,
2.2. Humanoid robots
2.13. Kaspar
2.14. SIG i SIG2
•
•
•
Control system: network of microcontrollers, PC running Windows,
Openness of design: yes,
Availability, price: unavailable.
2.2.2. SIG2
General description
SIG2 is a humanoid robot intended as a research platform dedicated to the investigation
of human-robot interactions. The robot is an improved version of the robot SIG designed
within the Kitano Symbiotic System Project. The modifications are concerned mainly
with the sound acquisition and recognition system, and have been achieved by equipping
the robot with a soft skin, re-shaping the ears, eliminating resonances of the casing, and
muffling the drive system. The robot is presented in figure 2.14.
Specific features
•
•
•
Manufacturer: Kyoto University, Japan,
Web: http://www.symbio.jst.go.jp/symbio/SIG/,
References: H. G. Okomo, et al., Social interaction of humanoid robot based on
audio-visual tracking, Proc. 18th Int. Conf. on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, Lecture Notes in AI, Springer-Verlag, Cairns, Australia, 2002,
21
Chapter 2. Overview of social robot designs
2.15. Kismet
•
•
•
•
•
•
•
•
•
•
Purpose: research platform,
Appearance: humanoid,
Mobility: none,
Physical parameters: height 0.6m,
Components: stationary base, torso, hands, head,
Scene detection: CCD cameras, microphones,
Robot-human communication: microphones, loudspeakers, speech recognition
and synthesis, face recognition, dialoguing,
Autonomy: auditory and visual object tracking,
Openness of design: none,
Availability, price: unavailable.
2.2.3. KISMET/MERTZ
General description
Kismet is an expressive robot constructed in the Artificial Intelligence Lab of MIT for the
investigation of human-robot interactions at the level of infant - caretaker. Essentially,
the robot consists of a head mounted on a stationary base. It has been endowed with
communication channels resembling those of humans: visual, auditory, and proprioceptive. In reaction to its environment the robot can show emotions ranging from calm to
disgust, and from sad to surprise. Kismet can also talk with humans, adjust its gaze
direction and its head orientation. While talking the robot recognises the intentions
of the human. The robot is presented in figure 2.15. MERTZ is a Kismet’s successor,
with aesthetic face and more efficient control systems, equipped with new learning and
human recognition capabilities. MERTZ can also mimic the human behaviour, and show
its attitude toward humans. The robot is presented in figure 2.16.
Specific features
•
•
•
•
•
22
Manufacturer: MIT, US,
Web: http://www.ai.mit.edu/projects/humanoid-robotics-group/kismet and
www.people.csail.mit.edu/lijin/robot.html,
References: C. L. Breazeal, Designing Sociable Robots, MIT Press, Cambridge,
Mass., 2002,
Purpose: research platform,
Appearance: humanoid,
2.2. Humanoid robots
2.16. MERTZ
•
•
•
•
•
•
•
•
•
Mobility: none,
Physical parameters: infant’s head,
Components: base, neck, head,
Scene detection: vision system, auditory system,
Robot-human communication: microphone, loudspeakers, facial expressions, vocalisation, gaze direction, head orientation, dialoguing,
Autonomy: detection and reaction to objects in the scene,
Control system: a network including Motorola 68332 microprocessors and PCs
running QNX, Linux and Windows NT,
Openness of design: none,
Availability, price: unavailable.
2.2.4. WE-4RII
General description
The Waseda WE-4RII robot (Waseda Eye No.4 Refined II) belongs to the emotion
expression humanoid robots family constructed at the Waseda University. The robot
consists of a movable torso equipped with a pair of hands and a head mounted on a 4
DOF neck. A distinguished feature of the robot is that its face is able to blush. The
robot has a remarkably reach collection of sensors, including visual, auditory, tactile,
23
Chapter 2. Overview of social robot designs
2.17. WE-4RII
temperature, and olfactory. WE-4RII has been devised to develop new mechanisms and
functionalities for naturally communicating humanoid robots. The robot is presented in
figure 2.17.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Waseda University, Japan,
Web: http://www.takanishi.mech.waseda.ac.jp/research/we/we-4rII/
index.htm,
References: H. Miwa, et al., Effective Emotional Expressions with, Emotion Expression Humanoid Robot WE-4RII, Proc. IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, pp. 2203-2208, 2004
Purpose: social robot,
Appearance: humanoid,
Mobility: none,
Physical parameters: height 0.97m, weight 59.3kg,
Components: torso, hands, neck, head,
Scene detection: vision, auditory, tactile, olfactory systems,
Robot-human communication: loudspeaker, facial expressions, gesticulation, determination of the sound direction, blushing,
Autonomy: detection and reaction to objects in the scene,
Control system: 3 PCs running Windows,
Openness of design: none,
Availability, price: unavailable.
2.2.5. TOYOTA ROLLING ROBOT
General description
This is a representative of 3 Toyota partner robots, endowed with humanoid features
such as agility and kindness, as well as an ability to skilfully operate diverse devices and
play the role of a personal assistant. The rolling robot moves on a balancing platform,
and has artificial lips and dextrous hands enabling the robot to play musical instruments
(trumpet, violin). The robot is presented in figure 2.18.
Specific features
•
•
•
24
Manufacturer: Toyota Motor Corporation, Japan,
Web: http://www.toyota.co.jp/en/special/robot/,
Purpose: partner robot,
2.2. Humanoid robots
2.18. Toyota Robots
•
•
•
•
•
•
•
•
•
Appearance: humanoid,
Mobility: balancing platform,
Physical parameters: height 1m, weight 35 kg,
Components: mobile base, torso, hands, head,
Scene detection: automotive technology sensors,
Robot-human communication:,
Autonomy: indoor navigation, manipulation,
Openness of design: none,
Availability, price: unavailable.
2.2.6. WAKAMARU
General description
Wakamaru has been devised to live with humans as a robotic companion. The robot acts
in accordance with the home schedule that it learns from the observations of inhabitants,
and actively engages itself into a conversation. Wakamaru’s appearance mimics that of
a 12th century Japanese warrior Ushiwakamaru. The robot is equipped with a pair
of 4 DOF hands and a head moved by a 3 DOF neck. It can be a home guard, that
in emergency is able to call for help or contact the owner via Internet. Thanks to a
sophisticated equipment Wakamaru is completely autonomous. The robot is presented
in figure 2.19.
Specific features
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Mitsubishi Heavy Industries, Ltd. Japan,
Web: http://www.mhi.co.jp/kobe/wakamaru,
Purpose: social robot,
Appearance: humanoid,
Mobility: wheeled mobile platform,
Physical parameters: height 1m, diameter 0.45m, weight 30kg, max speed 1km/h,
Components: mobile platform, torso, hands, head,
Scene detection: omni-direction and front cameras, microphones, touch/force sensors, ultrasonic sonars, infrared sensors,
Robot-human communication: microphones, loudspeakers, eye expressions, face
detection and recognition, voice recognition, speech synthesis,
Autonomy: point-to-point motion, collision avoidance, indoor navigation, people
following,
Control system: a multiprocessor configuration under Linux,
25
Chapter 2. Overview of social robot designs
2.19. Wakamaru
2.20. EMIEW and EMIEW2
•
•
Openness of design: none,
Availability, price: can be hired for USD 1000/day.
2.2.7. EMIEW2
General description
The robot EMIEW2 has been devised as an assistant for humans and it is a successor of
EMIEW. Its characteristic component is a self-balancing mobile platform enabling the
robot to move fast and agile. EMIEW2 can localise and recognise voice signals as well
as communicate with humans by means of speech and gestures. The robot is presented
in figure 2.20.
Specific features
•
•
26
Manufacturer: Hitachi, Japan,
References: http://www.hitachi.com/rd/research/emview2$_$01.html,
2.2. Humanoid robots
2.21. MDS
•
•
•
•
•
•
•
•
•
•
Purpose: social robot,
Appearance: humanoid,
Mobility: self-balancing mobile platform, wheeled-leg mechanism,
Physical parameters: height 0.8m, width 0.3m, depth 0.25m, weight 70kg, max
speed 6km/h,
Components: mobile platform, torso, hands, head,
Scene detection: vision cameras, microphone array, laser range finder,
Robot-human communication: microphones, loudspeakers, localisation of voice
signal source, speech recognition and synthesis, dialoguing, face recognition, gestures,
Autonomy: point-to-point motion, path following, collision avoidance, navigation
in populated places,
Openness of design: none,
Availability, price: unavailable.
2.2.8. MDS
General description
This robot has been designed at MIT Media Laboratory for the purpose of supporting
research and education in the area of human-robot interaction and social learning. The
robot distinguishes itself by technological refinement, high manufacturing precision and
aesthetic appearance. MDS (Mobile, Dextrous, Social) moves on a dynamically balancing platform, has dextrous hands, and a head with expressive face. The design is being
developed toward incorporating mobility, manipulability, advanced communication, and
interaction capabilities with humans. The robot is presented in figure 2.21.
Specific features
•
•
Manufacturer: MIT, US,
Web: http://robotic.media.mit.edu/projects/robots/mds/overview/
overview.html,
27
Chapter 2. Overview of social robot designs
2.22. Monty
•
•
•
•
•
•
•
•
•
•
•
Purpose: research platform,
Appearance: humanoid,
Mobility: balancing mobile platform,
Physical parameters: as 3 year old child,
Components: mobile platform, torso, hands, neck, head,
Scene detection: visual, auditory and tactile systems, laser range finder, capacitive
sensing of humans,
Robot-human communication: gestures, posture, facial expressions, speech synthesis,
Autonomy: point-to-point motion, path following, collision avoidance, navigation,
Control system: onboard DSPs and FPGAs, external PCs running Linux,
Openness of design: none,
Availability, price: unavailable.
2.2.9. MONTY
General description
The robot Monty belongs to a family of humanoid robots recently launched by Anybots,
Inc. Monty is intended to be a robotic servant, able to wash dishes or serve a cup of
coffee. The robot is shown in figure 2.22.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
28
Manufacturer: Anybots, Inc., Mountain View, CA, US,
Web: http://anybots.com/,
Purpose: robotic servant,
Appearance: humanoid,
Mobility: 2 wheel, balancing,
Physical parameters: height 1.70m, weight 72kg,
Components: mobile base, torso, 2 arms, head,
Scene detection: no data available,
Robot-human communication: no data available,
Autonomy: balancing,
Control system: remote control,
Availability, price: unavailable.
2.3. Android robots
2.23. BARTHOC
2.3. ANDROID ROBOTS
2.3.1. BARTHOC
General description
BARTHOC (Bielefeld Anthropomorphic RoboT for Human-Oriented Communication)
serves the purpose of the investigation of social and emotional factors in the human-robot
interactions. The robot is able to conversate, carry out requests, and react to gestures.
Its designers put special attention to the development of robot’s communication capabilities. As a result, BARTHOC can not only receive verbal messages, but also estimate
the interlocutor’s emotions. The robot is presented in figure 2.23.
Specific features
•
•
•
•
•
•
•
•
•
Manufacturer: University of Bielefeld, Germany,
Web: http://aiweb.techfak.uni-bielefeld.de/node/393,
References: M. Hackel, et al., Designing a sociable humanoid robot for interdisciplinary research, Adv. Robotics, 20(11), pp. 1219-1235, 2006,
Purpose: research platform,
Appearance: android,
Mobility: none,
Physical parameters: human size,
Components: torso, hands, neck, head with face,
Scene detection: microphone, vision cameras,
29
Chapter 2. Overview of social robot designs
2.24. ROMAN
•
•
•
•
Robot-human communication: microphone, loudspeakers, facial expressions,
speech recognition and synthesis, dialoguing, gesture recognition, people detection,
selection of the communication partner, carrying out voice requests,
Autonomy: object tracking, reaction to object in the environment,
Openness of design: none,
Availability, price: unavailable.
2.3.2. ROMAN
General description
ROMAN (RObot huMan interAction machiNe) has been designed at the University of
Kaiserslautern as a test platform for the examination of human-robot interactions. The
robot has the form of a very realistic head mounted on a neck. The emphasis has
been put on the communication interfaces that incorporate speech, gestures, and facial
expressions. The robot emotional structure includes several drives. A collection of 6 basic
emotions has been implemented. The robot is able to interpret the movements of its
interlocutor and adjust accordingly its behaviour. The robot is presented in figure 2.24.
Specific features
•
•
•
•
•
•
•
•
•
•
•
30
Manufacturer: University of Kaiserslautern, Germany,
Web: http://agrosy.informatik.uni_kl.de/en/robots/roman,
References: K. Mianowski, et al., Mechatronics of the humanoid robot ROMAN,
Proc. 6th Int. Workshop on Robot Motion and Control (RoMoCo), Bukowy Dworek,
PL, 2007,
Purpose: research platform,
Appearance: android,
Mobility: none,
Physical parameters: human’s head,
Components: neck, head with realistic face,
Scene detection: stereovision cameras, microphones,
Robot-human communication: environment detection, head motions, facial expressions (raising eyebrows, stretching lips, wrinkling nose),
Autonomy: reaction to object in the environment,
2.3. Android robots
2.25. Albert
•
•
•
Control system: DSPs, embedded PC,
Openness of design: none,
Availability, price: unavailable.
2.3.3. ALBERT
General description
Similarly to Kismet and MERTZ, this is a robotic head equipped with a very realistic
human face made of a special elastic polymer Frubber (foam-rubber) deceptively similar
to the human skin. The purpose of Albert is getting involved into interaction and
conversation with humans. The robot is able to understand statements in English,
initiate an eye contact, recognise human faces, and attract the interlocutor’s attention.
Albert can be mounted as a head on the biped Hubo of the Honda Asimo class. The
robot is presented in figure 2.25.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Hanson Robotics, US,
Web: http://www.hansonrobotics.com,
Purpose: social robot,
Appearance: android,
Mobility: none,
Physical parameters: human head,
Components: head with face,
Scene detection: microphone, micro-cameras,
Robot-human communication: microphone, loudspeakers, facial expressions, eye
contact tracking and maintaining, speech recognition and synthesis, natural language
processing, face recognition, dialoguing, personality simulation,
Autonomy: object tracking, reaction to object in the environment, eye contact
tracking,
Openness of design: none,
Availability, price: available for hire, USD 2000/day.
31
Chapter 2. Overview of social robot designs
2.26. iCat
2.4. ZOOMORPHIC ROBOTS
2.4.1. ICAT
General description
The robot iCat was developed in Philips laboratories as a hardware and software platform
for fostering the research on human-robot interactions. The robot was equipped with
movable eyes, eyelids, eyebrows, mouth, and head enabling the expression of emotions.
A camera served for face recognition. Tactile sensors and LEDs were installed in iCats
ears and feet. The robot is presented in figure 2.26.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Philips, Holland,
Web: http://www.research.philips.com/technologies/syst_softw/robotics/
index.html,
References: A. J. N. van Breemen, iCat: Experimenting with animabotics, AISB
2005 Creative Robotics Symp., Hatfield, England,
Purpose: research robot,
Appearance: zoomorphic,
Mobility: none,
Physical parameters: height 0.38m,
Components: feet, torso, head,
Scene detection: tactile sensors, microphone, infrared sensors,
Robot-human communication: loudspeaker, facial expressions, determination of
the sound source,
Autonomy: expression of emotions, face recognition,
Openness of design: limited,
Availability, price: no longer available.
2.4.2. PLEO
General description
The Pleo robot is an autonomous, interactive robot toy in a shape of a dinosaur baby,
commercially available from Ugobe Inc. It has 14 DOF and it can walk, has movable head
and tail. The robot is equipped with over 40 sensors (touch, motion, sound, light) and it
is programmed to play with a person, express simple emotions, feel hunger and fatigue,
change its mind and mood. With its colour camera the robot may see colours, detect
32
2.4. Zoomorphic robots
2.27. Pleo
objects in front of it, detect motion, track a moving objects. The robot is presented in
figure 2.27.
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Ugobe, US,
Web: http://www.pleoworld.com/,
Purpose: entertainment,
Appearance: zoomorphic,
Mobility: toddling, quadruped,
Physical parameters: height 0.20m, length 0.53m, weight 1.5kg,
Components: torso, legs, head, tail,
Scene detection: tactile sensors, microphones, vision system,
Robot-human communication: sounds acquisition, roaring, motion, detection,
body expressions,
Autonomy: toddling, expression of emotions, playing,
Control system: 32 bit Atmel ARM7 microprocessor board running LifeOS, (with
an artificial intelligence engine) powered by motor and vision subprocessors
Openness of design: limited (programmable, with the system documentation,
delivered)
Availability, price: on sale, EUR 300.
2.4.3. LEONARDO
General description
This robot, whose name is a tribute paid to Leonardo da Vinci, a scientist, inventor
and an artist, has been called a Stradivarius of expressive robots. Leonardo serves
primarily as a research platform for the implementation and the testing of human-robot
interaction algorithms, as well as of scene analysis and understanding. The robot, that
looks as a lovely furry animal toy, has been equipped with sophisticated tools for speech
analysis and visual image recognition, is able to understand people and learn simple
behaviours. Using its paws Leonardo can press buttons or shift small objects. Its main
communication means is the head that, thanks to movable eyes, mouth and neck, is able
to express a number of emotions. The robot is presented in figure 2.28.
33
Chapter 2. Overview of social robot designs
2.28. Leonardo
Specific features
•
•
•
•
•
•
•
•
•
•
•
•
•
34
Manufacturer: MIT, US,
Web: http://robotic.media.mit.edu/projects/robots/leonardo/overview/
overview.html,
References: J. Gray, et al., Leonardo: Goal assistance with divergent beliefs, http:
//videolectures.net/aaai07graylga/,
Purpose: research platform, robotic companion,
Appearance: zoomorphic,
Mobility: none,
Physical parameters: height 0.50m,
Components: torso, paws, neck, large-eared head,
Scene detection: visual and auditory systems,
Robot-human communication: facial and body expressions, sharing attention,
speech, face recognition,
Autonomy: expression of emotions, face recognition,
Openness of design: limited,
Availability, price: unavailable.
Chapter 3
Overview of social robot components
3.1. MOBILE BASES
Due to their unique appearance and very natural mobility properties we shall concentrate
on balancing wheeled mobile platforms, based on the concept of the inverted pendulum.
In recent years, this type of mobility has found many applications, e.g., in the design of
walking gaits for humanoid robots, robotic wheelchairs, and personal transport systems.
The movements of a balancing platform look pleasant, and in some aspects resemble
human movements. Natural balancing actions show that this construction is not only
acceptable for people, especially for children, but also safe.
3.1.1. BALANCING PLATFORMS OVERVIEW
The uniqueness of balancing systems has aroused a vivid interest within the robotic
community. Below we describe a collection of mobile bases belonging to the balancing
family.
Segway RMP 100
Segway RMP (Robotic Mobility Platform) 100 is a mobile platform particularly suitable
for indoor applications where maneuverability and a tight turning radius are advantageous. A dynamic stabilisation system helps the Segway RMP 100 to climb over
obstacles. The vehicle is shown in figure 3.1.
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Segway Inc., Bedford, NH, US,
Web: http://www.segway.com/police-government/products-solutions/
robotic-mobility-platform.php,
Purpose: robotic mobility platform, transporter,
Mobility: 2 wheel, balancing,
Physical parameters: height 0.69m, width 0.61m, depth 0.61m, weight 57kg, speed
1-10km/h, range 13-19km, carrying capacity 45-68 kg,
Components: wheeled mobile base, batteries, small upper table,
Scene detection: none,
Autonomy: balancing,
Control system: dynamic stabilisation,
Openness of design: none,
Availability, price: USD 17000.
nBot
The mobile platform nBot was constructed by P. Anderson form the Southern Methodist
University, Texas, US, and featured as NASA’s Cool Robot of the Week for 19 May 2003.
The platform is shown in figure 3.2.
•
•
•
•
Manufacturer: Southern Methodist University, University Park, TX, US,
Web: http://www.geology.smu.edu/~dpa-www/robo/nbot/,
Purpose: for outdoor performance,
Mobility: 2 wheel, balancing,
Chapter 3. Overview of social robot components
3.1. Segway RMP 100
3.2. nBot
•
•
•
•
•
•
•
•
Physical parameters: height 0.25m,
Components: aluminium chassis, DC motors with gear reduction, standard model
airplane wheels gyroscopes and tilt sensors,
Scene detection: none,
Autonomy: balancing,
Control system: Freescale HC11 microcontroller,
Controller: linear PID for rolling remote control, the balancing algorithm measures
two outputs from the robot and calculates the balancing torque,
Openness of design: partly open,
Availability, price: unavailable.
BallyBot
BallyBot is an experimental balancing robot, used as an experimental platform to study
sensor and control systems to be implemented for humanoid robots. The robot is shown
in figure 3.3.
•
•
•
•
•
36
Manufacturer: University of Western Australia,
Web: http://robotics.ee.uwa.edu.au/eyebot/,
Purpose: balancing research platform,
Mobility: 2 wheel, balancing,
Physical parameters: height 0.15m, weight <2kg,
3.1. Mobile bases
3.3. BallyBot
3.4. SegBot
•
•
•
•
•
•
•
Components: plexiglass chassis, DC motors with gear reduction, gyroscopes and
tilt sensors, encoders,
Scene detection: none,
Autonomy: balancing,
Control system: EyeBot controller with Freescale MC68332,
Controller: linear controllers LQR and pole placement for balancing, PID for rolling
remote control,
Openness of design: open,
Availability, price: unavailable.
SegBot
The SegBot appeared in 2004 as the final project for the Introduction to Mechatronics
class at the University of Illinois. The goal was to design and build a 2 wheel balancing
robot based on the same principles as Segway. The robot is shown in figure 3.4.
•
•
•
•
•
•
Manufacturer: University of Illinois, Urbana Champaign, US,
Web: http://coecsl.ece.uiuc.edu/ge423/spring04/group9,
Purpose: balancing research platform,
Mobility: 2 wheel, balancing,
Physical parameters: height 0.30m,
Components: aluminium chassis, DC motors, gyroscopes and tilt sensors, infrared
sensors, encoders, camera,
37
Chapter 3. Overview of social robot components
3.5. Joey
•
•
•
•
•
•
Scene detection: none,
Autonomy: balancing,
Control system: Spectrum Digital TMS320C6713 DSK,
Controller: linear controller LQR for balancing, wall following or remote control,
Openness of design: partly open,
Availability, price: unavailable.
Joey
This mobile platform has been devised at the Swiss Federal Institute of Technology,
Lausanne, Switzerland. The vehicle is able to make stationary U-turns. A controller
keeps the system in equilibrium. The robot is shown in figure 3.5.
•
•
•
•
•
•
•
•
•
•
•
•
Manufacturer: Swiss Federal Institute of Technology, Lausanne, Switzerland,
Web: http://leiwww.epfl.ch/joe,
Purpose: balancing small and lightweight research platform,
Mobility: 2 wheel, balancing,
Physical parameters: height 0.65m, weight 12kg, max speed 1.5m/s,
Components: chassis, DC motors, gyroscopes and tilt sensors,
Scene detection: none,
Autonomy: balancing,
Control system: DSPs,
Controller: remote control, linear pole placement for balancing,
Openness of design: partiy open,
Availability, price: unavailable.
uBot4
uBot4 is a wheeled, dynamically balancing, bimanual mobile manipulator, combining
manipulation and mobility capabilities in a very efficient way. It has been designed as a
research platform dedicated to the study of robotic tasks like pushing, pulling, digging,
grasping, single robot and cooperative transport, and also traversing a rough terrain. A
more advanced version, uBot5, has been described in chapter 2. The robot is shown in
figure 3.6.
•
•
•
•
38
Manufacturer: University of Massachusetts, Amherst, Massachusetts, US,
Web: http://www-robotics.cs.umass.edu/Robots/UBot,
Purpose: balancing small and lightweight research platform for mobile manipulation,
Mobility: 2 wheel, balancing,
3.2. Arms and hands
3.6. uBot4
•
•
•
•
•
•
•
•
Physical parameters: height 0.54m, width 0.48m, weight 11.5kg, max speed
4km/h,
Components: chassis, LCD touchscreen with speakers and a USB webcam, two
arms 0.5m length, gyroscopes and tilt sensors,
Scene detection: none,
Autonomy: balancing,
Control system: DSPs, embedded Bitsy, PC/104+ based Pentium,
Controller: linear PID for remote control, LQR and pole placement for balancing,
Openness of design: partly open (with an individual password),
Availability, price: unavailable,
3.1.2. CONCLUDING REMARKS
A comparison of robotic designs studied in chapter 2, and an analysis accomplished
in this subsection prove that balancing mobility, maneuverability, capability of climbing
over obstacles, and a sort of charm in motion demonstrated by balancing wheeled mobile
platforms make them preferred candidates for the mobile base of the LIREC robotic
companion. However, almost all existing constructions are research prototypes that are
not commercially available. The only available devices, coming from Segway Inc., are
too heavy, lacking the open architecture, and much too expensive. This being so, it
seems very rational to build a balancing platform dedicated to LIREC objectives.
3.2. ARMS AND HANDS
One of inseparable human skills is the ability to manipulate objects and communicate
by gestures – simply using his or her arms and hands. A social robot acting in human
surrounding should have a similar competence.
As we have mentioned above, there are two different tasks for robotic arms and hands:
manipulation and gesticulation. Each of these tasks presents different requirements with
respect to arm design, its equipment, and motoric skills. Manipulation requires a solid,
rigid construction, a precise positioning system with force control, a vision feedback,
and touch sensors. A device with such features will be quite heavy, complicated, and
likely very expensive. Gesticulation is not so challenging and usually requires a plain,
light arms structure, and a simple positioning system with an overcurrent protection
only. Possibly, to prevent from collision with other objects, the support from a vision
system will be required. Usually, an arm with a hand designed for gesticulation only
39
Chapter 3. Overview of social robot components
3.7. David Ng
3.8. AMTEC
is very light, its construction and control system are not complicated, it has low power
consumption and is cheaper than the arm designed for manipulation.
All the scenarios elaborated within the LIREC project are based on the following assumptions:
•
•
•
robot communicates with humans through voice, gesture and facial expressions,
robot undertakes only very simple manipulation tasks,
robot does not get into physical interaction with a human.
For these reasons, we are looking for robotic arms/hands designed primarily for gesticulation.
3.2.1. ARMS
David Ng – Robot Hand/Arm
Servo operated robotic hand and arm.
•
•
•
•
Simple construction, light weight,
Hand/arm is operated by 13 high torque servos,
Price: USD 6000,
Web: http://www.androidworld.com/prod61.htm.
SCHUNK AMTEC Robotics
The human-like 7 DOF light-weight arm (Amtec PowerCube technology).
•
•
•
•
40
Modular design of mechatronic modules,
Carbon Fiber Reinforced Polymer,
Operate off a 24VDC battery,
Control is done directly via PC or notebook, a PCI or USB interface,
3.2. Arms and hands
3.9. Neuronics
3.10. Festo
•
•
Physical parameters: length 1.37m, weight 16kg,
Web: http://www.amtec-robotics.com/robotersysteme_en.html.
Neuronics AG
Katana – Intelligent Personal Robot (IPR).
•
•
•
•
•
•
Powerful embedded controlboard (TI TMS320 32bit),
Linux with Xenomai Hard Real Time extensions,
Degrees of freedom: 5 to 6 (6 Motors),
Weight: 4.8 kg,
Price: CHF 19000,
Web: http://www.neuronics.ch.
Festo
Airic’s arm is inspired by nature. Based on a combination of mechatronics and the
biological model of a human being, it opens up new possibilities for future automated
movement processes.
•
•
•
•
•
30 muscles,
The muscles are Festo products,
Pneumatic,
Weight: 6.3kg,
Web: http://www.festo.com/inetdomino/coorp_sites/en/
ffeed49f2394ea43c12572b9006f7032.htm.
41
Chapter 3. Overview of social robot components
3.11. Barrett
3.12. Shadow
Barrett Technology
WAM Arm s a highly dexterous, naturally backdrivable manipulator. The only arm sold
in the world with direct-drive capability supported by Transparent Dynamics between
the motors and joints.
•
•
•
•
•
•
•
Human-like kinematics,
4 or 7 DOF,
Modular construction,
Servo-electronics,
Lightweight design (3.3 kg or 5.8 kg),
Open control architecture,
Web: http://www.barrett.com/robot/products-arm.htm.
3.2.2. HANDS
Shadow Dextrous Hand
The Shadow Dextrous Hand is an advanced humanoid robot hand system that reproduces
as closely as possible the 24 degrees-of-freedom of the human hand.
•
•
•
•
•
•
•
42
Pneumatic,
Hall effect sensor with typical resolution of 0.2 degrees senses the rotation of each
joint.
software provided under GNU GPL
Physical parameters: size: typical human male, weight 3.9kg
Communication: CAN,
Price: EUR 90000 ,
Web: http://www.shadowrobot.com/.
3.2. Arms and hands
3.13. UBHand 3
3.14. Faulhaber Hand
UBHand 3
The overall hand is based on an innovative mechanical architecture, which adopts deformable elements as joint hinges (compliant mechanisms).
•
•
•
•
Endo-skeletal structure,
Any kind of artificial muscles,
16 DOF (max. 20)
Web: http://www-lar.deis.unibo.it/activities/dexterousmanipulation/
UBHandIII/index.html.
Faulhaber Hand
This robotic hand consists of three fingers, each with four joints in three degrees of
freedom. The fourth finger, designed as a thumb, has four degrees of freedom.
•
•
•
•
•
•
Separately controllable fingers,
5 Mbps high speed bus is incorporated in the robotic hand itself and developed
specifically for this application,
Each finger joint features a company-designed contactless angle sensor as well as a
torque sensor,
Control system, a signal processor on a plug-in PCI card, is integrated in a standard
PC,
Electronically commutated DC motors (EC motors),
Web: http://www.faulhaber-group.com/n273481/n.html.
Dainichi Company, Ltd. Kani, Japan
The man-type robot hand, Gifu Hand III, has been developed as an object for a research
of grasping and manipulating an object and operating a machine device, and it is small
and lightweight.
•
5 fingers, 20 joints, and 16 DOF,
43
Chapter 3. Overview of social robot components
3.15. Gifu Hand III
3.16. BH8-series BarrettHand
•
•
•
•
Small servomotors inside the fingers,
Weight: 1.4kg,
Price: USD 51400,
Web: http://www.kk-dainichi.co.jp/e/gifuhand.html.
Barrett Technology
The BH8-series BarrettHand is a multi-fingered programmable grasper with the dexterity
to secure target objects of different sizes, shapes, and orientations. and compact form,
it is totally self-contained.
•
•
•
•
•
•
•
Three multi-jointed fingers,
Human-scaled,
Brushless, DC, servo motors,
Industry-standard serial communications,
C-Language command library for PCs,
Weight: 1.18kg,
Web: http://www.barrett.com/robot/products-hand.htm.
Touch Bionics
The i-LIMB Hand is the world’s first fully articulating and commercially available bionic
hand.
•
•
•
•
•
•
44
Five individually powered fingers,
High-strength plastics,
Lightweight,
Control system uses two-input myoelectric (muscle) signal to open and close the
hand,
Price: USD 18000,
Web: http://www.touchbionics.com/.
3.3. Vision acquisition system
3.17. i-LIMB Hand
3.2.3. CONCLUDING REMARKS
Concluding, one can find only a few complete arm solutions in the market. As a rule, offered products are appropriate rather for manipulation than gesticulation task. Available
constructions are quite heavy, complicated and expensive. Truly dedicated to humanoid
robots are the arms developed by Festo and David Ng. Festo does not mention the
arm in its offer, it looks as only an advertisement product. Ready to buy is David Ng
arm, unfortunately the available information if far from complete. The arm construction
bases on aluminium tubes powered by high torque servos. We do not know either the
arm true weight, or control and sensor systems or interfaces. Moreover, on the short
films avaliable on the web page it easy to see, that arm’s performance leaves much to
be desired. Similar situation is in the case of robotic hands. Avaliable hands were
constructed mainly for manipulation tasks. From our point of view, worth noticing is
the UBHand III because of its lightweight and simple construction. Unfortunately, the
hand itself does not contain driving motors and is commercially unavailable.
3.3. VISION ACQUISITION SYSTEM
The vision system designed for the social robot perception should be able to perform
the following tasks:
•
•
•
•
objects (human) localization,
face detection and localization,
motion detection,
object tracking.
These tasks require, that the acquisition system should feature:
•
•
•
wide viewing angle (e.g. for motion detection and object localization and tracking),
reasonable resolution of the ROI presenting the human face (e.g. for the recognition
of emotions),
high frame rate (e.g. for motion estimation and object tracking).
The first two requirements can be fulfilled by the acquisition system having the ability
to change both the viewing angle (Zoom) and the viewing direction (Pan-Tilt). The
last requirement implies high image data transmission rate between the camera and the
computer and hence determines the selection of the camera interface.
3.3.1. COMMERCIAL VISION SYSTEMS
The commercially available vision systems are addressed to typical tasks (inspection,
surveillance, traffic supervision, factory automation, robot navigation, obstacle avoidance, etc.). The most popular groups are discussed in more details in this section.
45
Chapter 3. Overview of social robot components
General conclusion, however, is that we cannot use any of the offered systems in our
project. The main reason is that the commercial systems are not flexible enough to
be adopted to the specific tasks in social robotics. Moreover, they are not open, no
user modifications are allowed. Therfore we consider assembling the vision system using
available components (cameras, zoom lenses, pan/tilt units, frame grabbers etc.).
Surveillance and security systems
Many companies are offering commercial vision systems dedicated to security and surveillance. Some tasks (object localization and tracking, face recognition, etc.) are similar
to those of the social robot visual perception system. Unfortunately, those systems are
closed and cannot be easily adopted (expanded) to fit our needs:
http://www.sourcesecurity.com/companies/company-listing/flir-commercial
-vision-system-b-v.html,
http://www.sourcesecurity.com/product-filter/cctv/image-capture/video
-motion-detectors.1.html,
http://www.videortechnical.com/index_dt.php,
http://verint.com/video_solutions/index.cfm,
http://www.security.honeywell.com/uk/,
http://www.everfocus.de/en/index.htm,
http://www.gallaghersms.com/,
http://www.multipix.com/_security/index.php.
Industrial machine vision
Some examples, not applicable in the project:
http://www.ni.com/vision/vbai.htm – automated inspection,
http://www.ni.com/vision/vdm.htm – vision development (e.g. OCR),
http://www.ni.com/vision/cvs.htm – compact vision systems (embedded processors),
http://www.matrox.com/imaging/products/system.cfm,
http://www.jai.com/EN/Pages/home.aspx – traffic solutions,
http://www.machinevisiononline.org/.
Smart cameras
A smart camera is a machine vision system which contains: image capture circuitry,
embedded processor, which can extract information from images, and interface devices
used to make results available to other devices. It is a self-contained, standalone vision
system with built-in image sensor in the housing of an industrial video camera. It
contains all necessary communication interfaces, e.g. Ethernet, as well as industry-proof
24V I/O lines for connection to a PLC, actuators, relays or pneumatic valves. It is not
necessarily larger than an industrial or surveillance camera. In most cases, smart camera
functions are limited to early processing, segmentation (thresholding, gradient), pattern
matching, blob coloring (labeling), silhouette parameters calculation (moment invariants), etc. More information on selected smart cameras can be found in the following
web pages:
http://pennwell.365media.com/visionsystemsdesign/category/37.html,
http://www.vision-components.de/,
46
3.3. Vision acquisition system
3.18. Smart cameras: VC4438, NI-1742, DVT
http://www.cognex.com/ProductsServices/VisionSystems/InSight.aspx?id=110,
http://www.ni.com/vision/smartcamera.htm,
http://www.matrox.com/imaging/products/smart_cameras.cfm,
http://www.vision-components.de/,
http://www.sickivp.se/sickivp/en.html,
http://www.cimtecautomation.com/DVT_cameras_sensors.htm.
Example smart camera – VC Vision Components GmbH VC4438
(http://www.vision-components.de) is shown in figure 3.18:
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Sensor: 1/3”, 640 (H)×480 (V) Pixel
Shutter: High-speed: up to 5s, Low-speed: up to 8s adjustable integration time
Integration: Full Frame Progressive Scan
Frame rate: 63 fps (126 fps with 2 times binning)
Acquisition: Asynchronous, program controlled or external trigger, full frame
A/D conversion: 1×25 MHz / 10 Bit
Processor: 8000 MIPS, 1 GHz Texas Instruments TMS320C64xx
Image display: B&W or pseudocolor from 3x8 bit RGB lookup table
Image/Data memory: 64 MBytes SDRAM
Flash memory: 4 MBytes Flash EPROM (non-volatile memory) for programs and
data, programmable in the system
Dig. I/Os: 4 inputs / 4 outputs optically decoupled 24V, outputs 4×500mA
Interfaces: RS232 up to 115.200 Baud max. and 100Mbit Ethernet
Video output: SVGA 800x600 (VESA standard)
Supply voltage: 24V ±20
Electrical connections: I/O (DC IN, PLC, 12-pin), V24 (6-pin), Trig (Trigger/keypad, 6-pin), VGA Out (10-pin)
Dimensions: approx. 110×50×35 mm, approx. 400 g.
Stereo
Binocular stereovision systems are available, which can perform depth map building
based on the disparity of the stereo pair of images. MobileRobots offers the MobileRanger Stereovision System with embedded range-finding as an add-on to the PeopleBot (http://www.activrobots.com/ACCESSORIES/MobileRanger.html). The binocular and trinocular camera heads are shown in figures 3.19 and 3.20. More solutions can
be found in:
http://www.ptgrey.com/products/stereo.asp?gclid=CNvT-fOI4ZUCFRSb1Qodx3QSXA,
http://www.videredesign.com/vision/svs_intro.htm,
http://www.ai.sri.com/software/SVS,
http://www.newtonlabs.com/cognachrome/,
47
Chapter 3. Overview of social robot components
3.19. Stereo camera heads of MobileRanger
3.20. Other stereo camera head examples
http://www.videredesign.com/vision/stereo_manuals.htm,
http://www.dis.uniroma1.it/~iocchi/stereo/,
http://www.linuxdevices.com/news/NS8326899521.html,
http://linuxdevices.com/articles/AT3172103326.html,
http://www.newtonlabs.com/cognachrome/,
http://www.videredesign.com/vision/stereo_manuals.htm,
http://www.dis.uniroma1.it/~iocchi/stereo/.
3.3.2. PAN-TILT-ZOOM
A way to achieve the required performance of the image acquisition system is the PTZ
(Pan-Tilt-Zoom) capability. We will consider two PTZ architectures: soft PTZ and
mechanical PTZ.
Soft PTZ
One of the possible image acquisition architectures (a single camera) requires:
•
•
•
•
48
high quality wide angle lens,
variable focus capability,
high speed camera interface,
high resolution image sensor with:
◦ global shutter,
◦ programmable ROI,
◦ binning capability.
3.3. Vision acquisition system
3.21. Logitech QuickCam Pro 9000 camera
3.22. AXIS 212 surveillance camera
In this way we can realize a soft PTZ functionality of the image acquisition system. The
limited resolution of the image sensors is the main constraint of such a solution.
An example of a camera near to the above specification is Logitech QuickCam Pro 9000
(shown in figure 3.21). It features:
•
•
•
•
•
•
•
•
•
Auto-focus,
Auto-follow and zoom on both one and multiple people,
RightLight 2 auto lighting compensation,
Carl Zeiss lens,
True 2 megapixel sensor (1600x1200), with software enhancement up to 8,
True high-definition resolution 960×720,
Integrated microphone,
USB 2.0 interface,
Web: http://www.logitech.com/index.cfm/38/3056&cl=us,en.
Linux driver for QuickCam USB cameras (qc-usb) works well with v4l2.
Another example of such a solution, AXIS 212 PTZ Network Camera (http://www.
axis.com/products/cam_212/) is shown in figure 3.22. It achieves full overview as well
as instant pan/tilt/zoom without any moving parts. This is based on two factors:
•
•
a wide-angle lens combined with a 3 megapixel sensor,
utilization of the full “windowing” possibilities: the camera captures predetermined
sections of the overview without mechanical motion.
The advantages of no moving parts are considerable:
•
•
no noise,
no delay for mechanical movement.
Mechanical PTZ
A PTZ camera with a motorized lens and a mechanical PTU (Pan-Tilt Unit) seems to
be the most effective alternative. The tasks of motion detection and object localization
49
Chapter 3. Overview of social robot components
3.23. Canon VC-C4 PTZ camera
3.24. Canon VC-C50i PTZ camera
can be performed in wide lens mode. Human face recognition as well as more precise
object localization and recognition will use tele lens mode. Thus, the required image
sensor resolution remains within reasonable limits (e.g. VGA or CCIR-D1). The main
drawback of this solution is its mechanical complexity. Moreover, in our application,
the level of mechanical noise could be unacceptably high for sound (voice) acquisition
system of the robot.
There are many PTZ camera modules available, especially for security/surveillance purposes. Most of them are analog video cameras (PAL/NTSC color system):
Canon VC-C4 Camera for Video Conferencing (formerly offered with some versions of
PeopleBot) is shown in figure 3.23.
•
•
•
•
•
•
•
•
•
•
•
•
•
Genuine Canon 16x Optical Lens, the highest zoom ratio in its class
1/4” CCD with 410,000 pixels
Auto and manual focus
Auto and manual exposure
Backward compatibility with previous Canon models
Motorized Pan range of 100◦ each way (VC-C4R 170◦ each way)
Motorized Tilt range of 90◦ up, 30◦ down (VC-C4R 10◦ up, 90◦ down)
Intelligent image processing to achieve higher image compression for video streaming
9 cameras can be cascaded together and operated by one I/R controller or via RS
232 (serial) computer control
Infra-red remote control included
Up to 9 preset motion positions for accurate framing
Compact size, only 4” W×4.48” D×3.58” H
Weight: <0.5kg.
Canon VC-C50i Communication Camera (http://www.usa.canon.com/consumer/) is
shown in figure 3.24.
•
50
Total Number of Pixels [NTSC] 630K total (340K effective); [PAL] 740K total (400K
effective)
3.3. Vision acquisition system
3.25. SONY EVI-HD1 high definition color PTZ video camera
•
•
•
•
•
•
•
•
•
Minimum Illumination 1 lux (visible light mode, at 1.30 electric shutter speed)
Focusing Wide: 0.01 to inf./tele. 1.6 to inf., AutoManual
Zoom 26 times (at infinite) optical, 12x digital
Pan Angle Range [VC-C50i] ± 100◦ [VC-C50iR] -90◦ to +10◦
Tilt Angle Range [VC-C50i] -30◦ to +90◦ , [VC-C50iR] -90◦ to +10◦
Moving Speed 1-90◦ /s
White Balance: auto/manual/one touch
Dimensions [VC-C50i] 100×125×96mm; [VC-C50iR] 130×125×98mm
Weight [VC-C50i] 420g; [VC-C50iR] 490g
Many of the PTZ cameras provide both Video and HD/SD image output formats. SONY
EVI-HD1 High Definition Color PTZ Video Camera (shown in figure 3.25) is one of the
examples:
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
Web: http://www.sony.pl/biz/view/ShowProduct.action?product=EVI-HD1
Image sensor 1/3-type CMOS
Effective pixels Approx. 2 Megapixels (16:9)
Lens 10x optical Zoom, 40x with digital zoom, f = 3.4 to 33.9 mm, F 1.8 to F 2.1
Minimum object distance 100 mm (wide)
Horizontal viewing angle 8◦ (tele) to 70◦ (wide) at HD signal output
Focus system Auto / Manual
Exposure control Auto / Manual / Priority AE / Exposure compensation / Bright
Shutter speed 1/2 to 1/10,000s
White balance Auto / Indoor / Outdoor / One push auto / Manual
Pan ±100◦ (Max. speed 300◦ /s) Tilt ±25◦ (Max. speed 125◦ /s)
Video output HD HD-SDI, Analog Component (Y/Pb/Pr) SD VBS, Y/C
Power requirements 12 V DC (10.8 to 13.0 V DC)
Power consumption Max. 30 W (at DC 12 V)
Dimensions 259 (W)×150 (H)×169 (D) mm
Mass Approx. 2 kg (4 lb 7 oz)
PTZ cameras are frequently equipped with IP networking interface (examples follow).
Sony SNC-P5 Colour PTZ IP Camera,1/4-type progressive scan CCD (http://pro.
sony.com/bbsc/ssr/app-security/cat-ip/product-SNCP5/) is shown in figure 3.26:
•
•
•
•
•
•
Interfaces/Ports RJ-45 10/100Base-TX, Analog Video Out
Protocol TCP/IP, ARP, ICMP, DHCP, DNS, HTTP, FTP, SMTP, NTP, SNMP,
RTP, RTCP, PPPoE
Optical Zoom 3x
Dimensions 130mm Height×130mm Width×110mm Depth
Connectivity Technology Cable
Video Resolution 160×120 to 640×480, MPEG-4 and Motion JPEG
51
Chapter 3. Overview of social robot components
3.26. Sony SNC-P5 colour PTZ IP camera
3.27. Canon VB-C50i networking communication camera
•
•
•
•
Sensor Type CCD
Lens Type Optical 3x zoom lens
File Formats MPEG-4 Motion JPEG
Power Consumption 6.2W
Canon VB-C50i Networking Communication Camera, replacement for the VC-C4 (http:
//www.canon-europe.com/For_Home/Product_Finder/Web_Cameras/Web_Cameras/vc_
c50i/i) is shown in figure 3.27:
•
•
•
•
•
•
•
Total Number of Pixels 1/4” CCD 340,000 Effective Pixels
Minimum Illumination 1 Lux (1/30 shutter speed)
Zoom 26X Optical Zoom Lens/12X Digital Zoom
Pan Angle Range ±100◦
Tilt Angle Range +90◦ /-30◦
Dimensions VB-C50i: 104.6×130.3×108.7mm VB-C50iR: 104.6×130.3×110.6 mm
Weight VB-C50i: Approx. 1.4 lbs (620g) VB-C50iR: 1.5lbs. (700g)
Canon VB-C300 (shown in figure 3.27) offers both Video and Networking capabilities (http://www.canon-europe.com/For_Home/Product_Finder/Web_Cameras/Web_
Cameras/VB_C300/):
•
•
•
•
•
•
•
•
52
Video Signal: Conforms to PAL (output image when connected to VB-C300)
Image Sensor: 1/4 inch CCD
Total number of pixels: 310,000 effective
Lens: f = 3.0 - 7.2mm F 2.0 - 3.4, 2.4x power zoom (shooting at infinity)
Focusing: Auto/Manual Wide-angle: 0.01m to inf. Telephoto: 1.6m to inf
Horizontal field of view: 70.8 (W)×29.6 (T)
Pan Angle Range 340◦ (±170◦ )
Tilt Angle Range Upright Position = 115◦ (-25◦ to 90◦ ) Inverted = 115◦ (-90◦ to
25◦ )
3.3. Vision acquisition system
3.28. Canon VB-C300 PTZ camera
3.29. A family of motozoom lenses by ABAKUS
•
•
•
•
•
•
•
•
•
•
•
•
Moving Speed Pan = Max. 90◦ /s Tilt = Max. 70◦ /s
Minimum subject illumination: Day mode: 1 lux (at 1/30s) Night Mode (0.25 lux at
1/30s shutter speed)
Imaging Mode: Dayl/Night mode (manual or schedule-based switch)
Infra-red cut filter: Power-driven insertion/removal (normal mode/night mode)
Protocol: HTTP, DHCP, WW-HTTP (Canon)
Video compression method: Motion-JPEG (for video), JPEG (for still pictures)
Image sizes: 680x480 / 320x240 / 160x120
Max. capture frame rate: 25 fps
Frame Rate: 0.1 fps 25 fps (max)
Camera Dimensions: 132 (W)×122 (D)×130 (H) mm (excluding attachment)
Weight: Approx 780g (Excluding Dome)
Power consumption: 10W Max (on PoE) or 13W Max (on AC)
More examples:
•
•
•
Axis Communications - AXIS 212, 214, 232D, 2130R PTZ Network Cameras (http:
//www.axis.com/products/)
Speco Indoor Outdoor Color Pan Tilt Zoom Speed PTZ Camera (http://www.
specotech.com/)
Other PTZ IP Cameras from Axis, Canon, Sony, Vivotek, Panasonic, JVC, Nextiva and Pixord (http://www.kintronics.com/neteye/PTZ%20cameras.htm,http:
//www.videobotics.com/frtable2.html,http://www.aegis-elec.com/products/
ptz-pan-tilt-zoom-cameras.html)
Zoom lenses
The zoom lenses are relatively expensive (if not integrated into the camera). Moreover,
they require special control devices. An example of C-mount motorized zoom lenses
family (http://www.abakus.co.uk/C_Mounts.htm) is shown in figure 3.29.
53
Chapter 3. Overview of social robot components
3.30. ABAKUS Regular Premium Zoom lens
An example of motorized zoom lens which fits our needs is Regular Premium Zoom
(shown in figure 3.30). Its main parameters:
•
•
•
•
•
•
•
•
•
•
•
•
•
Angular zoom range 8:1
Angle vertical 33 –4
Angle horizontal 44 –5.4
Angle diagonal 55 –7
Maximum Metal Dia. 51mm
Front to Flange distance 110mm
Image Size: Height×Width×Diagonal 3.6×4.8×6.0(mm)
Relative aperture range f/2.8-f/16 & fully close
Closest focus 1000 mm
Filter Standard 49mm, 0.75 pitch Standard 49mm, 0.75 pitch
Weight 340 g
Motor Control Focus, Zoom & Iris
Price £ 3840
One can avoid using zoom lenses by using two separate cameras: one with a wide-angle
lens for panoramic view and the other, with PTU and a narrow-angle lens, for close-ups.
Such a system for face detection/tracking, composed of a pair of PTZ and fixed (wide-angle) cameras is described in Takuma Funahasahi, Masafumi Tominaga, Takayuki Fujiwara, Hiroyasu Koshimizu, ”Hierarchical Face Tracking by Using PTZ camera”, Proc.
of the 6th IEEE Conf. on Automatic Face and Gesture Recognition (FGR 04), 2004.
http://doi.ieeecomputersociety.org/10.1109/AFGR.2004.1301570
http://www.springerlink.com/content/n7keng8mj6xdkku0/.
Another possible solution is to use a small camera with integrated zoom lens, mounted
in robot eye/head. The PTU functionality can be then replaced by DOFs of the head
itself.
An example of a camera, which can be used in such an acquisition system is Sony
FCB-EX11DP Ultra-Compact Color PAL Block Camera (shown in figure 3.31).
•
•
•
•
•
•
•
•
•
•
•
54
Image Sensor 1/4-type EXview HAD CCD
Signal System PAL
Number of Total Pixels 440,000 pixels
Horizontal Resolution On/Off (530 TV Line Resolution Mode On)
Lens 10x optical zoom, F1.8 to F2.9 f = 4.2mm to 42.0 mm
Digital Zoom 12x (120x with optical zoom)
Viewing Angle (H) 46.0 (wide end) to 4.6 (tele end)
Minimum object distance 0.2m (wide end) to 1.0m (tele end)
Sync system Internal/External (V-Lock)
Minimum illumination 1.0 lx (typical) (F1.8, 50 IRE)
Electronic shutter More than 50 dB
3.3. Vision acquisition system
3.31. Examples of Sony block cameras (FCB-EX11DP and FCB-H11)
3.32. DFK 31BF03-Z Firewire camera with motozoom
•
•
•
•
•
•
•
White balance Auto, ATW, Indoor, Outdoor, One-push, Manual
Focusing System Auto (Sensitivity: normal, low), One-push AF, Manual, Infinity,
Interval AF, Zoom Trigger AF
Video Output VBS: 1.0 Vp-p (sync negative), Y/C
Camera control interface: VISCA (TTL signal level), baud rate: 9.6 Kb/s, 19.2 Kb/s,
38.4 Kb/s, stop bit selectable
Power Consumption 6 to 12 V DC/1.9W (motors inactive), 2.8 W (motors active)
Weight 95 g
Dimensions 35.9×40.8×59.2 (mm)
Sony block cameras can be found at http://pro.sony.com/bbsc/ssr/
cat-industrialcameras/cat-block/.
Several Zoom camera examples with Firewire interface are offered by TheImagingSource (http://www.theimagingsource.com/en/products/zoom_cameras/firewire_
color/).
As an example we can quote the DFK 31BF03-Z specification (the camera shown in
figure 3.32):
•
•
•
•
•
•
•
•
•
•
•
•
Video formats @ Frame rate 1024×768 UYVY @ 15, 7.5, 3.75 fps 1024×768 BY8 @
30, 15, 7.5, 3.75 fps
Sensitivity 0.5 lx at 1/15s, gain 20 dB
Dynamic range ADC: 10 bit, output: 8 bit
SNR ADC: 9 bit at 25C, gain 0dB
Sensor ICX204AK
Type progressive scan
Format 1/3 ”
Resolution H: 1024, V: 768
Pixel size H: 4.65 m, V: 4.65 m
Focal Length (wide) 4.8 mm
F-Stop (wide) 1.6
MOD (wide) 1 cm
55
Chapter 3. Overview of social robot components
3.33. PTU-D46-17 Pan-Tilt Unit
•
•
•
•
•
•
•
•
•
•
•
•
•
Focal Length (tele) 46 mm
F-Stop (tele) 3.7
MOD (tele) 40 cm
Focus auto/man
Iris auto/man
Shutter 1/10000 to 30 s
Gain 0 to 36 dB
Offset 0 to 511
White balance -2 dB to +6 dB
Supply voltage 8 to 30 VDC
Current consumption approx 200 mA at 12 VDC
Dimensions H: 50.6 mm, W: 50.6 mm, L: 130 mm
Mass 380 g
PTU – Pan-Tilt units
A Zoom Camera requires a PTU in order to be aimed at the human face. There are
different sources of such positioning devices, offering rather large units:
http://www.omnitech.com/srs_ugv_ptu.htm
http://stason.org/TULARC/science-engineering/robotics/116-Robotics-Pan
-Tilt-Mechanisms-part1.html
http://www.axsys.com/p-pantilt.php
Directed Perception provides computer controlled pan/tilt tracking mounts for robotics,
pan/tilt web cams, computer vision, teleconferencing, security, automation, etc. The
PTU-D46 models seem to fit our needs.
http://www.dperception.com/products_family_ptu-d46.html
The PTU-D46-17 Pan-Tilt Unit is shown in figure 3.33. Its main parameters:
•
•
•
•
•
•
•
•
•
•
•
56
Max Payload Weight (nominal) 6 lbs.
Position Resolution 0.05143◦ (1/2 step), 0.0123◦ (1/8th step)
Min Pan Speed 0.0123◦ /s
Max Pan Speed 300◦ /s
Min Tilt Speed 0.0123◦ /s
Max Tilt Speed 300◦ /s
Pan Range ± 180◦
Tilt Range +31◦ /-80◦
Weight 3 lbs (not including Controller: 8 oz)
Height 5.2”
Operating Voltage 9–30VDC
3.3. Vision acquisition system
It is offered with the Color Lo-Res StereoCam system for range-finding and 3-D object
detection and recognition as an add-on to PeopleBot (http://www.activrobots.com/
ACCESSORIES/stereocam.html#pan-tilt).
3.3.3. CAMERA INTERFACES
In the image acquisition system high data transmission rates are required between the
camera and the rest of the vision system. We will consider several most popular camera
interface standards, pointing out to their advantages (+) and disadvantages (–).
Composite video
•
•
•
•
•
•
+ popular, inexpensive analog cameras for CCTV,
+ guaranteed and stable frame rate (25/30 fps),
+ frequently with integrated zoom or even PTZ,
– limited resolution (PAL/NTSC color standard),
– need for a frame grabber (digitizer),
– no global shutter and/or progressive scan (interlaced image in most cases).
Used in many PTZ and zoom cameras mentioned above (Canon, Sony) http://www.
usa.canon.com/consumer/>
http://www.canon-europe.com/For_Home/Product_Finder/Web_Cameras/Web_Cameras/
http://www.expandore.com/product/sony/PTZ/
http://pro.sony.com/bbsc/ssr/cat-industrialcameras/cat-block/
Firewire (IEEE 1394)
•
•
•
•
•
•
•
•
+ digital computer interface,
+ standardized data transfer and camera control protocol (DCAM/IIDC),
+ high data throughput (400Mbps 1394a, 800Mbps 1394b),
+ low-level protocol managed by the interface hardware,
+ drivers availability for Win and Linux,
+ operating voltage 8–30V, current limit 1.5A,
+ isochronous image transfer,
– limited availability of integrated PTZ cameras.
Used in many cameras:
http://www.theimagingsource.com/en/products/cameras/firewire_color/
http://www.prosilica.com/products/cv_series.html
http://www.prosilica.com/products/ec_series.html
http://www.baslerweb.com/beitraege/unterbeitrag_en_26297.html
http://www.b2bvideosource.com/IEEE_1394_CAMERAS.html
http://www.b2bvideosource.com/IEEE_1394B_CAMERAS.html
http://www.1394ta.com/index.html
http://en.wikipedia.org/wiki/FireWire#FireWire_400_.28IEEE_1394-1995.29
http://www.prosilica.com/support/why_firewire.htm
There are also Video/Firewire converters available (Canopus ADVC55, TheImagingSource DFG 1394-1e, etc.).
http://www.canopus.com/products/ADVC55/index.php
http://www.theimagingsource.com/en/products/converters/dfg13941e/overview/
57
Chapter 3. Overview of social robot components
3.34. Video/Firewire converter DFG 1394-1e
DFG 1394-1e specification (the device is shown in figure 3.34):
•
•
•
•
•
•
•
•
•
•
Video formats PAL/NTSC, RS-170/CCIR
Max resolution PAL: 768x576 at 25 Hz, NTSC: 640x480 at 30 Hz
Analog inputs 2×RCA (cinch), 1×Y/C (S-Video) - Multiplexed
Bus interface IEEE 1394-1995/1394a, 6-pin, 400 Mb/s
Color formats UYVY, RGB 32, RGB 24, RGB 8
Square pixels yes
Supply voltage 8 to 30 VDC
Current consumption approx 180 mA at 12 VDC
Dimensions H: 32 mm, W: 58 mm, L: 95 mm
Mass 100 g
USB 2.0
•
•
•
•
•
•
•
•
•
+ popular digital interface,
+ wide spectrum of available cameras (web cameras inter alia),
+ high data throughput (480Mbps),
+ drivers availability for Win and Linux,
– asynchronous image transfer,
– non-standard (proprietary) protocols,
– low-level protocol managed by the host-processor,
– limited availability of integrated PTZ cameras,
– operating voltage 4.7-5.5V, current limit 0.5A.
Used in majority of webcams and in many communication, surveillance and industrial
cameras:
http://www.logitech.com/index.cfm/38/3056&cl=us,en
http://www.theimagingsource.com/en/products/cameras/
http://www.ids-imaging.de/inhalte/uEye_table/Table_uEye_e.php
http://www.ids-imaging.com/frontend/overview.php
http://www.everythingusb.com/usb2/faq.htm
CameraLink
•
•
•
•
•
•
58
+ synchronous data transfer,
+ high-speed serial data and cabling standard,
+ very high data throughput (2.38 Gbps),
– frame grabber required,
– proprietary drivers for different camera models,
– limited availability of integrated PTZ cameras.
3.3. Vision acquisition system
The new standard is used in many cameras:
http://www.machinevisiononline.org/public/articles/articlesdetails.cfm?id=
2028
http://www.vision-systems.com/display_article/223650/19/none/none/EmTrd/
High-speed-cameras-and-Camera-Link
http://www.baslerweb.com/beitraege/beitrag_en_17693.html
http://www.machinevisiononline.org/public/articles/articlesdetails.cfm?id=
1108
http://www.machinevisiononline.org/public/articles/index.cfm?cat=129
There are also some vide interfaces available: http://www.matrox.com/imaging/products/
solios_gige/home.cfm
http://www.machinevisiononline.org/buyers_guide/newproducts/details.cfm?id=
1030
http://www.i-cubeinc.com/chartlinks/pixci_e4.htm
http://www.alacron.com/sales/camera_channel_link.htm
GigE Vision
•
•
•
•
•
+ very high data throughput (1000 Mbps),
+ standard gigabit Ethernet hardware,
+ very long cable lengths (100m),
+ standardized camera control registers and command structure (GenICam),
– no PTZ models available (yet?).
http://www.machinevisiononline.org/public/articles/index.cfm?cat=167
http://www.machinevisiononline.org/public/articles/details.cfm?id=2761
http://www.baslerweb.com/beitraege/unterbeitrag_en_64027.html
http://www.baslerweb.com/beitraege/beitrag_en_17693.html
http://www.prosilica.com/products/specifications/ge_series_specs.html
http://www.prosilica.com/products/gb_series.html
http://www.prosilica.com/products/gc_series.html
http://www.prosilica.com/products/ge_series.html
http://emva.org/genicam/current_status_of_genicam%E2%84%A2
http://www2.microeye.eu/inhalte/uEye_table/Table_GigE_e.php
http://www.gigalinx.net/camera_link_to_gige.html
http://www.matrox.com/imaging/products/solios_gige/home.cfm
IP cameras
•
•
•
•
•
+ standard communication protocols (http, ...),
+ wired and wireless connectivity,
+ wide availability, relatively low price,
+ compact PTZ versions available,
– low frame rates (10-15fps for VGA).
http://www.kintronics.com/neteye/PTZ%20cameras.htm
http://www.linksys.com/servlet/Satellite?c=L_Product_C2
59
Chapter 3. Overview of social robot components
http://www.usa.canon.com/consumer/controller?act=ModelInfoAct&fcategoryid=
160&modelid=14897#ModelTechSpecsAct
http://www.linuxdevices.com/news/NS7726407381.html
http://www.ptzcamera.net/
http://www.itechnews.net/2008/09/08/canon-vb-c60-ptz-network-camera/
http://www.linksys.com/servlet/Satellite?c=L_Product_C2\&childpagename=US%
2FLayout&pagename=Linksys%2FCommon%2FVisitorWrapper&cid=1143837459487
http://verint.com/video_solutions/section2b.cfm?article_level2_category_id=
7&article_level2a_id=359&article_level2b_id=708
A bundle of useful links to vision hardware providers can be found at http://www.cs.
cmu.edu/~cil/v-hardware.html.
3.3.4. CONCLUDING REMARKS
The selection of the image acquisition system for a socially interactive robot should
depend on the robot size and shape. In the case of a small robot, the described in 3.3.2
soft pan/tilt/zoom should be considered. This type of image acquisition system, with
no moving parts, silent, of low power consumption, will probably be the best choice.
If we can use mechanical PTZ (e.g. in the case of Pioneer or PeopleBot), there are many
suitable PTZ cameras available. For the indoor usage, zooming factor should be approx.
10, and the viewing angle in wide lens mode at least 60o .
Special attention should be paid to the camera size, if we envision putting it into the
robot head (eye?). Almost all available zoom cameras are far too large for such an
application. The only possibility we can see at the moment is using one of the block
zoom cameras from Sony (e.g. FCB-EX11DP is 35.9×40.8×59.2mm in size).
The Firewire interface (IEEE 1394) should be the right choice, as a Firewire Camera module is available in YARP. In the case when we have to use a video camera
(PAL/NTSC standard), there are several Video to Firewire converters available (e.g.
TheImagingSource DFG 1394-1e, described earlier).
The motors in PTZ and zoom cameras are controlled in different ways, depending on
the camera interface. In the case of IP, Firewire, USB or GigE cameras, the control
commands share the link with the image data. Most of the analog cameras make use of
the standard VISCA protocol (over an asynchronous serial link) for the motor control
and the camera settings (exposition etc.)
In the case of motion detection and analysis, the essential feature of the image sensor
is a global shutter. If enabled, the order of scanning the image field (interlaced or
progressive) does not affect the image quality.
3.4. SOUND ACQUISITION AND EMISSION SYSTEMS
The sound acquisition for communication and speech recognition is one of the serious
problem in a wide range of applications, starting from conference communication systems, hearing aids designs, through Internet phones and voice controlled devices, ending
with car or powered hang gliding communication. The problem arises in human-robot
communication systems as well. Acoustic background noise, echo signals, and robot
driven system din cause the signal degradation, which decreases the performance of
speech recognition systems.
60
3.4. Sound acquisition and emission systems
The acoustic signal disturbances can be eliminated on two stages: on the analogue stage
– during the microphone signals amplification and analogue circuit processing – and on
the digital stage – with use of variety of hardware and software systems, after signal
conversion to digital form. Until now, there are elaborated many computer algorithms
for audio signal preprocessing, designed especially for speech signal enhancement. They
utilise single microphones or microphone arrays signals, are based on a priori models or
on-line model estimation, and can be applied in the robotic companion sound systems.
However, considering the robotic companion hardware, the problem of an appropriate
sound acquisition system should not be neglected. It is important that analogue input
circuits could provide acoustic signals with required characteristics and free of noise and
disturbances.
While constructing a robot speech recognition system one comes across two sources of
signal disturbances: external sounds coming from the robot surrounding, and all the
sounds produced by the robot itself. Admittedly the disturbances belonging to the first
class are undesirable from the speech recognition point of view, however they can be
interesting from the surrounding detection point of view, and thus should be introduced
to the robot input, possibly with an additional channel. Nonetheless, the robot internal
sounds should be completely separated and possibly eliminated. It should be noticed,
that affording possibilities for sound sources localisation would be required feature of
the sound acquisition system.
A robot sound acquisition system should contain a set of properly arranged microphones,
a system of microphone signal conditioners (amplifiers/filters), and analog-digital converters. Unfortunately, there are no ready to use systems dedicated to the robot sound
acquisition systems. In the underneath subsections we present a short overview of robot
sound acquisition system components and briefly describe a testbed for such systems.
The sound emission task appears to be much easier than its acquisition. However, in
the case of sound systems for social robots some attention has to be put on the power
amplifier and loudspeaker parameters, such as power consumption, efficiency, sound
clarity. Moreover, the whole system dimensions are one of the crucial selection criteria.
In subsection 3.4.5 we characterise example loudspeaker/amplifier systems suitable for
the sound emission in the social robot case.
3.4.1. MICROPHONES
Omnidirectional electret microphones
Electret microphones are the most popular and universal microphone type. There are
many providers of these devices, however, comparing the parameters of different microphones one can conclude that they performance is quite similar and they can be
categorised in few classes. Below, there are given exemplary electret microphones types
utilised in experiments with a robot sound acquisition system.
Electret microphone BCM-9765P (Bestar Electronic Industry)
•
•
•
•
•
•
Dimensions: φ9.7mm×4.5mm,
Directivity: omnidirectional,
Sensitivity: -44±2dB (at L=50cm, f=1kHz,
VS=3V, RL=2.2kΩ, 0dB=1V/Pa),
Frequency: 20–16000Hz,
Operating voltage: 1.5–10V,
S/N ratio: >60dB.
61
Chapter 3. Overview of social robot components
Electret microphone WM-55A103 (Panasonic)
•
•
•
•
•
•
•
Dimensions: φ9.7mm×5.0mm,
Directivity: omnidirectional
Sensitivity: -47±4dB (at L=50cm, f=1kHz, VS=3V,
0dB=1V/Pa),
Frequency: 100–16000Hz,
Operating voltage: 1.5–10V,
Current consumption: <0.5mA,
S/N ratio: >60dB.
Unidirectional dynamic microphones
There is a large set of different unidirectional dynamic microphones, unfortunately their
size practically precludes their application in a robotic companion. Below, there is
presented one, small size unidirectional microphone used during experiments with a robot
sound acquisition system. Worth noticing is the second presented microphone (MM216),
which is a special construction double microphone designed on purpose for using in
noisy environments, which actually is not real unidirectional microphones, but due to
its construction favours sounds coming from a single direction over surrounding sounds.
Dynamic microphone MDU 43 (Tonsil)
•
•
•
•
•
•
•
Dimensions: φ21mm×48mm,
Directivity: unidirectional (cardioid),
Sensitivity: -62dB (0dB=1V/Pa)
Frequency: 100–12000Hz,
Impedance: 200Ω,
S/N ratio: >60dB,
Web: http://www.skleptonsil.pl/go/_info/?id=767.mm
Differential microphone MM 216 (Tonsil)
•
•
•
•
•
•
•
Dimensions: 22.7mm×11mm,
Directivity: omnidirectional,
Sensitivity: -62dB (0dB=1V/Pa)
Frequency: 300–3400Hz,
Impedance: 600Ω,
S/N ratio: >60dB,
Web: http://www.skleptonsil.pl/go/_info/?id=777.
3.4.2. ANALOG SIGNAL CONDITIONING
Before microphone signals are digitalised they have to be amplified and filtered. In these
tasks operational amplifier circuits are utilised. The amplifier constructed for tests is
a two-stage amplifier with the first stage being differential. This allows for connecting two
microphones simultaneously, what can be used for the robot internal sounds cancellation.
The bandwidth of the amplifier is limited to the voice frequencies. Below two types of
operational amplifiers applied in the design are presented.
Operational amplifier TL072 (Texas Instruments)
•
•
•
•
•
62
Bandwidth: 3√MHz,
Noise: 18 nV/ Hz,
Input bias current: 65 pA,
Supply voltage: ±9 to ±15 V (typical),
Web: http://focus.ti.com/lit/ds/symlink/tl072.pdf.
3.4. Sound acquisition and emission systems
3.35. AMTEC A/D converters module
Operational amplifier AD823 (Analog Devices)
•
•
•
•
•
Bandwidth: 16
√ MHz,
Noise: 16 nV/ Hz,
Input bias current: 3 pA,
Supply voltage: ±3 to ±30 V (typical),
Web: http://www.analog.com/static/imported-files/data_sheets/ad823.pdf.
3.4.3. A/D CONVERTERS
There is a large number of different A/D converters cards commercially available in the
market. Below we shortly characterise an A/D converters pc/104 module chosen for the
experiments.
PCM-3718H/HG multifunction module
PCM-3718H/HG (figure 3.35) is a 12-bit multifunction pc/104 module with programmable gain.
•
•
•
•
•
•
•
16×single-ended or 8 differential analog inputs,
12-bit A/D converter, up to 100 kHz sampling rate with DMA transfer,
Two 8-bit digital input/output TTL level channels,
Input impedance: 10MΩ,
I/O connectors: 2x20pin box header,
Power consumption: 5V at 400mA,
Drivers: Windows and Linux (open source).
3.4.4. TESTBED FOR SOUND ACQUISITION SYSTEM
To experiment with the influence of the robot driving system din and the surrounding
noise on the robot sound acquisition system, and to analyse the properties of different
microphones and conditioners layouts, an experimental testbed has been prepared (see
figure 3.36). The main goals of the experiments are to eliminate the driving system
din, to emphasise the voice frequencies band, to extract sounds coming from the robot
environment, and to provide the necessary information for the sound localisation system.
3.4.5. LOUDSPEAKERS AND POWER AMPLIFIERS
Enormous number of different type loudspeakers and power amplifiers is present in the
market. While assembling a sound emission system for a robotic companion the dimensions and the power consumption of the loudspeaker-amplifier combination play vital
63
Chapter 3. Overview of social robot components
3.36. Layout of the testbed for sound acquisition system
role. Of course, the sound quality is not to be overlooked. Below, we present example
loudspeakers and amplifiers belonging to two different classes: the first, simple and
inexpensive devices, but providing quality sufficient for voice emission, and the second,
containing more sophisticated and expensive equipment, but small and of high quality.
MAC5310 + BAS8100 modules
The MAC5310 Audio Codec and Audio
Power Amplifier Module can be used as
an audio input/output interface. Together
with BAS8100 Speaker Module it forms
a complete audio channel. The speaker is
characterised by small size and sensible parameters sufficient for voice emission.
Modules parameters:
•
•
•
•
•
•
•
•
•
•
High quality 16Bit audio ADC&DAC
codec,
Microphone preamplifier,
Power amplifier (1W at 8Ω),
Standard i2s interface,
Speaker power handling: 2W,
Speaker frequency response: 320–20000Hz,
Speaker impedance: 8Ω,
Manufacturer: Dr Robot Inc.,
Web: www.drrobot.com,
Price: ∼EUR 45.
Chopper amplifier module DK-DB00x
DK-DB00x is a family of D class audio amplifiers of high efficiency, reliable, shockproof, and small in size. They are different
from the traditional analog design which is bigger in size, heavier, and require radiators.
Amplifier parameters:
•
•
•
•
64
Two channel, full audio frequency,
OEM module with a case,
Output power: 2W×2, 30W×2, 50W×2, 100W×2,
Frequency response: 20–20000Hz,
3.4. Sound acquisition and emission systems
•
•
•
•
Signal to noise ratio: >87dB, THD <1%,
Power supply: DC 5V/12V/16V/24V/30V,
Manufacturer: Shenzhen DKing Technology Co., Ltd.,
Web: www.chinadking.com, dking.en.alibaba.com/product/
50166083/51264106/Digital_Boxes/Digital_Amplifier_
Modul.html.
AM-S30d amplifier
AM-S30d is a two channel chopper amplifier. It offers high quality
parameters and small size simultaneously, its main advantage is 90%
efficiency.
Amplifier parameters:
•
•
•
•
•
•
•
Two channel, full audio frequency,
OEM module without a case,
Output power: 2x20W (8Ω), 2x30W (4Ω),
Frequency response: 20–20000Hz,
Signal to noise ratio: >87dB, THD <0.05%,
Manufacturer: Flying Mole Co.,
Web: flyingmole.co.jp/new_en/index.html.
HiFi full range speaker Monacor SPH-30X/8
This full range HiFi speaker combines numerous highlights of the speaker
technology on a minimum of space: a Kevlar cone of extra low weight,
but nevertheless well damped, a motor system with full air ventilation, a
high-quality NEODYMIUM magnetic system with double magnet and a
newly developed basket, which provide good exposure in noisy environments
Speaker parameters:
•
•
•
•
•
•
Dimensions: 80.5mm×80.5mm×42mm,
Output power: 20W,
Frequency response: 105–17000Hz,
Manufacturer: Monacor International GmbH & Co.,
Web: monacor.djshop.pl/_monacor/produkt/4916,monacor.html,
Price: EUR 30.
3.4.6. CONCLUDING REMARKS
Because of a wide range of commercially available audio amplifiers and loudspeakers,
assembling of the robotic companion sound emission system is relatively simple to perform. Properly chosen loudspeakers with a power amplifier will do the job. The selection
will depend of the sound emission tasks: for voice emission cheap and popular solutions
would work, as for multimedia tasks more careful choice has to be made.
On the other hand, there are no commercially available, relevant and reliable sound
acquisition systems. Therefore, a specialised system should be designed to manage all
the disturbances issues. Admittedly, there is a wide suite of computer algorithms for
digital audio signals processing and mastering, but possibly all disturbances should be
eliminated at the source.
The robot sound acquisition system should contain a set of properly arranged microphones, a system of microphone signal conditioners, and analog-digital converters. A set
of omnidirectional microphones driving differential preamplifiers will help to eliminate
the robot din and to provide the necessary information for the sound localisation system. An additional, unidirectional microphone equipped with pan/tilt capability would
be desirable feature.
65
Chapter 4
Body specification of FLASH
This chapter presents a short characterisation of a tentative LIREC robotic companion
FLASH (Flexible LIREC Autonomous Social Helper). We begin with providing basic
technical features of the robot. Than, we shortly describe the construction of two main
hardware components: the mobile base and the hands.
The specification of the LIREC companion FLASH provided in this chapter results from
the requirements imposed on the robot by the robotic scenarios, and from an analysis of
existing social robots and robotic components accomplished in chapters 2 and 3. As a
conclusion, in this chapter we have defined the body of FLASH, made a recommendation
of the balancing mobile platform, and proposed a specific design of FLASH’s hands. Both
these components should be treated as preliminary prototypes that will be tested in small
scale experiments, and possibly re-designed in consequence of these experiments. The
remaining robot’s components, especially the visual and the auditory systems, so far
have not been defined uniquely. Instead, in chapter 3, we have formulated a number of
design recommendations that need to be refined in the course of further research.
4.1. BASIC TECHNICAL FEATURES
•
Purpose: research platform, robotic companion,
•
Appearance: mechanoid,
•
Components:
◦ mobile base,
◦ torso,
◦ head and neck,
◦ pair of hands,
◦ vision system,
◦ sound acquisition and emission systems,
◦ other sensor systems,
◦ on-board computers,
•
Operating conditions:
◦ Conducive indoor conditions,
◦ Daylight or lamplight illumination,
◦ Continuous Internet connection,
•
Mobility:
◦ Type: Wheeled, self-balancing,
◦ Sensors: Position measurement, obstacle detection, navigation,
◦ Autonomy: Obstacle avoidance and navigation to map-registered locations,
•
Manipulability:
◦ Type: Limited, with a pair of hands designed for gesticulation,
◦ Sensors: Position measurement, touch detection,
◦ Autonomy: Limited grasp?,
•
Communication:
◦ Human detection: Detection of persons, face detection,
◦ Individual recognition: Recognition of 1-2 persons, recognition of facial characteristics,
Chapter 4. Body specification of FLASH
◦
◦
Sound: Acquisition of surrounding sounds with directional, information and
voice band emphasis, sound emission with speech synthesis,
Body language: Recognition of gestures and facial expressions, gesticulation
and facial expression,
•
Hardware specification:
◦ Robot height: ∼1.2m,
◦ Robot diameter: ∼0.45m,
◦ Motion:
— Platform: self-balancing, class (2,0),
— Hands: even-gesticulating, 8 DOF each,
— Neck: lithe, 3 DOF,
— Head: expressive, max 20 DOF,
◦ I/O Systems:
— Visual: binocular stereo? with pan, tilt and zoom,
— Auditory: directional microphone, omnidirectional microphones array,
— Voice: loudspeaker,
◦ Power supply: lithium polymer battery,
•
Controller:
◦ CPU: multiprocessor configuration,
◦ Operating system: Linux,
◦ Control architecture: serial bus connection distributed processing based on
YARP.
4.2. BALANCING PLATFORM COSMOS
Following the recommendation from subsection 3.1.2, the robot FLASH should be equipped with a balancing mobile base dedicated to its functions and objectives. Besides of
being balancing, the base of FLASH should possess the following features:
•
•
•
•
•
lightweight and inexpensive,
safe and autonomous rolling and balancing,
open, modifiable control and communication architecture,
integrability under YARP,
adaptability to migration.
Further on in this section we shall describe a prototype of such a platform designed at
WRUT.
The construction of Cosmos (the first prototype of WRUT balancing platform) is quite
simple. The name Cosmos reflects the first reaction of students who saw the platform
in motion: they have exclaimed: “what a cosmos”, which in the language of young
Polish people expresses a mixture of surprise and admiration. The platform has been
built using common low-tech electrical parts, standard airplane model wheels, and cheap
commercially available sensors. The view of Cosmos is presented in figure 4.1. The
platform consists of the following components:
•
•
•
•
•
•
•
•
•
68
main control unit with Freescale MPC555 CPU with touch screen 128x240px,
voltage converter with DAC,
double H bridge - a motor driver,
power supply module +5V, +6V,
internal navigation system (INS) with accelerometer, gyro and ADC,
24 key keyboard,
LiPoly pack 14,8 4800mAh,
two motors 12V Graupner 500E with gearbox,
two magnetic encoders AS5040.
4.2. Balancing platform Cosmos
4.1. Cosmos
4.2. Chassis
69
Chapter 4. Body specification of FLASH
4.3. Sensors module
4.4. DC motor
Chassis
Cosmos has a 2 wheel, balancing mobile base. Its chassis has been made of aluminium,
the dimensions are displayed in figure 4.2. Due to simplicity of the design, all the
components can be installed quickly. All electronics modules, like controller, keyboard,
motor driver, power supply, etc. have been situated in the lower part of the Cosmos
body. Motors with gearbox and rotor-position sensors are fixed at the bottom. The
LiPoly battery is fixed in the middle part of the body.
Sensors
ADXRS150 Analog Devices digital rate gyroscope and a MMA7260 Freescale digital
3-axis accelerometer (see figure 4.3) have been installed on the platform in order to
measure the tilt angle of the platform as well as its angular velocity. The gyroscope
provides a measure of instantaneous angular change, but it produces a significant drift
when the gyroscope is operating. This may be due to the operating temperature or
inherent characteristics of the gyroscope itself. On the other hand, the inclinometer
provides an absolute measure of inclination, but the output signal is often corrupted
with noise. To overcome these problems a signal-level sensor fusion technique has been
used, based on the the Kalman filter. Two magnetic encoders are employed to measure
the position of each wheel.
Actuators
The actuation to balance the robot is provided by two widely used 500E DC motors
made by Graupner, Germany, shown in figure 4.4. Each motor has a gear reduction of
36:1, and a torque constant of 0.07 kgm/A. At the moment, the applied gearboxes show
an excessive backlash.
70
4.3. Arms and hands
4.5. Controller
Controller
PHYTEC phyCORE-MPC555 onboard module plays the role of the platform’s brainŹ.
The module consists of a powerful 32-Bit microcontroller PowerPC Freescale MPC555
(see figure 4.5) running at 40MHz, there is 1MB ROM and 1MB RAM onboard. MPC555
used in phyCORE-MPC555 is an advanced microcontroller with 64-bit float point unit,
2 channels QSPI interface, 2 SCI interfaces, 2 TPU3 and MIOS1 timers/counters blocks.
This allows that relatively large and computationally demanding programs be executed
efficiently. The controller is programmed in C. A user interface is realised by a touchscreen presented in figure 4.6.
Control algorithms
Control algorithms applied in balancing platforms should simultaneously steer the rolling
motion of the wheels and provide a stabilisation of the balancing movements of the platform. Temporarily, in Cosmos platform a common solution has been adopted, i.e. a
simple PID control for the wheels and a stabilisation algorithm based on the linearisation of the platform’s dynamics. Further research work has been undertaken toward
derivation and identification of a nonlinear model of the dynamics, and its employment
in the control algorithm. A general functional schematic of Cosmos has been displayed
in figure 4.7
4.3. ARMS AND HANDS
As we have concluded in the social robot hands overview, there is a need for designing
a robotic hand aimed to gesticulation. Below the hand construction designed for the
FLASH robot is outlined. On the design stage, the following assumptions have been
made:
•
•
•
•
•
•
•
design for gesticulation task,
not expensive, light construction,
safe and smooth movements,
easily controlled with a set of predefined gestures,
built of common, easily available elements,
reconfigurable,
open source and YARP based communication.
The construction (see figure 4.8) is characterised by the following parameters:
•
•
Complete arm with hand,
Bearing joints,
71
Chapter 4. Body specification of FLASH
4.6. Touchscreen
4.7. Functional schematic of Cosmos
72
4.3. Arms and hands
4.8. Design of the FLASH hand and arm
4.9. Arm links connection
•
•
•
•
•
•
•
•
5 high torque and high speed digital servos (Dynamixel), 3 lightweight micro servos,
Trajectory tracking, predefined gestures,
Physical parameters: total length ∼60cm, weight ∼1kg,
DOF: arm 5, hand 3,
Carrying components: carbon fiber tubes, aluminium elements,
Control system: distributed, PC/104 based, with limited force control,
Communication: internal RS485, external Ethernet powered by YARP,
Power supply: 18VDC.
The designed arm consists of 2 links connected via a single 1DOF joint (figure 4.9). The
arm is to be screwed to a robot body via 3DOF joint (figure 4.10), and is endowed with
another 1DOF joint, to which a hand can be mounted. The hand (fig 4.11) is formed
of four 1DOF fingers. The thumb and the index finger are driven by two separate
microservos. The other two fingers are driven by one, shared microservo.
73
Chapter 4. Body specification of FLASH
4.10. Arm shoulder
4.11. Hand construction
4.4. CONCLUDING REMARKS
In conclusion, we want to emphasise the following features of robot FLASH:
•
•
•
•
•
•
74
FLASH
FLASH
FLASH
FLASH
FLASH
FLASH
is a flexible research platform with stable and believable behaviour,
is dedicated to the implementation of a long term robotic companion,
admits changeable embodiments,
provides a hardware base for an autonomous actor in robotic scenarios,
is adaptable to migration,
has open architecture, software, and documentation.