Home Consortium Research Activities Deliverables Publications Links Contact

  Research Activity 7: Systems level integration & Evaluation

'...Key Experiments are derived from the research activities, and serve as integration platforms for their results. Each Key experiment focuses on one or more fundamental abilities of a cognitive robot...'


Partners involved

Objectives

Results

Final review presentation

Videos

Deliverables


Partners involved(RA leader)

IPA, all partners IPA is the partner in charge of the overall RA7 activity, and each Key Experiment has an identified leader, responsible for the implementation of one of the three scenarii:

  • The Robot Home Tour (LAAS)
  • The Curious Robot (UniBi)
  • Learning Skills and Tasks (UniKarl)

Objectives

The COGNIRON project is organised along 6 Research Activities. In order to demonstrate their outcomes and results in an integrated fashion, a seventh Research Activity (Research Activity 7) has been introduced. Research Activity 7 is a specific project activity, as it concentrates on the integration of several robot functions and their experimentation and evaluation in the context of well-defined scenarios, or Key Experiments, that enable to exhibit the cognitive capacities and to assess progress. The three Key Experiments are:

  • The Robot Home Tour
  • The Curious Robot
  • Learning Skills and Tasks

In this research area the three key experiments (KE) were implemented and tested. The key experiment settings are four robots realizing scenarios which are instantiated and detailed by so-called scripts. These describe certain sequences of possible robot interactions with the environment (including humans). Within the KEs, the COGNIRON functions (CF) were studied in the context of the research areas (RA) of the project. The COGNIRON functions are implemented as one or several software services which are smaller sub-components. The robots accomplish their missions based on a set of the services, including supporting services which are in some cases not a direct outcome of any RA, but needed to realize the scenario. A large effort in WP7 was devoted to development and implementation of these components and to the integration of the COGNIRON functions

cogniron key experiments

The key experiment robots and partner interactions based on the COGNIRON functions, services and supporting services. Please refer to the RA7 deliverable 2007 for descriptions of the function abbreviations.

Results

WP7.1 - Evaluation of functions and Key Experiments

The objectives of this work package were to evaluate the key experiments based on criteria that reflect technical evaluations and evaluations from a user's perspective and to report on the results in a joint document.

The four robots of the three key experiments were equipped with functionalities that are necessary to run the scripts of the individual KE. All scripts were recorded as videos with the real robots acting according to a storyboard that was specifically designed for the evaluation from a user perspective (see also the RA3 documents). In some cases the robots were remotely controlled since the final developments of critical functions were still in progress. Those functions that were ready and testable on a technical level were evaluated by data sets, small (selected) script runs and some further conceptual estimation of the scalability and stability. Some impressions of the evaluation activities of all three key experiments are shown in the figure below. All evaluation results were collected and consolidated into the final RA7 deliverable (D7.2007).

evaluation activities

Some impressions of the evaluation activities in WP7

WP7.2 - Key Experiment 1: The Robot Home Tour

Objectives

The key-experiment KE1 "Robot Home Tour" focuses on multi-modal human-robot interaction to learn the topology of the environments (apartments) and its artefacts, the identity and location of objects and their spatial-temporal relations. The objectives of KE1 as a part of RA7 is to provide a scenario to study dedicated research question (with a focus on RA 1, 2, 5, and 6) and as a test-bed for the evaluation of developed approaches. The robot platform in KE1 is BIRON. In the last project phase, KE1 had a special emphasis on evaluation with a combination of live user trials and video studies. Therefore, the major objectives can be summarized as (i) to implement an evaluation- ready system and to actually evaluate it, and (ii) to sustain and ensure a life time of developed approaches and generated data beyond the end of the COGNIRON project by supporting dissemination and training activities.

Scenario

A robot, taught and shown around by a human, discovers a home-like environment and builds up an understanding of its surroundings. The robot engages a dialogue with the human for naming locations and disambiguating information. The 'Home Tour' key experiment demonstrates the dialogue capacities of the robot, and the implementation of human-robot interaction skills as well as the continuous learning of both spaces and objects.

KE1

BIRON in the robot house

State of the setting

The KE1 has been implemented a release strategies with three different release threads: Evaluation, Training, and Final. All these threads have been finalized and full-integrated system evaluations have been carried out. A video study focused on the character of the robot and compared extrovert and introvert robot behaviour in the home tour setting. Three iterations of user trials with subjects of decreasing familiarity with robots have been conducted and compiled into a corpus of interactions with an autonomously operating robot.

Cogniron functions shown in KE

The cognitive function being demonstrated in KE1 are

  • CF-DLG: grounding-based dialogue system, the dialog style is adaptive and features as well extrovert as also introvert behaviours
  • CF-ROR: resolving object references on the basis of pointing gestures and object attributes to facilitate object and spatial learning. It supports the learning of spatial concepts in the KE
  • CF-DTT: dynamic topic tracking to identify and track conversational topics in dialogues to disambiguate instances
  • CF-AM: an implementation of active memory to fuse information of different domains to allow situation-aware processing and behaviours. In the final release, this architectural paradigm is consequently applied in KE1
  • CF-PTA: tracking persons and control person attention for situation-aware conversation control, featuring attention switches between person and a multimodal fusion of visual and acoustic percepts
  • CF-NAV: navigation between places identified by labels. Obstacle avoidance and short distance path planning are part of this function, too. It is not the focus of KE1, but applied as a support service.
  • CF-LOC: topological localization using different modalities: Laser-based and feature-based approaches are integrated. The two different approaches are transparently integrated in the system and their embedding in the interactive dialog model has been in the center of research.

WP7.3 - Key Experiment 2: The Curious Robot

Objectives

The objectives of this work-package are the final and full implementation and demonstration of Key Experiment 2 with two main underlying scientific issues:

  • The first issue addresses situation recognition and decision making to accomplish a task without an explicit request from the human but rather by inferring that its completion is necessary for a human action. The robot will express its intentions in a legible manner to the human on the basis of the results of the user studies carried out in work-packages 3 and 6.
  • Close human-robot interaction constitutes the second main issue for this Key Experiment. Handing an object to a human in a safe and comfortable manner (see WP3) requires knowledge on human perception, posture and workspace.

Scenario

A robot in an indoor environment has identified and is observing a person in the room and interpreting its activity. It identifies some potential needs from the person, for instance a drink. It then interrupts its own task to answer to the person's needs by fetching the object. The robot thus anticipates on a situation that may occur, and acts to facilitate the future action of a person. The 'Curious Robot' key experiment allows the demonstration and assessment of initiative taking and knowledge acquisition, as well as intentionality object recognition and close interaction with humans.

KE2

The robot Jido handing out a juice bottle to Luis

State of the setting

The challenge in KE 2 as part of the RA 7 is the capacity of the robot achieve a task based on explicit reasoning on its abilities and on the presence, state, needs, preferences and abilities of the person with which it interacts.

During the last project phase, we enhanced the robot equipment, extended its functional and decisional abilities and established an incremental development and illustration process. We have also developed a set of debugging and visualization tools in order to exhibit the data structures and the knowledge that are manipulated explicitly by the robot.

evaluation activities

an example of a fetch-and-carry task in interaction with a person

The focus and the main contributions come essentially from RA 2, 3, and 6 with contributions from RA 1 and 5. For a complete description see Deliverable D7-2007.

Cogniron functions shown in KE

The robot has been endowed with a set of Cogniron Functions and a set of complementary services. The integrated Cogniron functions are:

  • CF-PTA/CF-TBP - Person tracking, detection of attention and 3D multitracking of the head and the two hands simultaneously.
  • CF-RET - Reasoning on human-robot collaborative tasks: This is realized through SHARY (a robot supervisor specifically designed to integrate cooperative task achievement scheme inspired by joint activity theory) and HATP (the associated task planner).
  • CF-ACT - Detection and interpretation of human activities and postures. The fundamental component for activity recognition is the body-part tracking module called VooDoo from UniKarl and GEST (based on stereoscopic data).
  • CF-NHP - Navigation in human presence
  • CF-DLG - Multi modal dialogue essentially focused on the verbal and non-verbal communication acts associated to the establishment of a joint task between a robot and its human partner and the different decisions that are involved in the task achievement.
  • CF-MHP - Manipulation in human presence. This module is in charge of the planning of the arm's motion by taking into account the human partner when it is in close vicinity.
  • CF-OR - Object recognition and modelling

A complete robot software architecture has been implemented and is used for several situation involving fetch-and-carry tasks in presence of humans in an apartment-like environment. The robot is capable to interact and to hand objects to persons in various situations. Efforts have been devoted in order to endow the robot with the ability to performs its tasks in a legible and acceptable manner.

WP7.4 - Key Experiment 3: Learning skills and tasks

Objectives

The objectives of this work package are the final and full implementation and demonstration of key experiment 3: Learning Skills and Tasks. The script of this experiment had to be updated and more precise definition of the Cogniron functions were needed to realize it. The key-experiment is split into two different scripts and will be presented on two different robots. This was decided by the consortium due to the fact that two different learning methods are subject to research and that the robot used for the second script in current development and therefore it is not yet available. The new robot (Care-O-bot 3) has important CPU capacity and will possess a most modern sensory-actuation setting. The first script is concerned with Learning Skills: "Arranging and interacting with objects," while the second script is concerned with Learning Tasks: "Serving a guest." In the first script the topic of learning through imitation is investigated while in the second script learning on elementary operator level is the focus of interest.

Scenario

A person makes a few demonstrations of a certain task to a robot. During the demonstration the user possibly explains his actions with vocal comments. While watching the demonstrations the robot learns the relevant parts of the task (e.g. geometrical relations between objects) and new skills such as object-actions relations. Once the demonstrations are finished the robot tries to repeat the task on the basis of what it has learned, with questions to the demonstrator on how best to complete the task. The human instructor may correct the robot's action if asked or needed or provide additional demonstrations of the same task to enhance the robots knowledge on the task at hand. The 'Learning Skills and Task' Key Experiment allows the implementation and assessment of abilities related to learning skills, tasks and object models, dialogue, and understanding human gestures.

KE3

Learning Important Features of the Task and Object recognition and tracking

KE 3a Arranging and Interacting with Objects

The script KE 3a aims at evaluating the CFs CF-LIF and CF-RG (services imitateWhatTo, imitateHowTo and imitateMetric from RA 4). The software components are developed by EPFL and UH. The scenario includes a human demonstrator and a robot imitator. The demonstrator stands in front of the robot and demonstrates a simple manipulator task to the robot by displacing simple objects such as wooden cubes or cylinders of different colours. The demonstrations can be performed either by using motion sensors attached to the body of the demonstrator or by moving the robot's arms (kinaesthetic demonstration). Once a demonstration is finished the robot tries to reproduce the task. After each reproduction, the demonstrator evaluates the robot's performance at fulfilling the task's constraints and decides whether a new demonstration is required or not. The scenario finishes when the robot is able to generalize the skill correctly, i.e. when it reproduces the task correctly when faced with different initial configurations of the workspace.

KE 3b Learning to Serve a Guest

  • 1. The robot enters the room (remotely controlled)
  • 2. The robot positions the gripper to retrieve the object of interest from the user
  • 3. The robot varies the views and records the images sequence of the object model.
  • 4. The user (as tutor) shows the object of interest (alternatively to 3.)
  • 5. The object detection function builds a feature-point cloud model.
  • 6. The robot grasps the object (here based on a pre-defined position).
  • State of the setting

    The new hardware of the Care-O-bot 3 that is used in KE3 was released. The two robots of KE3 were programmed such that they are able to run services that relate to needed Cogniron functions to run episodes of the scripts. Last implementations, optimization and testing took place. The two robots were run at EPFL and IPA. The learning function CF-LCT was implemented and tested off-line on another setting at UKA. This was decided due to the fact that the recognition abilities did not provide information rich enough to display the capacity of the function. Evasive sensor equipment at UKA was used to test these parts.

    Cogniron functions shown in KE

    The Cogniron functions that are integrated on the robot for the first script of KE3 are:

    • CF-LIF Learning important features of the task: This that is the main task knowledge generalization function in this setting.
    • CF-RG Reproducing gestures: This function contains the task reproduction part.
    • Due to the reason that not the full functionalities of CF-DLG and CF-OR were needed in this KE they were replaced by suitable components available at EPFL.

    The test results are given in the deliverable D7.2007. They show that observational learning of e.g. simple pick-and-place tasks is possible using the approach developed in Cogniron. Furthermore, it is shown how the selection of different fitness functions influence the learning result.

    In order to run the second script the new Care-O-bot 3 was programmed to include the following functions:

    • Various low-level functions to control the hardware parts
    • CF-OR Object recognition: This function is here implemented as a complete trainable 3D object detector that can retrieve the 3D position of an object in a large scene in less than a second. The training can take place in two modes: either the user shows the object to the robot or the user gives the object to the robot and the robot varies the views. Collision-free object grasping and placing: These supporting services are used for the task reproduction.
    • To better initialize the Voodoo software from UKA (used to realize CF-GR/CF-ACT) we implemented and test a new supporting service for person detection and body-model initialization (see report on RA2).
    • The speech recognition could not be implemented on the new robot due to delays in the integration of the vision components described above.

    Deliverables

    2004 - D7.1.1 Specification of the 3 key experiments
    2004 - D7.2.1 Set up of the Key Experiment "Robot Home Tour"
    2004 - D7.3.1 Set up of the Key Experiment "Curious Robot"
    2004 - D7.4.1 Set up of the Key Experiment "skill/task learning"
    2005 - RA7 Refined Specification of Key Experiments and Implementation Status
    2006 - RA7 Refined Specification of Key Experiments and Implementation Status
    2007 - RA7 Refined Specification of Key Experiments and Implementation Status

    Final review presentation

    RA7 presentationby Jens Kubacki (IPA)

    KE1 Presentation by Marc Hanheide (UniBi)

    KE2 Presentation by Rachid Alami (UniBi)

    KE3-A Presentation by Aude Billard (EPFL)

    KE3-B Presentation by Martin Haegele (IPA)

    Videos

    A collection of videos could be found there

    .

     

    An Integrated Project funded by the European Commission's Sixth Framework Programme