schreiter-et-al-2024-thor-magni-a-large-scale-indoor-motion-capture-recording-of-human-movement-and-robot-interaction
Introduction to THOR-MAGNI
Dataset Overview: THOR-MAGNI is a comprehensive dataset focused on indoor human and robot navigation and interaction, designed to enhance research on social human navigation.
Purpose: It aims to model and predict human motion, analyze interactions between humans and robots, and investigate visual attention in social contexts.
Key Features: Interaction-focused, includes rich context annotations, and consists of walking trajectories, gaze-tracking data, lidar, and camera streams recorded from a mobile robot.
Uniqueness: It provides a greater variety of contextual features and scenario variations than existing datasets, addressing previous gaps in human motion modeling.
Background and Significance
Rapid Growth Field: The focus on human motion modeling and interaction with robots is increasing, driven by the need for safer and more efficient algorithms in human-robot interaction (HRI) settings.
Importance of Data: Quality human motion data is critical for developing human-aware path planning, collision avoidance, and understanding human activities.
Current Limitations: Existing datasets often lack comprehensive contextual integration and exogenous factors influencing human behavior, which THOR-MAGNI aims to rectify by including varied contextual variables.
Dataset Construction and Features
Data Composition
Includes 3.5 hours of motion recordings across 52 sessions, utilizing multiple sensor types (lidar, eye tracking).
Records various scenarios of social human-human and human-robot interactions, focusing on diverse activities like navigating and object transport.
Contextual Features
Environmental Factors: Participant movement is influenced by surrounding dynamics, such as the presence of other robots, obstacles, and semantic cues.
Agent Cues: Human behaviors are complemented by cues such as head orientation, gaze direction, and past activity patterns, contributing to a richer understanding of intentions and actions.
Technical Configuration
Recording Setup: Data was collected in controlled laboratory settings using motion capture systems and various sensors (3D lidar, RGB cameras).
Multi-modal Data: Synchronization between motion capture and eye-tracking data is achieved using advanced calibration and data alignment techniques.
Scenario Design
Different Scenarios: The dataset features several scenarios that vary by conditions and types of interaction, allowing for the isolation of specific factors affecting motion and interaction dynamics.
Key Conditions:
Navigation styles (differential vs. omnidirectional)
Task-based interactions (e.g., transporting objects, joint navigation with robots).
Roles: Participants are categorized into roles (Visitors vs. Carriers) to simulate realistic interactions and task environments.
Example Scenarios
Scenario 1: Focuses on social behavior in a static environment.
Scenario 2: Introduces roles and tasks (Visitors and Carriers working concurrently).
Scenario 3: Studies the impact of different robot motion styles on human behavior during navigation conditions.
Comparative Analysis
Comparison to Existing Datasets: THOR-MAGNI holds improvements over previous datasets (e.g., THOR, UCY, ETH) due to its extended duration, context variability, and inclusion of robot interaction.
Data Representation: It utilizes various metrics for evaluating motion dynamics in crowded environments, enhancing the understanding of social interactions and human behaviors.
Data Management and Utilization
Tools Provided: Alongside the dataset, tools for data visualization and preprocessing are offered to assist researchers in extracting insights efficiently.
Future Directions: Ongoing efforts will focus on real-world data collection, expanding environmental variations, and refining human-robot collaboration studies.
Conclusion
Significance of THOR-MAGNI: This dataset serves as a pivotal resource for advancing studies in human motion prediction, HRI, and contextual behavior analysis in shared environments, ultimately fostering the development of more adaptive and intuitive robotic systems.