Human-to-Robot Handovers
Solution design and development
- Inferences must be generated automatically by a model that uses images or videos as input
- No prior knowledge of the objects is available. The only prior knowledge available to the models is the high-level set of categories of the containers (cup, drinking glass, food box).
- The use of prior 3D object models is not allowed (e.g., the reconstructed shapes of the containers in 3D).
- Each subject is instructed to use only one hand and always the same.
- The location where the robot needs to deliver the container must be inferred using perception from the initial location and not hard-coded (to be confirmed).
- Learning across executions of configurations is not allowed: participants must not update/fine-tune the vision/robotic algorithms using the data/measurements captured during the execution of the configurations of the benchmark by the subjects (test time).
- Online solutions - i.e., solutions that can be run on a continuous stream as for the case of human-to-robot handover - are preferred. To encourage this type of solution, organisers will refer to the CORSMAL real-to-simulation framework that allows the participants to observe models would perform for a human-to-robot handover.
- Calls to exisiting Large Language Models are allowed.
- Tactile sensors are allowed as part of the solution.
- Both RGB and RGB-D inputs can be used for the solution.