A Benchmark for Recipe Understanding in Artificial Agents

Jens Nevens, Robin De Haes, Rachel Ringe, Mihai Pomarlan, Robert Porzel, Katrien Beuls, Paul Van Eecke

Research output: Chapter in Book/Report/Conference proceedingConference paper

1 Citation (Scopus)
15 Downloads (Pure)

Abstract

This paper introduces a novel benchmark that has been designed as a test bed for evaluating whether artificial agents are able to understand how to perform everyday activities, with a focus on the cooking domain. Understanding how to cook recipes is a highly challenging endeavour due to the underspecified and grounded nature of recipe texts, combined with the fact that recipe execution is a knowledge-intensive and precise activity. The benchmark comprises a corpus of recipes, a procedural semantic representation language of cooking actions, qualitative and quantitative kitchen simulators, and a standardised evaluation procedure. Concretely, the benchmark task consists in mapping a recipe formulated in natural language to a set of cooking actions that is precise enough to be executed in the simulated kitchen and yields the desired dish. To overcome the challenges inherent to recipe execution, this mapping process needs to incorporate reasoning over the recipe text, the state of the simulated kitchen environment, common-sense knowledge, knowledge of the cooking domain, and the action space of a virtual or robotic chef. This benchmark thereby addresses the growing interest in human-centric systems that combine natural language processing and situated reasoning to perform everyday activities.
Original languageEnglish
Title of host publicationProceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
EditorsNicoletta Calzolari, Min-Yen Kan, Veronique Hoste, Alessandro Lenci, Sakriani Sakti, Nianwen Xue
PublisherELRA and ICCL
Pages22-42
Number of pages21
ISBN (Electronic)9782493814104
Publication statusPublished - May 2024
EventLREC-COLING 2024 - Lingotto Conference Centre, Turin, Italy
Duration: 20 May 202425 May 2024
https://lrec-coling-2024.org

Publication series

Name2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, LREC-COLING 2024 - Main Conference Proceedings

Conference

ConferenceLREC-COLING 2024
Country/TerritoryItaly
CityTurin
Period20/05/2425/05/24
Internet address

Bibliographical note

Funding Information:
The research reported on in this paper received funding from the EU's H2020 RIA programme under grant agreement no. 951846 (MUHAI), the Research Foundation Flanders (FWO) through a postdoctoral grant awarded to PVE (grant no. 76929), and from the Collaborative Research Center (SFB) 1320 EASE - Everyday Activity Science and Engineering, University of Bremen (www.easecrc. org), sub-project P01 \u201CEmbodied Semantics for the Language of Action and Change\u201D.

Publisher Copyright:
© 2024 ELRA Language Resource Association: CC BY-NC 4.0.

Keywords

  • benchmark
  • recipe execution
  • natural language understanding
  • situated reasoning

Fingerprint

Dive into the research topics of 'A Benchmark for Recipe Understanding in Artificial Agents'. Together they form a unique fingerprint.

Cite this