Data: CRAFT: A Benchmark for Causal Reasoning About Forces and inTeractions
| dc.contributor.author | Ates, Tayfun | |
| dc.contributor.author | Atesoglu, M. Samil | |
| dc.contributor.author | Yigit, Cagatay | |
| dc.contributor.author | Kesen, Ilker | |
| dc.contributor.author | Kobas, Mert | |
| dc.contributor.author | Erdem, Erkut | |
| dc.contributor.author | Erdem, Aykut | |
| dc.contributor.author | Goksun, Tilbe | |
| dc.contributor.author | Yuret, Deniz | |
| dc.date.accessioned | 2025-10-24T11:06:04Z | |
| dc.date.issued | 2021-06-07 | |
| dc.description.abstract | Humans are able to perceive, understand and reason about physical events. Developing models with similar physical understanding capabilities is a long-standing goal of artificial intelligence. As a step towards this goal, in this work, we introduce CRAFT, a new visual question answering dataset that requires causal reasoning about physical forces and object interactions. It contains 58K video and question pairs that are generated from 10K videos from 20 different virtual environments, containing various objects in motion that interact with each other and the scene. Two question categories from CRAFT include previously studied <em>descriptive</em> and <em>counterfactual</em> questions. Besides, inspired by the theories of force dynamics in cognitive linguistics, we introduce new question categories that involve understanding the interactions of objects through the notions of <em>cause</em>, <em>enable</em>, and <em>prevent</em>. Our results demonstrate that even though these tasks seem to be simple and intuitive for humans, the evaluated baseline models, including existing state-of-the-art methods, do not yet deal with the challenges posed in our benchmark dataset. | |
| dc.description.abstract | CRAFT was supported in part by GEBIP 2018 Award of the Turkish Academy of Sciences to E. Erdem and T. Goksun, BAGEP 2021 Award of the Science Academy to A. Erdem, and AI Fellowship to Ilker Kesen provided by the KUIS AI Lab. | |
| dc.description.uri | https://dx.doi.org/10.5281/zenodo.4904783 | |
| dc.description.uri | http://dx.doi.org/10.5281/zenodo.4904783 | |
| dc.description.uri | https://dx.doi.org/10.5281/zenodo.4904782 | |
| dc.identifier.doi | 10.5281/zenodo.4904783 | |
| dc.identifier.openaire | doi_dedup___::825f071267c21163365963913c6836d8 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.14288/31146 | |
| dc.language.iso | eng | |
| dc.publisher | Zenodo | |
| dc.rights | OPEN | |
| dc.subject | causal reasoning | |
| dc.subject | visual question answering | |
| dc.subject | temporal reasoning | |
| dc.subject | intuitive physics | |
| dc.subject | force dynamics | |
| dc.title | CRAFT: A Benchmark for Causal Reasoning About Forces and inTeractions | |
| dc.type | Dataset | |
| dspace.entity.type | Data | |
| local.import.source | OpenAire |
