Commit
·
c009e13
1
Parent(s):
f20b415
Updated ReadMe
Browse files
README.md
CHANGED
@@ -88,7 +88,31 @@ configs:
|
|
88 |
|
89 |
### Dataset Summary
|
90 |
|
91 |
-
This dataset is a consolidation of SpaRTUN and StepGame datasets with an extension of additional spatial characterization and reasoning path generation. The methodology is explained in our ACL 2024 paper - [SpaRC and SpaRP: Spatial Reasoning Characterization and Path Generation for Understanding Spatial Reasoning Capability of Large Language Models]().
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
92 |
|
93 |
### Languages
|
94 |
|
|
|
88 |
|
89 |
### Dataset Summary
|
90 |
|
91 |
+
This dataset is a consolidation of SpaRTUN and StepGame datasets with an extension of additional spatial characterization and reasoning path generation. The methodology is explained in our ACL 2024 paper - [SpaRC and SpaRP: Spatial Reasoning Characterization and Path Generation for Understanding Spatial Reasoning Capability of Large Language Models](). The dataset format and fields are normalized across the two upstream benchmark datasets -- SpaRTUN and StepGame. The datasets are primarily a Spatial Question Answering datasets, which are enriched with verbalized reasoning paths. The prominent fields of interests are:
|
92 |
+
|
93 |
+
- *context*: Textual description of the spatial context.
|
94 |
+
- *question*: A question about finding spatial relations between two entities in the context.
|
95 |
+
- *targets*: Answer i.e. list of spatial relations between the entities in the question.
|
96 |
+
- *target_choices*: List of all the spatial relations to choose from.
|
97 |
+
- *target_scores*: Binarized Multi-label representation of targets over target_choices.
|
98 |
+
- *reasoning*: Verbalized reasoning path as deductively-verified CoT for training or few-shot examples.
|
99 |
+
|
100 |
+
Additionally, the fields with the metadata are:
|
101 |
+
|
102 |
+
- *context_id*: An identifier from the source data corresponding to the context. `context_id` is unique. A single context can have multiple questions (e.g. SpaRTUN). Hence, (context_id, question_id) is a unique identifier for a dataset instance.
|
103 |
+
- *question_id*: An identifier from the source data corresponding to the question.
|
104 |
+
- *num_hop*: Ground truth number of hop required for the question.
|
105 |
+
- *symbolic_context*: A json string describing the symbolic context.
|
106 |
+
- *symbolic_entity_map*: A json string that maps symbolic entities to their complete descriptive names used.
|
107 |
+
- *symbolic_question*: A json string describing the symbolic question.
|
108 |
+
- *num_context_entities*: Number of entities in the context.
|
109 |
+
- *num_question_entities*: Number of entities in the question.
|
110 |
+
- *question_type*: Type of the question. Only `FR` i.e. Find Relation type questions are currently present.
|
111 |
+
- *canary*: A canary string present only in the `test`.
|
112 |
+
- *reasoning_types*: The type of reasoning copied from the source data required for answering the question.
|
113 |
+
- *spatial_types*: The type of spatial relations copied from the source data required for answering the question.
|
114 |
+
- *source_data*: The upstream source of the data (either SpaRTUN or StepGame) for a given instance.
|
115 |
+
- *comments*: Additional comments specific to the upstream data.
|
116 |
|
117 |
### Languages
|
118 |
|