kilogram / README.md
anya-ji's picture
Link dataset to paper (#1)
ef45dfa verified
|
raw
history blame
No virus
2.33 kB
---
annotations_creators:
- crowdsourced
language:
- en
multilinguality:
- monolingual
pretty_name: KiloGram
size_categories:
- 1K<n<10K
source_datasets:
- original
tags:
- tangrams
- reference-games
- vision-language
viewer: false
---
Preprocessed training and evaluation data from KiloGram.
KiloGram dataset and code repo: https://github.com/lil-lab/kilogram
---
# File Formats
## Training Set
Texts: `train_*.json` are all in the format of `{tangramName: list(annotations)}`.
Images: Colored images with parts (under `/color`) are named in the format of `tangramName_{idx}.png`, where `idx` corresponds to the index of the annotation in the text file.
## Validation, Development, Heldout Set
Texts: `{whole, part}_{black, color}.json` are in the format of `{"targets": list(imageFileNames), "images": list(imageFileNames), "texts": list(annotations)}`. We flattened all the contexts and concatenated them into one list for each entry.
E.g. the first 10 elements in `"targets"` are the image file name of the target of the first context repeated 10 times; the first 10 of `"images"` are the image file names in that context; and the first 10 of `"texts"` are the corresponding 10 annotations in that context.
`/controlled` contains experiments with constrained contexts controlled for number of parts, and `/random` contains ones without. (See Appendix A.8 in paper)
`/development/texts/augmented/aug_dev.json` and `images/augmented.tar.bz2` are experiments in the same format as above used to evaluate the effect of adding part information.
Intermediate files:
`*/text/controlled/eval_batch_data.json` are in the format of
`{tangramName: {numOfParts: list({"target": [tangramName_{idx}, annotation], "distractors": list(list([tangramName_{idx}, annotation]))})}}`, used to generate controlled experiment jsons. Note: annotations are descriptions concatenated by "#" instead of in natural English.
# Citation
```bibtex
@misc{ji2022abstractvisualreasoningtangram,
title={Abstract Visual Reasoning with Tangram Shapes},
author={Anya Ji and Noriyuki Kojima and Noah Rush and Alane Suhr and Wai Keen Vong and Robert D. Hawkins and Yoav Artzi},
year={2022},
eprint={2211.16492},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2211.16492},
}
```