5/11/2025: COCONut-Pancap-50K panoptic grounding captions are vailable at huggingface. For the dense masks subset, we are still working on the final inspection.
3/28/2025: COCONut-Pancap region30k is released for the interest of region-level instruction data. More tutorials are coming!
You can switch to download COCONut-S by adding “–split coconut_s” to the command.
python download_coconut.py --split coconut_s
The data will be saved at “./coconut_datasets” by default, you can change it to your preferred path by adding “–output_dir YOUR_DATA_PATH”.
To use COCONut-Large, you need to download the panoptic masks from huggingface and copy the images by the image list from the objects365 image folder. Then add them on top of COCONut-B, to consist the full COCONut-Large dataset.
[CVPR2024] 🥥COCONut: Crafting the Future of Segmentation Datasets with Exquisite Annotations in the Era of ✨Big Data✨
Xueqing Deng, Qihang Yu, Peng Wang, Xiaohui Shen, Liang-Chieh Chen
🚀 Contributions
🔥 1st large-scale human verified dataset for segmentation, more info can be found at our website.
🔥 COCONut is now available at Kaggle and huggingface, welcome to download!
📢 News!
TODO
Dataset Splits
Please refer to 🔗preparing datasets for exploring training and evaluation.
Get Started
We only provide the annotation, for those who are interested to use our annotation will need to download the images from the links: COCONut-S images, COCONut-B images and relabeled COCO-val images.
We provide two methods to download the dataset annotations, details are as below。
🔗Kaggle download link
You can use the web UI to download the dataset directly on Kaggle.
If you find our dataset useful, we really appreciate if you can upvote our dataset on Kaggle,
🔗Huggingface dataset preview
Directly download the data from huggingface or git clone the huggingface dataset repo will result in invalid data structure.
We recommend you to use our provided download script to download the dataset from huggingface.
You can switch to download COCONut-S by adding “–split coconut_s” to the command.
The data will be saved at “./coconut_datasets” by default, you can change it to your preferred path by adding “–output_dir YOUR_DATA_PATH”.
To use COCONut-Large, you need to download the panoptic masks from huggingface and copy the images by the image list from the objects365 image folder. Then add them on top of COCONut-B, to consist the full COCONut-Large dataset.
Tutorials
visualization on COCONut panoptic masks
panoptic segmentation
instance segmentation
semantic segmentation
open-vocabulary segmentation
object detection
FAQ
We summarize the common issues in FAQ.md, please check this out before you create any new issues.
More visualization on COCONut annotation
Terms of use
Acknowledgement
Bibtex
If you find our dataset useful, please cite: