Skip to content

Latest commit

 

History

History
40 lines (35 loc) · 1.72 KB

Data.md

File metadata and controls

40 lines (35 loc) · 1.72 KB

Pretraining

Please download the 558K subset of the LAION-CC-SBU dataset with BLIP captions we use in the paper here. Put the downloaded data under the folder playground/data.

playground/
└── data
    └── pretrain
        ├── blip_laion_cc_sbu_558k.json
        ├── blip_laion_cc_sbu_558k_meta.json
        └── images

Instruction Tuning

Please download the annotation of the final mixture our instruction tuning data llava_v1_5_mix665k.json, and download the images from constituting datasets:

After downloading all of them, organize the data as follows in ./playground/data,

playground/
└── data
    ├── llava_v1_5_mix665k.json
    ├── coco
    │   ├── val2014
    │   └── train2017
    ├── gqa
    │   └── images
    ├── ocr_vqa
    │   └── images
    ├── textvqa
    │   └── train_images
    └── vg
        ├── VG_100K
        └── VG_100K_2