Evaluating Pre-trained Models on Task Datasets
LAVIS provides pre-trained and finetuned model for off-the-shelf evaluation on task dataset. Let’s now see an example to evaluate BLIP model on the captioning task, using MSCOCO dataset.
Preparing Datasets
First, let’s download the dataset. LAVIS provides automatic downloading scripts to help prepare most of the public dataset, to download MSCOCO dataset, simply run
cd lavis/datasets/download_scripts && python download_coco.py
This will put the downloaded dataset at a default cache location cache
used by LAVIS.
If you want to use a different cache location, you can specify it by updating cache_root
in lavis/configs/default.yaml
.
If you have a local copy of the dataset, it is recommended to create a symlink from the cache location to the local copy, e.g.
ln -s /path/to/local/coco cache/coco
Evaluating pre-trained models
To evaluate pre-trained model, simply run
bash run_scripts/blip/eval/eval_coco_cap.sh
Or to evaluate a large model:
bash run_scripts/blip/eval/eval_coco_cap_large.sh