You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+62-28Lines changed: 62 additions & 28 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -27,6 +27,11 @@ The vision features (detr, resnet, clip, vit) are available at https://huggingfa
27
27
Alternatively, you may download the extracted vision features (detr, resnet, clip) from [vision_features](https://drive.google.com/file/d/13B0hc_F_45-UlqPLKSgRz-ALtFQ8kIJr/view?usp=share_link) and unzip the files under `vision_features`
28
28
29
29
## Extract Features (optional)
30
+
31
+
The processed vision features for ScienceQA are available at https://huggingface.co/cooelf/vision_features/tree/main.
32
+
33
+
The following instructions show how we obtain those features.
34
+
30
35
Download the image files from [Google Drive](https://drive.google.com/drive/folders/1w8imCXWYn2LxajmGeGH_g5DaL2rabHev?usp=sharing) and unzip all the images (train, dev, test) in the same folder (). The structure should be:
31
36
32
37
```
@@ -43,54 +48,83 @@ images
43
48
│ └── image.png
44
49
```
45
50
46
-
Run ```extract_features.py --data_root images --output_dir vision_features --img_type detr```
51
+
Run ```extract_features.py --data_root images --output_dir vision_features --img_type vit```
47
52
48
53
If you hope to use your own images, please structure those images in the way above, or modify the script ```extract_features.py```.
49
54
55
+
## Extract Captions (optional)
56
+
57
+
The processed captions for ScienceQA are available at ```data/instruct_captions.json```.
58
+
59
+
The following instructions show how we obtain those features.
60
+
61
+
Intall lavis and prepare Vicuna weights to use InstructBLIP for caption extraction.
Our trained models are available at [models](https://drive.google.com/file/d/1FtTYOJPHnWnFfCxNC6M3gar4RAX5E21b/view?usp=share_link). To use our trained models, please put the them under the ```models``` folder.
102
+
Our trained models are available at https://huggingface.co/cooelf/mm-cot/tree/main. To use our trained models, please put the them under the ```models``` folder.
0 commit comments