-
Notifications
You must be signed in to change notification settings - Fork 18
/
model_conf_tests.json
135 lines (135 loc) · 9.34 KB
/
model_conf_tests.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
[
["AIDC-AI/Ovis1.6-Llama3.2-3B", "-A", "flash_attention_2"],
["AIDC-AI/Ovis1.6-Gemma2-9B", "-A", "flash_attention_2"],
["AIDC-AI/Ovis1.6-Gemma2-27B", "-A", "flash_attention_2"],
["AIDC-AI/Ovis1.5-Gemma2-9B", "-A", "flash_attention_2"],
["AIDC-AI/Ovis1.5-Llama3-8B", "-A", "flash_attention_2"],
["BAAI/Aquila-VL-2B-llava-qwen", "-A", "flash_attention_2", "--load-in-4bit"],
["BAAI/Aquila-VL-2B-llava-qwen", "-A", "flash_attention_2"],
["BAAI/Bunny-v1_0-2B-zh", "--load-in-4bit"],
["BAAI/Bunny-v1_0-2B-zh"],
["BAAI/Bunny-v1_0-3B", "--load-in-4bit"],
["BAAI/Bunny-v1_0-3B"],
["BAAI/Bunny-v1_0-3B-zh"],
["BAAI/Bunny-v1_0-4B", "--load-in-4bit"],
["BAAI/Bunny-v1_0-4B"],
["BAAI/Bunny-v1_1-4B", "--load-in-4bit"],
["BAAI/Bunny-v1_1-4B"],
["BAAI/Emu2-Chat", "--load-in-4bit"],
["BAAI/Emu2-Chat", "--max-memory=0:78GiB,1:20GiB"],
["BAAI/Emu3-Chat", "--load-in-4bit", "-A", "flash_attention_2"],
["BAAI/Emu3-Chat", "-A", "flash_attention_2"],
["HuggingFaceTB/SmolVLM-Instruct", "-A", "flash_attention_2", "--load-in-4bit"],
["HuggingFaceTB/SmolVLM-Instruct", "-A", "flash_attention_2"],
["OpenGVLab/InternVL-Chat-V1-5", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL-Chat-V1-5", "--device-map", "cuda:0", "--max-tiles", "40", "--load-in-4bit"],
["OpenGVLab/InternVL-Chat-V1-5", "--device-map", "cuda:0", "--max-tiles", "40"],
["OpenGVLab/InternVL-Chat-V1-5", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2_5-1B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2_5-1B", "--device-map", "cuda:0", "--max-tiles", "12"],
["OpenGVLab/InternVL2_5-2B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2_5-2B", "--device-map", "cuda:0", "--max-tiles", "12"],
["OpenGVLab/InternVL2_5-4B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2_5-4B", "--device-map", "cuda:0", "--max-tiles", "12"],
["OpenGVLab/InternVL2_5-8B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2_5-8B", "--device-map", "cuda:0", "--max-tiles", "12"],
["OpenGVLab/InternVL2_5-26B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2_5-26B", "--device-map", "cuda:0", "--max-tiles", "12"],
["OpenGVLab/InternVL2_5-38B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2_5-38B", "--device-map", "cuda:0", "--max-tiles", "12"],
["OpenGVLab/InternVL2_5-78B", "--device-map", "cuda:0", "--max-tiles", "12", "--load-in-4bit"],
["OpenGVLab/InternVL2-1B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL2-1B", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2-2B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL2-2B", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2-8B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL2-8B", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2-26B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL2-26B", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2-40B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL2-40B", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2-Llama3-76B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/Mini-InternVL-Chat-2B-V1-5", "--load-in-4bit"],
["OpenGVLab/Mini-InternVL-Chat-2B-V1-5", "--max-tiles", "40", "--load-in-4bit"],
["OpenGVLab/Mini-InternVL-Chat-2B-V1-5", "--max-tiles", "40"],
["OpenGVLab/Mini-InternVL-Chat-2B-V1-5"],
["Qwen/Qwen-VL-Chat", "--load-in-4bit"],
["Qwen/Qwen-VL-Chat"],
["Qwen/Qwen2-VL-2B-Instruct-AWQ", "-A", "flash_attention_2"],
["Qwen/Qwen2-VL-2B-Instruct", "-A", "flash_attention_2"],
["Qwen/Qwen2-VL-7B-Instruct-AWQ", "-A", "flash_attention_2"],
["Qwen/Qwen2-VL-7B-Instruct", "-A", "flash_attention_2"],
["Qwen/Qwen2-VL-72B-Instruct-AWQ", "-A", "flash_attention_2"],
["Salesforce/xgen-mm-phi3-mini-instruct-dpo-r-v1.5"],
["Salesforce/xgen-mm-phi3-mini-instruct-interleave-r-v1.5"],
["Salesforce/xgen-mm-phi3-mini-instruct-singleimg-r-v1.5"],
["Salesforce/xgen-mm-phi3-mini-instruct-r-v1"],
["TIGER-Lab/Mantis-8B-Fuyu", "--device-map", "cuda:0", "--load-in-4bit"],
["TIGER-Lab/Mantis-8B-Fuyu", "--device-map", "cuda:0"],
["adept/fuyu-8b", "--device-map", "cuda:0", "--load-in-4bit"],
["adept/fuyu-8b", "--device-map", "cuda:0"],
["allenai/MolmoE-1B-0924", "-A", "flash_attention_2", "--load-in-4bit", "--use-double-quant"],
["allenai/MolmoE-1B-0924", "-A", "flash_attention_2", "--load-in-4bit"],
["allenai/MolmoE-1B-0924", "-A", "flash_attention_2"],
["allenai/Molmo-7B-D-0924", "-A", "flash_attention_2", "--load-in-4bit", "--use-double-quant"],
["allenai/Molmo-7B-D-0924", "-A", "flash_attention_2", "--load-in-4bit"],
["allenai/Molmo-7B-D-0924", "-A", "flash_attention_2"],
["allenai/Molmo-7B-O-0924", "-A", "flash_attention_2", "--load-in-4bit", "--use-double-quant"],
["allenai/Molmo-7B-O-0924", "-A", "flash_attention_2", "--load-in-4bit"],
["allenai/Molmo-7B-O-0924", "-A", "flash_attention_2"],
["allenai/Molmo-72B-0924", "-A", "flash_attention_2", "--load-in-4bit", "--use-double-quant"],
["allenai/Molmo-72B-0924", "-A", "flash_attention_2", "--load-in-4bit"],
["echo840/Monkey-Chat", "--load-in-4bit"],
["echo840/Monkey-Chat"],
["failspy/Phi-3-vision-128k-instruct-abliterated-alpha", "-A", "flash_attention_2", "--load-in-4bit"],
["failspy/Phi-3-vision-128k-instruct-abliterated-alpha", "-A", "flash_attention_2"],
["fancyfeast/joy-caption-alpha-two", "--load-in-4bit", "-A", "flash_attention_2"],
["fancyfeast/joy-caption-alpha-two", "-A", "flash_attention_2"],
["fancyfeast/joy-caption-pre-alpha", "--load-in-4bit", "-A", "flash_attention_2"],
["fancyfeast/joy-caption-pre-alpha", "-A", "flash_attention_2"],
["google/paligemma2-3b-ft-docci-448", "-A", "flash_attention_2"],
["google/paligemma2-10b-ft-docci-448", "-A", "flash_attention_2"],
["internlm/internlm-xcomposer2d5-7b", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["internlm/internlm-xcomposer2d5-7b", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["internlm/internlm-xcomposer2-4khd-7b", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["llava-hf/llava-1.5-13b-hf", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["llava-hf/llava-1.5-13b-hf", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["llava-hf/llava-1.5-7b-hf", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["llava-hf/llava-1.5-7b-hf", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["llava-hf/llava-v1.6-34b-hf", "-A", "flash_attention_2", "--load-in-4bit"],
["llava-hf/llava-v1.6-34b-hf", "-A", "flash_attention_2"],
["llava-hf/llava-v1.6-vicuna-13b-hf", "-A", "flash_attention_2", "--load-in-4bit"],
["llava-hf/llava-v1.6-vicuna-13b-hf", "-A", "flash_attention_2"],
["llava-hf/llava-v1.6-vicuna-7b-hf", "-A", "flash_attention_2", "--load-in-4bit"],
["llava-hf/llava-v1.6-vicuna-7b-hf", "-A", "flash_attention_2"],
["lmms-lab/llava-onevision-qwen2-0.5b-ov", "-A", "flash_attention_2"],
["lmms-lab/llava-onevision-qwen2-7b-ov", "-A", "flash_attention_2"],
["meta-llama/Llama-3.2-11B-Vision-Instruct", "-A", "flash_attention_2", "--load-in-4bit"],
["meta-llama/Llama-3.2-11B-Vision-Instruct", "-A", "flash_attention_2"],
["meta-llama/Llama-3.2-90B-Vision-Instruct", "-A", "flash_attention_2", "--load-in-4bit"],
["microsoft/Florence-2-base-ft", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["microsoft/Florence-2-base-ft", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["microsoft/Florence-2-large-ft", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["microsoft/Florence-2-large-ft", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["microsoft/Phi-3-vision-128k-instruct", "-A", "flash_attention_2", "--load-in-4bit"],
["microsoft/Phi-3-vision-128k-instruct", "-A", "flash_attention_2"],
["microsoft/Phi-3.5-vision-instruct", "-A", "flash_attention_2", "--load-in-4bit"],
["microsoft/Phi-3.5-vision-instruct", "-A", "flash_attention_2"],
["mistralai/Pixtral-12B-2409"],
["mx262/MiniMonkey", "-A", "flash_attention_2", "--load-in-4bit"],
["mx262/MiniMonkey", "-A", "flash_attention_2"],
["openbmb/MiniCPM-V-2_6-int4", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["openbmb/MiniCPM-V-2_6", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["openbmb/MiniCPM-V-2_6", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["openbmb/MiniCPM-Llama3-V-2_5", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["openbmb/MiniCPM-Llama3-V-2_5", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["qnguyen3/nanoLLaVA", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["qnguyen3/nanoLLaVA", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["qnguyen3/nanoLLaVA-1.5", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["qnguyen3/nanoLLaVA-1.5", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["qresearch/llama-3-vision-alpha-hf", "--device", "cuda:0", "--load-in-4bit"],
["qresearch/llama-3-vision-alpha-hf", "--device", "cuda:0"],
["rhymes-ai/Aria", "-A", "flash_attention_2"],
["vikhyatk/moondream2", "-A", "flash_attention_2", "--load-in-4bit"],
["vikhyatk/moondream2", "-A", "flash_attention_2"]
]