-
Notifications
You must be signed in to change notification settings - Fork 18
/
model_conf_tests.alt.json
46 lines (46 loc) · 3.11 KB
/
model_conf_tests.alt.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
[
["omlab/omchat-v2.0-13B-single-beta_hf", "-A", "flash_attention_2"],
["BAAI/Bunny-Llama-3-8B-V", "--load-in-4bit"],
["BAAI/Bunny-Llama-3-8B-V"],
["BAAI/Bunny-v1_1-Llama-3-8B-V", "--load-in-4bit"],
["BAAI/Bunny-v1_1-Llama-3-8B-V"],
["HuggingFaceM4/idefics2-8b", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["HuggingFaceM4/idefics2-8b", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["HuggingFaceM4/idefics2-8b-AWQ", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["HuggingFaceM4/idefics2-8b-chatty", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["HuggingFaceM4/idefics2-8b-chatty", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["HuggingFaceM4/idefics2-8b-chatty-AWQ", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["OpenGVLab/InternVL2-4B", "--device-map", "cuda:0", "--load-in-4bit"],
["OpenGVLab/InternVL2-4B", "--device-map", "cuda:0"],
["OpenGVLab/Mini-InternVL-Chat-4B-V1-5", "--max-tiles", "40", "--load-in-4bit"],
["OpenGVLab/Mini-InternVL-Chat-4B-V1-5", "--load-in-4bit"],
["OpenGVLab/Mini-InternVL-Chat-4B-V1-5"],
["THUDM/cogagent-chat-hf", "--load-in-4bit"],
["THUDM/cogagent-chat-hf"],
["THUDM/cogvlm-chat-hf", "--load-in-4bit"],
["THUDM/cogvlm-chat-hf"],
["THUDM/cogvlm2-llama3-chat-19B", "--load-in-4bit"],
["THUDM/cogvlm2-llama3-chat-19B"],
["THUDM/cogvlm2-llama3-chinese-chat-19B", "--load-in-4bit"],
["THUDM/cogvlm2-llama3-chinese-chat-19B"],
["THUDM/glm-4v-9b", "--device-map", "cuda:0", "--load-in-4bit"],
["THUDM/glm-4v-9b", "--device-map", "cuda:0"],
["TIGER-Lab/Mantis-8B-clip-llama3", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["TIGER-Lab/Mantis-8B-clip-llama3", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["TIGER-Lab/Mantis-8B-siglip-llama3", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["TIGER-Lab/Mantis-8B-siglip-llama3", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["cognitivecomputations/dolphin-vision-72b", "-A", "flash_attention_2", "--load-in-4bit", "--device-map", "cuda:0"],
["cognitivecomputations/dolphin-vision-7b", "-A", "flash_attention_2", "--load-in-4bit", "--device-map", "cuda:0"],
["cognitivecomputations/dolphin-vision-7b", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["llava-hf/llava-v1.6-mistral-7b-hf", "-A", "flash_attention_2", "--load-in-4bit"],
["llava-hf/llava-v1.6-mistral-7b-hf", "-A", "flash_attention_2"],
["omlab/omchat-v2.0-13B-single-beta_hf", "-A", "flash_attention_2"],
["openbmb/MiniCPM-V", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["openbmb/MiniCPM-V", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["openbmb/MiniCPM-V-2", "-A", "flash_attention_2", "--device-map", "cuda:0", "--load-in-4bit"],
["openbmb/MiniCPM-V-2", "-A", "flash_attention_2", "--device-map", "cuda:0"],
["qihoo360/360VL-8B", "-A", "flash_attention_2", "--load-in-4bit"],
["qihoo360/360VL-8B", "-A", "flash_attention_2"],
["tiiuae/falcon-11B-vlm", "-A", "flash_attention_2", "--load-in-4bit"],
["tiiuae/falcon-11B-vlm", "-A", "flash_attention_2"]
]