-
Notifications
You must be signed in to change notification settings - Fork 1.1k
/
Copy pathhandler_map.py
220 lines (215 loc) · 10.8 KB
/
handler_map.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
from bfcl.model_handler.api_inference.claude import ClaudeHandler
from bfcl.model_handler.api_inference.cohere import CohereHandler
from bfcl.model_handler.api_inference.databricks import DatabricksHandler
from bfcl.model_handler.api_inference.deepseek import DeepSeekAPIHandler
from bfcl.model_handler.api_inference.fireworks import FireworksHandler
from bfcl.model_handler.api_inference.functionary import FunctionaryHandler
from bfcl.model_handler.api_inference.gemini import GeminiHandler
from bfcl.model_handler.api_inference.gogoagent import GoGoAgentHandler
from bfcl.model_handler.api_inference.gorilla import GorillaHandler
from bfcl.model_handler.api_inference.grok import GrokHandler
from bfcl.model_handler.api_inference.mistral import MistralHandler
from bfcl.model_handler.api_inference.nexus import NexusHandler
from bfcl.model_handler.api_inference.nova import NovaHandler
from bfcl.model_handler.api_inference.nvidia import NvidiaHandler
from bfcl.model_handler.api_inference.openai import OpenAIHandler
from bfcl.model_handler.api_inference.together import TogetherHandler
from bfcl.model_handler.api_inference.writer import WriterHandler
from bfcl.model_handler.api_inference.yi import YiHandler
from bfcl.model_handler.local_inference.bielik import BielikHandler
from bfcl.model_handler.local_inference.deepseek import DeepseekHandler
from bfcl.model_handler.local_inference.deepseek_coder import DeepseekCoderHandler
from bfcl.model_handler.local_inference.gemma import GemmaHandler
from bfcl.model_handler.local_inference.glaive import GlaiveHandler
from bfcl.model_handler.local_inference.glm import GLMHandler
from bfcl.model_handler.local_inference.granite import GraniteHandler
from bfcl.model_handler.local_inference.hammer import HammerHandler
from bfcl.model_handler.local_inference.hermes import HermesHandler
from bfcl.model_handler.local_inference.llama import LlamaHandler
from bfcl.model_handler.local_inference.llama_fc import LlamaFCHandler
from bfcl.model_handler.local_inference.minicpm import MiniCPMHandler
from bfcl.model_handler.local_inference.minicpm_fc import MiniCPMFCHandler
from bfcl.model_handler.local_inference.mistral_fc import MistralFCHandler
from bfcl.model_handler.local_inference.phi import PhiHandler
from bfcl.model_handler.local_inference.quick_testing_oss import QuickTestingOSSHandler
from bfcl.model_handler.local_inference.qwen import QwenHandler
from bfcl.model_handler.local_inference.salesforce import SalesforceHandler
from bfcl.model_handler.local_inference.falcon_fc import Falcon3FCHandler
# TODO: Add meta-llama/Llama-3.1-405B-Instruct
# Inference through API calls
api_inference_handler_map = {
"gorilla-openfunctions-v2": GorillaHandler,
"DeepSeek-V3": DeepSeekAPIHandler,
"o1-2024-12-17-FC": OpenAIHandler,
"o1-2024-12-17": OpenAIHandler,
"o3-mini-2025-01-31-FC": OpenAIHandler,
"o3-mini-2025-01-31": OpenAIHandler,
"gpt-4o-2024-11-20": OpenAIHandler,
"gpt-4o-2024-11-20-FC": OpenAIHandler,
"gpt-4o-mini-2024-07-18": OpenAIHandler,
"gpt-4o-mini-2024-07-18-FC": OpenAIHandler,
"gpt-4-turbo-2024-04-09": OpenAIHandler,
"gpt-4-turbo-2024-04-09-FC": OpenAIHandler,
"gpt-3.5-turbo-0125": OpenAIHandler,
"gpt-3.5-turbo-0125-FC": OpenAIHandler,
"claude-3-opus-20240229": ClaudeHandler,
"claude-3-opus-20240229-FC": ClaudeHandler,
"claude-3-5-sonnet-20241022": ClaudeHandler,
"claude-3-5-sonnet-20241022-FC": ClaudeHandler,
"claude-3-5-haiku-20241022": ClaudeHandler,
"claude-3-5-haiku-20241022-FC": ClaudeHandler,
"nova-pro-v1.0": NovaHandler,
"nova-lite-v1.0": NovaHandler,
"nova-micro-v1.0": NovaHandler,
"open-mistral-nemo-2407": MistralHandler,
"open-mistral-nemo-2407-FC": MistralHandler,
"open-mixtral-8x22b": MistralHandler,
"open-mixtral-8x22b-FC": MistralHandler,
"open-mixtral-8x7b": MistralHandler,
"mistral-large-2407": MistralHandler,
"mistral-large-2407-FC": MistralHandler,
"mistral-medium-2312": MistralHandler,
"mistral-small-2402": MistralHandler,
"mistral-small-2402-FC": MistralHandler,
"firefunction-v1-FC": FireworksHandler,
"firefunction-v2-FC": FireworksHandler,
"Nexusflow-Raven-v2": NexusHandler,
"gemini-2.0-flash-lite-preview-02-05-FC": GeminiHandler,
"gemini-2.0-flash-lite-preview-02-05": GeminiHandler,
"gemini-2.0-flash-001-FC": GeminiHandler,
"gemini-2.0-flash-001": GeminiHandler,
"gemini-2.0-pro-exp-02-05-FC": GeminiHandler,
"gemini-2.0-pro-exp-02-05": GeminiHandler,
"gemini-1.5-pro-002": GeminiHandler,
"gemini-1.5-pro-002-FC": GeminiHandler,
"gemini-1.5-pro-001": GeminiHandler,
"gemini-1.5-pro-001-FC": GeminiHandler,
"gemini-1.5-flash-002": GeminiHandler,
"gemini-1.5-flash-002-FC": GeminiHandler,
"gemini-1.5-flash-001": GeminiHandler,
"gemini-1.5-flash-001-FC": GeminiHandler,
"gemini-1.0-pro-002": GeminiHandler,
"gemini-1.0-pro-002-FC": GeminiHandler,
"meetkai/functionary-small-v3.1-FC": FunctionaryHandler,
"meetkai/functionary-medium-v3.1-FC": FunctionaryHandler,
"databricks-dbrx-instruct": DatabricksHandler,
"command-r-plus-FC": CohereHandler,
"command-r7b-12-2024-FC": CohereHandler,
"snowflake/arctic": NvidiaHandler,
"nvidia/nemotron-4-340b-instruct": NvidiaHandler,
"BitAgent/GoGoAgent": GoGoAgentHandler,
# "yi-large-fc": YiHandler, # Their API is under maintenance, and will not be back online in the near future
"palmyra-x-004": WriterHandler,
"grok-beta": GrokHandler,
"meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo": TogetherHandler,
"meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo": TogetherHandler,
"meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo": TogetherHandler,
"meta-llama/Llama-3.3-70B-Instruct-Turbo": TogetherHandler,
"mistralai/Mixtral-8x7B-Instruct-v0.1": TogetherHandler,
"mistralai/Mistral-7B-Instruct-v0.1": TogetherHandler,
"Qwen/Qwen2.5-7B-Instruct-Turbo": TogetherHandler,
"Qwen/Qwen2.5-72B-Instruct-Turbo": TogetherHandler,
"deepseek-ai/DeepSeek-V3": TogetherHandler,
}
# Inference through local hosting
local_inference_handler_map = {
"google/gemma-2-2b-it": GemmaHandler,
"google/gemma-2-9b-it": GemmaHandler,
"google/gemma-2-27b-it": GemmaHandler,
"meta-llama/Meta-Llama-3-8B-Instruct": LlamaHandler,
"meta-llama/Meta-Llama-3-70B-Instruct": LlamaHandler,
"meta-llama/Llama-3.1-8B-Instruct-FC": LlamaFCHandler,
"meta-llama/Llama-3.1-8B-Instruct": LlamaHandler,
"meta-llama/Llama-3.1-70B-Instruct-FC": LlamaFCHandler,
"meta-llama/Llama-3.1-70B-Instruct": LlamaHandler,
"meta-llama/Llama-3.2-1B-Instruct": LlamaHandler,
"meta-llama/Llama-3.2-3B-Instruct": LlamaHandler,
"meta-llama/Llama-3.3-70B-Instruct-FC": LlamaFCHandler,
"meta-llama/Llama-3.3-70B-Instruct": LlamaHandler,
"Salesforce/xLAM-1b-fc-r": SalesforceHandler,
"Salesforce/xLAM-7b-fc-r": SalesforceHandler,
"Salesforce/xLAM-7b-r": SalesforceHandler,
"Salesforce/xLAM-8x22b-r": SalesforceHandler,
"Salesforce/xLAM-8x7b-r": SalesforceHandler,
"mistralai/Ministral-8B-Instruct-2410": MistralFCHandler,
"microsoft/Phi-3-mini-4k-instruct": PhiHandler,
"microsoft/Phi-3-mini-128k-instruct": PhiHandler,
"microsoft/Phi-3-small-8k-instruct": PhiHandler,
"microsoft/Phi-3-small-128k-instruct": PhiHandler,
"microsoft/Phi-3-medium-4k-instruct": PhiHandler,
"microsoft/Phi-3-medium-128k-instruct": PhiHandler,
"microsoft/Phi-3.5-mini-instruct": PhiHandler,
"NousResearch/Hermes-2-Pro-Mistral-7B": HermesHandler,
"NousResearch/Hermes-2-Pro-Llama-3-8B": HermesHandler,
"NousResearch/Hermes-2-Theta-Llama-3-8B": HermesHandler,
"NousResearch/Hermes-2-Pro-Llama-3-70B": HermesHandler,
"NousResearch/Hermes-2-Theta-Llama-3-70B": HermesHandler,
"ibm-granite/granite-20b-functioncalling": GraniteHandler,
"MadeAgents/Hammer2.1-7b": HammerHandler,
"MadeAgents/Hammer2.1-3b": HammerHandler,
"MadeAgents/Hammer2.1-1.5b": HammerHandler,
"MadeAgents/Hammer2.1-0.5b": HammerHandler,
"THUDM/glm-4-9b-chat": GLMHandler,
"Qwen/Qwen2-1.5B-Instruct": QwenHandler,
"Qwen/Qwen2-7B-Instruct": QwenHandler,
"Qwen/Qwen2.5-0.5B-Instruct": QwenHandler,
"Qwen/Qwen2.5-1.5B-Instruct": QwenHandler,
"Qwen/Qwen2.5-3B-Instruct": QwenHandler,
"Qwen/Qwen2.5-7B-Instruct": QwenHandler,
"Qwen/Qwen2.5-14B-Instruct": QwenHandler,
"Qwen/Qwen2.5-32B-Instruct": QwenHandler,
"Qwen/Qwen2.5-72B-Instruct": QwenHandler,
"Team-ACE/ToolACE-8B": LlamaHandler,
"openbmb/MiniCPM3-4B": MiniCPMHandler,
"openbmb/MiniCPM3-4B-FC": MiniCPMFCHandler,
"watt-ai/watt-tool-8B": LlamaHandler,
"watt-ai/watt-tool-70B": LlamaHandler,
"deepseek-ai/DeepSeek-V2.5": DeepseekCoderHandler,
"deepseek-ai/DeepSeek-Coder-V2-Instruct-0724": DeepseekCoderHandler,
"deepseek-ai/DeepSeek-Coder-V2-Lite-Instruct": DeepseekCoderHandler,
"deepseek-ai/DeepSeek-V2-Chat-0628": DeepseekHandler,
"deepseek-ai/DeepSeek-V2-Lite-Chat": DeepseekHandler,
"ZJared/Haha-7B": QwenHandler,
"speakleash/Bielik-11B-v2.3-Instruct": BielikHandler,
"NovaSky-AI/Sky-T1-32B-Preview": QwenHandler,
"Qwen/QwQ-32B-Preview": QwenHandler,
"tiiuae/Falcon3-10B-Instruct-FC": Falcon3FCHandler,
"tiiuae/Falcon3-7B-Instruct-FC": Falcon3FCHandler,
"tiiuae/Falcon3-3B-Instruct-FC": Falcon3FCHandler,
"tiiuae/Falcon3-1B-Instruct-FC": Falcon3FCHandler,
"uiuc-convai/CALM-8B": LlamaHandler,
"uiuc-convai/CALM-70B": LlamaHandler,
"uiuc-convai/CALM-405B": LlamaHandler
}
# Deprecated/outdated models, no longer on the leaderboard
outdated_model_handler_map = {
# "gorilla-openfunctions-v0": GorillaHandler,
# "o1-preview-2024-09-12": OpenAIHandler,
# "o1-mini-2024-09-12": OpenAIHandler,
# "gpt-4o-2024-08-06": OpenAIHandler,
# "gpt-4o-2024-08-06-FC": OpenAIHandler,
# "gpt-4o-2024-05-13": OpenAIHandler,
# "gpt-4o-2024-05-13-FC": OpenAIHandler,
# "gpt-4-1106-preview-FC": OpenAIHandler,
# "gpt-4-1106-preview": OpenAIHandler,
# "gpt-4-0125-preview-FC": OpenAIHandler,
# "gpt-4-0125-preview": OpenAIHandler,
# "gpt-4-0613-FC": OpenAIHandler,
# "gpt-4-0613": OpenAIHandler,
# "claude-2.1": ClaudeHandler,
# "claude-instant-1.2": ClaudeHandler,
# "claude-3-sonnet-20240229": ClaudeHandler,
# "claude-3-sonnet-20240229-FC": ClaudeHandler,
# "claude-3-5-sonnet-20240620": ClaudeHandler,
# "claude-3-5-sonnet-20240620-FC": ClaudeHandler,
# "claude-3-haiku-20240307": ClaudeHandler,
# "claude-3-haiku-20240307-FC": ClaudeHandler,
# "gemini-1.0-pro-001": GeminiHandler,
# "gemini-1.0-pro-001-FC": GeminiHandler,
# "meetkai/functionary-small-v3.1-FC": FunctionaryHandler,
# "mistral-tiny-2312": MistralHandler,
# "glaiveai/glaive-function-calling-v1": GlaiveHandler,
# "google/gemma-7b-it": GemmaHandler,
# "deepseek-ai/deepseek-coder-6.7b-instruct": DeepseekHandler,
}
HANDLER_MAP = {**api_inference_handler_map, **local_inference_handler_map}