diff --git a/README.md b/README.md index 2a7c2b386..0a89a45d0 100644 --- a/README.md +++ b/README.md @@ -158,6 +158,7 @@ At present, we have introduced several key features to showcase our current capa We offer extensive model support, including dozens of large language models (LLMs) from both open-source and API agents, such as LLaMA/LLaMA2, Baichuan, ChatGLM, Wenxin, Tongyi, Zhipu, and many more. - News + - πŸ”₯πŸ”₯πŸ”₯ [CodeQwen1.5-7B-Chat](https://huggingface.co/Qwen/CodeQwen1.5-7B-Chat) - πŸ”₯πŸ”₯πŸ”₯ [Qwen1.5-32B-Chat](https://huggingface.co/Qwen/Qwen1.5-32B-Chat) - πŸ”₯πŸ”₯πŸ”₯ [Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta) - πŸ”₯πŸ”₯πŸ”₯ [gemma-7b-it](https://huggingface.co/google/gemma-7b-it) diff --git a/README.zh.md b/README.zh.md index 360ad7903..162b8654c 100644 --- a/README.zh.md +++ b/README.zh.md @@ -152,6 +152,7 @@ ζ΅·ι‡ζ¨‘εž‹ζ”―ζŒοΌŒεŒ…ζ‹¬εΌ€ζΊγ€APIδ»£η†η­‰ε‡ εη§ε€§θ―­θ¨€ζ¨‘εž‹γ€‚ε¦‚LLaMA/LLaMA2、Baichuan、ChatGLMγ€ζ–‡εΏƒγ€ι€šδΉ‰γ€ζ™Ίθ°±η­‰γ€‚ε½“ε‰ε·²ζ”―ζŒε¦‚δΈ‹ζ¨‘εž‹: - ζ–°ε’žζ”―ζŒζ¨‘εž‹ + - πŸ”₯πŸ”₯πŸ”₯ [CodeQwen1.5-7B-Chat](https://huggingface.co/Qwen/CodeQwen1.5-7B-Chat) - πŸ”₯πŸ”₯πŸ”₯ [Qwen1.5-32B-Chat](https://huggingface.co/Qwen/Qwen1.5-32B-Chat) - πŸ”₯πŸ”₯πŸ”₯ [Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta) - πŸ”₯πŸ”₯πŸ”₯ [gemma-7b-it](https://huggingface.co/google/gemma-7b-it) diff --git a/dbgpt/configs/model_config.py b/dbgpt/configs/model_config.py index 3ff9a01a5..ed8d296f8 100644 --- a/dbgpt/configs/model_config.py +++ b/dbgpt/configs/model_config.py @@ -108,6 +108,8 @@ def get_device() -> str: # https://huggingface.co/Qwen/Qwen1.5-32B-Chat "qwen1.5-32b-chat": os.path.join(MODEL_PATH, "Qwen1.5-32B-Chat"), "qwen1.5-72b-chat": os.path.join(MODEL_PATH, "Qwen1.5-72B-Chat"), + # https://huggingface.co/Qwen/CodeQwen1.5-7B-Chat + "codeqwen1.5-7b-chat": os.path.join(MODEL_PATH, "CodeQwen1.5-7B-Chat"), # (Llama2 based) We only support WizardLM-13B-V1.2 for now, which is trained from Llama-2 13b, see https://huggingface.co/WizardLM/WizardLM-13B-V1.2 "wizardlm-13b": os.path.join(MODEL_PATH, "WizardLM-13B-V1.2"), # wget https://huggingface.co/TheBloke/vicuna-13B-v1.5-GGUF/resolve/main/vicuna-13b-v1.5.Q4_K_M.gguf -O models/ggml-model-q4_0.gguf