Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat(model): Support Mixtral-8x7B #959

Merged
merged 4 commits into from
Dec 21, 2023
Merged

Conversation

fangyinc
Copy link
Collaborator

Closes #956

Other

  • Refactor model adapter

@github-actions github-actions bot added the enhancement New feature or request label Dec 21, 2023
@fangyinc fangyinc changed the title feat: Support Mixtral-8x7B feat(model): Support Mixtral-8x7B Dec 21, 2023
@github-actions github-actions bot added the model Module: model label Dec 21, 2023
@csunny
Copy link
Collaborator

csunny commented Dec 21, 2023

The idea of abstracting an adapter into a standalone module to manage the adaptation of models under different architectures and different inference frameworks is brilliant. I will take a look at the related code later to continuously improve.

@fangyinc
Copy link
Collaborator Author

The idea of abstracting an adapter into a standalone module to manage the adaptation of models under different architectures and different inference frameworks is brilliant. I will take a look at the related code later to continuously improve.

Yes, we can also import only the required dependencies.

@fangyinc
Copy link
Collaborator Author

With the introduction of chat_templating in huggingface, most model adaptations will be simpler, and we do not have to rely on FastChat.

Copy link
Collaborator

@csunny csunny left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Copy link
Collaborator

@Aries-ckt Aries-ckt left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

r+

@Aries-ckt Aries-ckt merged commit 6b982e2 into eosphoros-ai:main Dec 21, 2023
6 checks passed
vshy108 pushed a commit to vshy108/DB-GPT that referenced this pull request Jan 18, 2024
vshy108 pushed a commit to vshy108/DB-GPT that referenced this pull request Feb 6, 2024
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707199703 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198697 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198521 +0800

Add requirements.txt

Create only necesasary tables

Remove reference info in chat completion result
Set disable_alembic_upgrade to True
Comment _initialize_awel
Comment mount_static_files
Fix torch.has_mps deprecated

Add API key

Comment unused API endpoints

Install rocksdict to enable DiskCacheStorage

Fix the chat_knowledge missing in chat_mode

Update requirements.txt

Re-enable awel and add api key check for simple_rag_example DAG

Merge main bdf9442

Disable disable_alembic_upgrade

Compile bitsandbytes from source and enable verbose

Tune the prompt of chat knowledge to only refer to context

Add the web static files and uncomment previous unused APIs

Add back routers

Enable KNOWLEDGE_CHAT_SHOW_RELATIONS

Display relation based on CFG.KNOWLEDGE_CHAT_SHOW_RELATIONS

Stop reference add to last_output if KNOWLEDGE_CHAT_SHOW_RELATIONS is false

Fix always no reference

Improve chinese prompts

Update requirements.txt

Improve prompt

Improve prompt

Fix prompt variable name

Use openhermes-2.5-mistral-7b.Q4_K_M.gguf

1. Fix the delete issue of LlamaCppModel
2. Disable verbose log
3. Update diskcache
4. Remove conda-pack

Update chinese prompt and process the model response

Extract result from varying tags

Add back missing content_matches and put tags regex into variable

Update english prompt and decide CANNOT_ANSWER based on language configuration

Add 3 new models entries and upgrade bitsandbytes

Add few chat templates

Update model conversation with fastchat code

Revert "Update model conversation with fastchat code"

This reverts commit a5dc4b5.

Revert "Add few chat templates"

This reverts commit e6b6c99.

Add OpenHermes-2.5-Mistral-7B chat template

Fix missing messages and offset in chat template

Update fschat

Remove model adapter debugging logs and added conversation template

Update chinese chat knowledge prompt

Avoid to save the long chat history messages

Update chinese chat knowledge prompt

Temporary workaround to make the GGUF file use different chat template

Use ADD_COLON_SINGLE instead of FALCON_CHAT for separator style

Allow no model_name in chat completion request

Use starling-lm-7b-alpha.Q5_K_M.gguf

Add empty string as system for openchat_3.5 chat template

Undo response regex in generate_streaming

refactor: Refactor storage and new serve template (eosphoros-ai#947)

feat(core): Add API authentication for serve template (eosphoros-ai#950)

ci: Add python unit test workflows (eosphoros-ai#954)

feat(model): Support Mixtral-8x7B (eosphoros-ai#959)

feat(core): Support multi round conversation operator (eosphoros-ai#986)

chore(build): Fix typo and new pre-commit config (eosphoros-ai#987)

feat(model): Support SOLAR-10.7B-Instruct-v1.0 (eosphoros-ai#1001)

refactor: RAG Refactor (eosphoros-ai#985)

Co-authored-by: Aralhi <[email protected]>
Co-authored-by: csunny <[email protected]>

Upgrade english prompt for chat knowledge
vshy108 pushed a commit to vshy108/DB-GPT that referenced this pull request Feb 13, 2024
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707199703 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198697 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198521 +0800

Add requirements.txt

Create only necesasary tables

Remove reference info in chat completion result
Set disable_alembic_upgrade to True
Comment _initialize_awel
Comment mount_static_files
Fix torch.has_mps deprecated

Add API key

Comment unused API endpoints

Install rocksdict to enable DiskCacheStorage

Fix the chat_knowledge missing in chat_mode

Update requirements.txt

Re-enable awel and add api key check for simple_rag_example DAG

Merge main bdf9442

Disable disable_alembic_upgrade

Compile bitsandbytes from source and enable verbose

Tune the prompt of chat knowledge to only refer to context

Add the web static files and uncomment previous unused APIs

Add back routers

Enable KNOWLEDGE_CHAT_SHOW_RELATIONS

Display relation based on CFG.KNOWLEDGE_CHAT_SHOW_RELATIONS

Stop reference add to last_output if KNOWLEDGE_CHAT_SHOW_RELATIONS is false

Fix always no reference

Improve chinese prompts

Update requirements.txt

Improve prompt

Improve prompt

Fix prompt variable name

Use openhermes-2.5-mistral-7b.Q4_K_M.gguf

1. Fix the delete issue of LlamaCppModel
2. Disable verbose log
3. Update diskcache
4. Remove conda-pack

Update chinese prompt and process the model response

Extract result from varying tags

Add back missing content_matches and put tags regex into variable

Update english prompt and decide CANNOT_ANSWER based on language configuration

Add 3 new models entries and upgrade bitsandbytes

Add few chat templates

Update model conversation with fastchat code

Revert "Update model conversation with fastchat code"

This reverts commit a5dc4b5.

Revert "Add few chat templates"

This reverts commit e6b6c99.

Add OpenHermes-2.5-Mistral-7B chat template

Fix missing messages and offset in chat template

Update fschat

Remove model adapter debugging logs and added conversation template

Update chinese chat knowledge prompt

Avoid to save the long chat history messages

Update chinese chat knowledge prompt

Temporary workaround to make the GGUF file use different chat template

Use ADD_COLON_SINGLE instead of FALCON_CHAT for separator style

Allow no model_name in chat completion request

Use starling-lm-7b-alpha.Q5_K_M.gguf

Add empty string as system for openchat_3.5 chat template

Undo response regex in generate_streaming

refactor: Refactor storage and new serve template (eosphoros-ai#947)

feat(core): Add API authentication for serve template (eosphoros-ai#950)

ci: Add python unit test workflows (eosphoros-ai#954)

feat(model): Support Mixtral-8x7B (eosphoros-ai#959)

feat(core): Support multi round conversation operator (eosphoros-ai#986)

chore(build): Fix typo and new pre-commit config (eosphoros-ai#987)

feat(model): Support SOLAR-10.7B-Instruct-v1.0 (eosphoros-ai#1001)

refactor: RAG Refactor (eosphoros-ai#985)

Co-authored-by: Aralhi <[email protected]>
Co-authored-by: csunny <[email protected]>

Upgrade english prompt for chat knowledge
vshy108 pushed a commit to vshy108/DB-GPT that referenced this pull request Feb 13, 2024
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707199703 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198697 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198521 +0800

Add requirements.txt

Create only necesasary tables

Remove reference info in chat completion result
Set disable_alembic_upgrade to True
Comment _initialize_awel
Comment mount_static_files
Fix torch.has_mps deprecated

Add API key

Comment unused API endpoints

Install rocksdict to enable DiskCacheStorage

Fix the chat_knowledge missing in chat_mode

Update requirements.txt

Re-enable awel and add api key check for simple_rag_example DAG

Merge main bdf9442

Disable disable_alembic_upgrade

Compile bitsandbytes from source and enable verbose

Tune the prompt of chat knowledge to only refer to context

Add the web static files and uncomment previous unused APIs

Add back routers

Enable KNOWLEDGE_CHAT_SHOW_RELATIONS

Display relation based on CFG.KNOWLEDGE_CHAT_SHOW_RELATIONS

Stop reference add to last_output if KNOWLEDGE_CHAT_SHOW_RELATIONS is false

Fix always no reference

Improve chinese prompts

Update requirements.txt

Improve prompt

Improve prompt

Fix prompt variable name

Use openhermes-2.5-mistral-7b.Q4_K_M.gguf

1. Fix the delete issue of LlamaCppModel
2. Disable verbose log
3. Update diskcache
4. Remove conda-pack

Update chinese prompt and process the model response

Extract result from varying tags

Add back missing content_matches and put tags regex into variable

Update english prompt and decide CANNOT_ANSWER based on language configuration

Add 3 new models entries and upgrade bitsandbytes

Add few chat templates

Update model conversation with fastchat code

Revert "Update model conversation with fastchat code"

This reverts commit a5dc4b5.

Revert "Add few chat templates"

This reverts commit e6b6c99.

Add OpenHermes-2.5-Mistral-7B chat template

Fix missing messages and offset in chat template

Update fschat

Remove model adapter debugging logs and added conversation template

Update chinese chat knowledge prompt

Avoid to save the long chat history messages

Update chinese chat knowledge prompt

Temporary workaround to make the GGUF file use different chat template

Use ADD_COLON_SINGLE instead of FALCON_CHAT for separator style

Allow no model_name in chat completion request

Use starling-lm-7b-alpha.Q5_K_M.gguf

Add empty string as system for openchat_3.5 chat template

Undo response regex in generate_streaming

refactor: Refactor storage and new serve template (eosphoros-ai#947)

feat(core): Add API authentication for serve template (eosphoros-ai#950)

ci: Add python unit test workflows (eosphoros-ai#954)

feat(model): Support Mixtral-8x7B (eosphoros-ai#959)

feat(core): Support multi round conversation operator (eosphoros-ai#986)

chore(build): Fix typo and new pre-commit config (eosphoros-ai#987)

feat(model): Support SOLAR-10.7B-Instruct-v1.0 (eosphoros-ai#1001)

refactor: RAG Refactor (eosphoros-ai#985)

Co-authored-by: Aralhi <[email protected]>
Co-authored-by: csunny <[email protected]>

Upgrade english prompt for chat knowledge
vshy108 pushed a commit to vshy108/DB-GPT that referenced this pull request Feb 13, 2024
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707199703 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198697 +0800

parent 3f70da4
author penghou.ho <[email protected]> 1701341533 +0800
committer penghou.ho <[email protected]> 1707198521 +0800

Add requirements.txt

Create only necesasary tables

Remove reference info in chat completion result
Set disable_alembic_upgrade to True
Comment _initialize_awel
Comment mount_static_files
Fix torch.has_mps deprecated

Add API key

Comment unused API endpoints

Install rocksdict to enable DiskCacheStorage

Fix the chat_knowledge missing in chat_mode

Update requirements.txt

Re-enable awel and add api key check for simple_rag_example DAG

Merge main bdf9442

Disable disable_alembic_upgrade

Compile bitsandbytes from source and enable verbose

Tune the prompt of chat knowledge to only refer to context

Add the web static files and uncomment previous unused APIs

Add back routers

Enable KNOWLEDGE_CHAT_SHOW_RELATIONS

Display relation based on CFG.KNOWLEDGE_CHAT_SHOW_RELATIONS

Stop reference add to last_output if KNOWLEDGE_CHAT_SHOW_RELATIONS is false

Fix always no reference

Improve chinese prompts

Update requirements.txt

Improve prompt

Improve prompt

Fix prompt variable name

Use openhermes-2.5-mistral-7b.Q4_K_M.gguf

1. Fix the delete issue of LlamaCppModel
2. Disable verbose log
3. Update diskcache
4. Remove conda-pack

Update chinese prompt and process the model response

Extract result from varying tags

Add back missing content_matches and put tags regex into variable

Update english prompt and decide CANNOT_ANSWER based on language configuration

Add 3 new models entries and upgrade bitsandbytes

Add few chat templates

Update model conversation with fastchat code

Revert "Update model conversation with fastchat code"

This reverts commit a5dc4b5.

Revert "Add few chat templates"

This reverts commit e6b6c99.

Add OpenHermes-2.5-Mistral-7B chat template

Fix missing messages and offset in chat template

Update fschat

Remove model adapter debugging logs and added conversation template

Update chinese chat knowledge prompt

Avoid to save the long chat history messages

Update chinese chat knowledge prompt

Temporary workaround to make the GGUF file use different chat template

Use ADD_COLON_SINGLE instead of FALCON_CHAT for separator style

Allow no model_name in chat completion request

Use starling-lm-7b-alpha.Q5_K_M.gguf

Add empty string as system for openchat_3.5 chat template

Undo response regex in generate_streaming

refactor: Refactor storage and new serve template (eosphoros-ai#947)

feat(core): Add API authentication for serve template (eosphoros-ai#950)

ci: Add python unit test workflows (eosphoros-ai#954)

feat(model): Support Mixtral-8x7B (eosphoros-ai#959)

feat(core): Support multi round conversation operator (eosphoros-ai#986)

chore(build): Fix typo and new pre-commit config (eosphoros-ai#987)

feat(model): Support SOLAR-10.7B-Instruct-v1.0 (eosphoros-ai#1001)

refactor: RAG Refactor (eosphoros-ai#985)

Co-authored-by: Aralhi <[email protected]>
Co-authored-by: csunny <[email protected]>

Upgrade english prompt for chat knowledge
Hopshine pushed a commit to Hopshine/DB-GPT that referenced this pull request Sep 10, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request model Module: model
Projects
None yet
Development

Successfully merging this pull request may close these issues.

[Feature][model] Support Mixtral-8x7B-Instruct model
3 participants