Skip to content

Commit

Permalink
0.7.0
Browse files Browse the repository at this point in the history
  • Loading branch information
matatonic committed Apr 8, 2024
1 parent c203ab3 commit ac9782b
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions docker-compose.yml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ services:
#command: ["python", "vision.py", "-m", "internlm/internlm-xcomposer2-vl-7b", "--use-flash-attn", "-d", "cuda:0"]
#command: ["python", "vision.py", "-m", "internlm/internlm-xcomposer2-vl-7b-4bit", "--use-flash-attn"]
#command: ["python", "vision.py", "-m", "llava-hf/llava-v1.6-34b-hf", "--load-in-4bit", "--use-flash-attn"] # very, very slow...
command: ["python", "vision.py", "-m", "echo840/Monkey-Chat"]
#command: ["python", "vision.py", "-m", "echo840/Monkey-Chat"]
#command: ["python", "vision.py", "-m", "echo840/Monkey"]
#command: ["python", "vision.py", "-m", "openbmb/OmniLMM-12B"] # WIP
#command: ["python", "vision.py", "-m", "deepseek-ai/deepseek-vl-7b-chat"] # WIP
Expand All @@ -33,7 +33,7 @@ services:
#command: ["python", "vision.py", "-m", "llava-hf/llava-1.5-13b-hf", "--load-in-4bit", "--use-flash-attn"]
#command: ["python", "vision.py", "-m", "openbmb/MiniCPM-V", "--use-flash-attn", '-d', 'cuda:0']
#command: ["python", "vision.py", "-m", "llava-hf/llava-1.5-7b-hf", "--load-in-4bit", "--use-flash-attn"]
#command: ["python", "vision.py", "-m", "vikhyatk/moondream2", "--use-flash-attn"]
command: ["python", "vision.py", "-m", "vikhyatk/moondream2", "--use-flash-attn"]
#command: ["python", "vision.py", "-m", "vikhyatk/moondream1"] # broken
#command: ["python", "vision.py", "-m", "YanweiLi/Mini-Gemini-2B"] # be sure to review and run prepare_minigemini.sh before starting minigemini
runtime: nvidia
Expand Down

0 comments on commit ac9782b

Please sign in to comment.