Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Continuous batching does not support video inputs for this model: MiniCPM-V-2.6 #2478

Open
1 of 3 tasks
jiaolongxue opened this issue Oct 24, 2024 · 6 comments
Open
1 of 3 tasks
Labels
Milestone

Comments

@jiaolongxue
Copy link

System Info / 系統信息

cuda12.1

Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?

  • docker / docker
  • pip install / 通过 pip install 安装
  • installation from source / 从源码安装

Version info / 版本信息

v0.16.0

The command used to start Xinference / 用以启动 xinference 的命令

docker run --name xinference-local -d -e XINFERENCE_MODEL_SRC=modelscope -e MODELSCOPE_CACHE=/data/modelscope/hub -e XINFERENCE_HOME=/data/inference/home/ -e VLLM_USE_MODELSCOPE=False -v /data:/data -p 9997:9997 --gpus all xinference:v0.16.3-cuda121 xinference-local -H 0.0.0.0 --log-level debug

Reproduction / 复现过程

上传视频,询问视频内容,提示错误

Traceback (most recent call last):
  File "/usr/local/lib/python3.10/dist-packages/gradio/queueing.py", line 527, in process_events
    response = await route_utils.call_process_api(
  File "/usr/local/lib/python3.10/dist-packages/gradio/route_utils.py", line 261, in call_process_api
    output = await app.get_blocks().process_api(
  File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1786, in process_api
    result = await self.call_function(
  File "/usr/local/lib/python3.10/dist-packages/gradio/blocks.py", line 1350, in call_function
    prediction = await utils.async_iteration(iterator)
  File "/usr/local/lib/python3.10/dist-packages/gradio/utils.py", line 583, in async_iteration
    return await iterator.__anext__()
  File "/usr/local/lib/python3.10/dist-packages/gradio/utils.py", line 576, in __anext__
    return await anyio.to_thread.run_sync(
  File "/usr/local/lib/python3.10/dist-packages/anyio/to_thread.py", line 56, in run_sync
    return await get_async_backend().run_sync_in_worker_thread(
  File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 2144, in run_sync_in_worker_thread
    return await future
  File "/usr/local/lib/python3.10/dist-packages/anyio/_backends/_asyncio.py", line 851, in run
    result = context.run(func, *args)
  File "/usr/local/lib/python3.10/dist-packages/gradio/utils.py", line 559, in run_sync_iterator_async
    return next(iterator)
  File "/usr/local/lib/python3.10/dist-packages/gradio/utils.py", line 742, in gen_wrapper
    response = next(iterator)
  File "/usr/local/lib/python3.10/dist-packages/xinference/core/chat_interface.py", line 221, in predict
    response = model.chat(
  File "/usr/local/lib/python3.10/dist-packages/xinference/client/restful/restful_client.py", line 523, in chat
    raise RuntimeError(
RuntimeError: Failed to generate chat completion, detail: [address=0.0.0.0:46675, pid=77] Continuous batching does not support video inputs for this model: MiniCPM-V-2.6

Expected behavior / 期待表现

能够正确理解视频内容

@XprobeBot XprobeBot added the gpu label Oct 24, 2024
@XprobeBot XprobeBot modified the milestones: v0.15, v0.16 Oct 24, 2024
Copy link

github-actions bot commented Nov 6, 2024

This issue is stale because it has been open for 7 days with no activity.

@github-actions github-actions bot added the stale label Nov 6, 2024
Copy link

This issue was closed because it has been inactive for 5 days since being marked as stale.

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Nov 12, 2024
@948024326
Copy link

我也报错了这个问题 请问有解决吗?

@qinxuye
Copy link
Contributor

qinxuye commented Nov 13, 2024

我们看下

@qinxuye qinxuye removed the stale label Nov 19, 2024
@qinxuye qinxuye reopened this Nov 19, 2024
@likenamehaojie
Copy link

同样的问题 ,不能对视频进行问答

@likenamehaojie
Copy link

@qinxuye

@XprobeBot XprobeBot modified the milestones: v0.16, v1.x Nov 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

5 participants