Skip to content

internvl inference: how set max_num_frames value? #137

@Bendywater

Description

@Bendywater

When I inference video based internvl model, I find the frames of video is not transformer in vllm and only there are 32 frames. Therefore, I want to know how to set num_frames for video. I try this server: --media-io-kwargs '{"video": {"num_frames": 60}}' but I don't konw if it is right.

env CUDA_VISIBLE_DEVICES=2 vllm serve /mnt/models/InternVL3_5-4B \
    -dp 1 \
    --gpu-memory-utilization 0.8 \
    --max-model-len 40K \
    --served-model-name InternVL3_5-4B \
    --media-io-kwargs '{"video": {"num_frames": 60}}' \
    --port 12400 \
    --trust-remote-code
``

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions