-
Notifications
You must be signed in to change notification settings - Fork 2.9k
model: Intern vl 2.5 #3351
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
model: Intern vl 2.5 #3351
Conversation
depends on #3203 |
6cf5f01
to
506dda1
Compare
|
||
all_frames = [] | ||
|
||
def load_image_internvl(image_file, input_size=448, max_num=12): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we need to introduce parameters for max/min numbers passed from the API, similar to what is done in lmdeploy?
dict(type='image_url', image_url=dict(max_dynamic_patch=12, url='https://raw.githubusercontent.com/OpenGVLab/InternVL/main/internvl_chat/examples/image1.jpg')),
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is that parameter frequently used? If not, and other vl models don't support that parameter, this might be of low-priority
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Perhaps the user wants to control the limit on the size of image slices per request?
self.vision_model = InternVisionModel( | ||
config=config.vision_config, quant_config=quant_config | ||
) | ||
self.language_model = InternLM2ForCausalLM( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Add support to qwen/llama/.. language_model.
Edit
the
This modification will work. It appears there are some discrepancies with the inference results from lmdeploy. SGLang is unable to generate Chinese output
Result
|
f4de116
to
a7734d7
Compare
This LGTM, but too large. I think decouple should be good. |
@mickqian Why we close this? Are there any follow up? |
Motivation
Support InternVL2_5, as requested in #3092
Modifications
Checklist