Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to input multiple images using VLM-R1 #135

Open
WuYeeh opened this issue Mar 8, 2025 · 3 comments
Open

How to input multiple images using VLM-R1 #135

WuYeeh opened this issue Mar 8, 2025 · 3 comments

Comments

@WuYeeh
Copy link

WuYeeh commented Mar 8, 2025

This is a very good job, but I seem to be in trouble. When I use VLM-R1 to train Qwen2.5VL for tasks which need input multiple images, the images seem cannot be input into the Trainer. LLaMA-Factory is easy to input multiple images. How should I modify the training code in VLM-R1? Looking forward to your reply, thank you.

@snakeztc
Copy link
Contributor

snakeztc commented Mar 9, 2025

We are working on supporting this feature. Stay tuned.

@gmyFighting
Copy link

In grpo_trainer.py, the 'image' variable is a list. You simply need to add all your images to this list.

@WuYeeh
Copy link
Author

WuYeeh commented Mar 10, 2025

@gmyFighting Thanks, I get it. But I want to know if I want to place my images in the designated position of the prompt (different images in different position), can I add an identifier at the desired location like LLaMA-Factory?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants