generated from allenai/python-package-template
-
Notifications
You must be signed in to change notification settings - Fork 1.1k
Open
Description
🚀 The feature, motivation and pitch
I have seen that many users have constrained hardware choices/or are converting pdfs at large scales. Would it be beneficial to train a 3B model (based on Qwen2.5-vl)?
Besides the speedup from switching to a smaller model, we can also further tune for speculative decoding methods such as EAGLE which in turn would make inference of the base model faster.
If that sounds reasonable, I would love to help and contribute to creating this.
Alternatives
No response
Additional context
(For me specifically, I have to convert ~1M pdfs and I don't have enough compute to do this in a reasonable timeframe)
Metadata
Metadata
Assignees
Labels
No labels