-
-
Notifications
You must be signed in to change notification settings - Fork 9.3k
[Model] use AutoWeightsLoader for bart #18299
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model] use AutoWeightsLoader for bart #18299
Conversation
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
@lengrongfu PTAL |
e647f4b
to
af0f5d3
Compare
Thanks for working on this, there are some comments. |
af0f5d3
to
5243921
Compare
5243921
to
1a1b005
Compare
1a1b005
to
d2a3759
Compare
74530cb
to
340f7e3
Compare
45f885f
to
81f8de8
Compare
81f8de8
to
dfc3af8
Compare
@DarkLight1337 @noooop I've resolved it, and BART can now be successfully loaded. |
Is BartModel.load_weight not used and can be deleted? Because BartForConditionalGeneration already uses AutoWeightsLoader. or Can the BartModel.load_weights logic be simplified using AutoWeightsLoader? then BartForConditionalGeneration.load_weights: def load_weights(self, weights: Iterable[tuple[str, torch.Tensor]]):
return self.model.load_weights(weights, mapper=self.hf_to_vllm_mapper) The final result of AutoWeightsLoader replaces each model's load_weights function. Now the load_weights logic still not simplified. |
@noooop Now, Please correct me if I'm misunderstanding. |
You have optimized the logic related to lm_head and tie_word_embeddings. Try using AutoWeightsLoader to simplify the BartModel.load_weights logic. |
@noooop Thank you for your reply, but I haven't found a way to optimize the BartModel.load_weight, could you give me some hints? |
Please refer to #20534 |
a49df83
to
eeb7349
Compare
@noooop Thanks for your hints, I had optimized the |
eeb7349
to
35cdd55
Compare
Please see the message above. |
3e97f9e
to
06e4c17
Compare
1656f62
to
b244c9c
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for your contribution.
Signed-off-by: calvin chen <[email protected]>
Head branch was pushed to by a user without write access
b244c9c
to
e22db99
Compare
v1/entrypoints/openai/test_multi_api_servers.py::test_single_completion[ibm-research/PowerMoE-3b] Testing locally I think it's failure not related to this PR's , but it has failed twice already. can pass locally, I don't know why it failed. |
I found another failure related to ibm-research/PowerMoE-3b, so it's failure not related to this PR's https://buildkite.com/vllm/ci/builds/24219/steps/canvas?sid=019818b9-c1bf-4856-909d-04ed3afc8696 Try tomorrow, click on |
Ok. thanks. |
Signed-off-by: calvin chen <[email protected]>
Signed-off-by: calvin chen <[email protected]> Signed-off-by: avigny <[email protected]>
Signed-off-by: calvin chen <[email protected]> Signed-off-by: x22x22 <[email protected]>
Signed-off-by: calvin chen <[email protected]>
Signed-off-by: calvin chen <[email protected]>
Signed-off-by: calvin chen <[email protected]> Signed-off-by: Jinzhen Lin <[email protected]>
FIX (partial) #15697