-
Notifications
You must be signed in to change notification settings - Fork 482
Open
Labels
contributions welcomePR from open source contributors welcome to solve this issue.PR from open source contributors welcome to solve this issue.feature requestNew feature or requestNew feature or requestgood first issueGood for newcomersGood for newcomers
Description
🚀 The feature, motivation, and pitch
With additive (delta-style) parameter-efficient tuning methods such as LoRa, we should be able to make a slightly more mem-efficient hydra architecture by using a single block that does ~frozen_head + tunable_weights
for the learner/policy head's fwd-pass and simply frozen_head
for the reference, instead of maintaining 2x heads.
CC @LouisCastricato and @cat-state for pointing this out
Alternatives
No response
Additional context
No response
OpdoopOpdoop
Metadata
Metadata
Assignees
Labels
contributions welcomePR from open source contributors welcome to solve this issue.PR from open source contributors welcome to solve this issue.feature requestNew feature or requestNew feature or requestgood first issueGood for newcomersGood for newcomers