Skip to content

Conversation

maxreciprocate
Copy link
Collaborator

This PR

  • truncates input into pipeline to bert's context length (512)
  • puts pipeline on processes' local gpus for 2x speedup in training on single gpu (1:02 v 2:19 for 128 gradient steps), a bit more on multigpu (1:07 v 9:26 on 8 gpus) at cost of 0.5G extra VRAM
  • removes depreciation warning of return_all_scores usage in pipeline, on transformers version 4.24.0

@Dahoas Dahoas merged commit f552561 into master Nov 10, 2022
@maxreciprocate maxreciprocate deleted the fix-sentiments branch November 10, 2022 16:24
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants