-
Notifications
You must be signed in to change notification settings - Fork 528
[Model] DistilBERT #922
[Model] DistilBERT #922
Conversation
Codecov Report
|
Codecov Report
@@ Coverage Diff @@
## master #922 +/- ##
=========================================
- Coverage 88.32% 87.8% -0.53%
=========================================
Files 67 71 +4
Lines 6330 6723 +393
=========================================
+ Hits 5591 5903 +312
- Misses 739 820 +81
|
Job PR-922/1 is complete. |
Job PR-922/5 is complete. |
Job PR-922/6 is complete. |
Job PR-922/7 is complete. |
Job PR-922/8 is complete. |
Job PR-922/9 is complete. |
@@ -0,0 +1,211 @@ | |||
# coding: utf-8 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
How do you intend to maintain and verify this script?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
They're currently not maintained, and not documented. Ideally we should test them in CI (adding fairseq, tf, pytorch transformer). But I don't have the bandwidth to do that now...
Job PR-922/10 is complete. |
Job PR-922/11 is complete. |
|
||
Usage: | ||
|
||
pip3 install pytorch-transformers |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
That package seems not maintained anymore. Why not convert from transformers package? Can be addressed in a separate PR.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
At the time of writing this script, there was only the pytorch-transformer package. I think it's still useful to some users, who started to use bert with pytorch-transformers
.
Job PR-922/13 is complete. |
Job PR-922/12 is complete. |
Job PR-922/14 is complete. |
Job PR-922/15 is complete. |
@mli Super happy DistilBERT was added as a model option! |
It's a |
Description
The distilBERT model from https://arxiv.org/abs/1910.01108
Checklist
Essentials
Changes
Comments