-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding support for Distilbert #21
base: main
Are you sure you want to change the base?
Conversation
Thank you! What differs DistilBERT from BERT? |
@marco-nicola what do you think friend? I’ll go for it but a bit worried about code duplication for just a few differences. |
Preferably just use the DistilBERT config (extend code in BERT) so there's no need for duplicate code. |
Got it, in that case extending the converter/preprocessing.go and converter/mapper.go for BERT would be the proper way to manage the differences in layer identifiers, together with the configuration. Let me know what you think, I can modify the PR for you to check this approach. |
I'm looking into supporting |
@mooijtech I am in vacation with family so it is a bit difficult for me to follow up on this now. I'll back to you next week and we'll figure it out together how to proceed with |
Based on the Bert's code for language modeling and text encoding tasks, these changes add support for DistilBert architecture #7 .