This might sounds like a bit of re-structuring but for the sake of future compatibility, I propose the following,
- Move to
huggingface trainer: This will help the repo to automatically adapt to deepspeed and all the exclusive features of transformers library.
- We don't have to re-invent the wheel. Given that we are using huggingface trainer, we only need to implement the following functions for a trainer for different tasks.
-- data_loader
-- DataCollator
-- compute_metrics
-- predictions (if needed)
- In case if we want to
finetune our full model, we don't have to change a lot in the surface level.
I would love to take some responsibility if needed. Let me know. @jaketae @tianjianjiang @wilsonyhlee
This might sounds like a bit of re-structuring but for the sake of future compatibility, I propose the following,
huggingfacetrainer: This will help the repo to automatically adapt todeepspeedand all the exclusive features of transformers library.--
data_loader--
DataCollator--
compute_metrics--
predictions(if needed)finetuneour full model, we don't have to change a lot in the surface level.I would love to take some responsibility if needed. Let me know. @jaketae @tianjianjiang @wilsonyhlee