Back to Basics
Collection
A collection of "Back to Basics" models/aretfacts
•
1 item
•
Updated
This model is a fine-tuned version of distilbert/distilgpt2 on the dany0407/eli5_category dataset, as part of the HuggingFace Casual Language modelling guide. It achieves the following results on the evaluation set:
Distil gpt2 fine-tuned on eli5 dataset
exploration and re-aquainting myself with fineruning/training at home
dany0407/eli5_category
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss |
|---|---|---|---|
| 3.9157 | 1.0 | 1325 | 3.8137 |
| 3.8203 | 2.0 | 2650 | 3.8041 |
| 3.7755 | 3.0 | 3975 | 3.8031 |
Base model
distilbert/distilgpt2