HW 2 - DistilBERT fine-tuned

Task

Fine-tuned DistilBERT for text classification on a classmate's HW1 dataset.

  • Dataset: ddecosmo/hw_text_dataset
  • Text column: Text
  • Label column: label (classes: ['asu', 'bucknell', 'cmu', 'duq', 'ucsd'])
  • Train/Eval split: 80/20 (stratified if available)

Training

  • Base model: distilbert-base-uncased
  • Epochs: 3, LR=5e-5, WD=0.01, warmup=10%
  • Batch size: 16
  • Best model by: F1 (macro)

Results (Test)

  • Accuracy: 0.4000
  • F1 (macro): 0.1231
  • Precision (macro): nan
  • Recall (macro): nan

Notes & Limitations

  • Small student dataset; results may vary with seeds.
  • Labels mapped as: {'asu': 0, 'bucknell': 1, 'cmu': 2, 'duq': 3, 'ucsd': 4}

AI Tool Disclosure

This notebook used ChatGPT for scaffolding code and documentation. All dataset selection, training, evaluation, and uploads were performed by the student.

Downloads last month
10
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train george2cool36/hw2_text_finetune_distilbert