Model Card for {{ model_id | default("Model ID", true) }}

This is finetuned version of DistilBERT that is used for sentiment analysis on NFL news titles.

Model Details

Model Description

This model uses the DistilBERT model to classify NFL news article titles as positive or negative.

  • Developed by: Devin DeCosmo
  • Model type: Binary Sentiment Analysis
  • Language(s) (NLP): English
  • License: MIT
  • Finetuned from model: DistilBERT

Uses

This is used for sentiment analysis of NFL articles, but could possibly be used for other article titles.

Direct Use

The direct use is to classify NFL articles as positive or negative.

Out-of-Scope Use

If the dataset was expanded, this could be used for sentiment analysis on other types of articles or find other features like bias towards a team or player.

Bias, Risks, and Limitations

This is trained off a small dataset of 100 titles, this small dataset could be liable to overfitting and is not robust.

Recommendations

The small dataset size means this model is not highly generalizable.

How to Get Started with the Model

Use the code below to get started with the model.

Training Details

Training Data

James-kramer/football_news

This is the training dataset used. It consists of 100 original titles used for validation along with 1000 synthetic pieces of data from training.

Training Procedure

This model was trained with DistilBERT using binary classification, a training split of 80%, and 5 epochs. I initially used more but this converged extremely quickly.

Evaluation

Testing Data, Factors & Metrics

Testing Data

James-kramer/football_news The testing data was the 'original' split, the 100 original titles in this set.

Factors

This dataset is evaluating whether the food is positive, "1", or negative, "0".

Metrics

The testing metric used was accuracy to ensure the highest accuracy of the model possible. I also considered testing time. This small langauge model ran extremely quickly with 102 steps per second.

Results

After training with the initial dataset, this model reached an accuracy of 100% in validation. This is likely due to the simplicity of the task, binary classification, along with distilBERT being made for tasks such as this.

Summary

This model reached a high accuracy with our current model, but this perfomance can not be confirmed to continue as the dataset was very small. Additional testing with more samples would be highly beneficial.

Downloads last month
13
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train ddecosmo/finetuned_model