distilbert-base-fire-class

Model description

The distilbert-base-fire-class model is a distilled version of BERT, specifically fine-tuned for classification tasks related to the dataset on tweets about fire incidents. This model can be used for analyzing and classifying tweets that mention fire-related events or discussions.

Intended uses & limitations

Intended Uses

The distilbert-base-fire-class model is specifically designed for distinguishing whether a text is discussing fire-related incidents or not. It can be applied in various contexts such as:

  • Monitoring social media: Detecting and analyzing tweets or posts about fire incidents to help in early detection and response.
  • News filtering: Identifying news articles or reports that focus on fire events for more efficient information processing.
  • Emergency response systems: Assisting in categorizing incoming reports or alerts based on their relevance to fire-related topics.
  • Research and analysis: Supporting studies on public reactions and discussions about fire incidents by classifying relevant texts.

Limitations

While the model is effective in its intended use, there are several limitations to consider:

  • Dataset Specificity: The model is trained on a specific dataset of tweets about fire, which may limit its accuracy when applied to other types of texts or platforms.
  • Language and Context: The model may not perform as well on texts that include slang, regional dialects, or highly context-specific language that was not well-represented in the training data.
  • Generalization: This model is focused on fire-related text classification and may not generalize well to other topics without further fine-tuning or additional training data.

Training and evaluation data

The model was trained using the tweets_about_fire dataset. This dataset consists of tweets specifically mentioning fire incidents, providing a focused context for training the classification model.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 500
  • num_epochs: 3

Training results

Training Loss Epoch Step Validation Loss
0.4456 1.0 368 0.4997
0.2904 2.0 736 0.6664
0.2408 3.0 1104 0.5989

Framework versions

  • Transformers 4.38.2
  • Pytorch 2.2.1+cu121
  • Datasets 2.18.0
  • Tokenizers 0.15.2

Downloads last month
15
Safetensors
Model size
67M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Evaluation results