deberta-v3-base-zyda-2-sentiment

This model is a fine-tuned version of agentlans/deberta-v3-base-zyda-2 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.0408
  • Mse: 0.0408

Model description

More information needed

Intended uses & limitations

Example use:

import torch
import numpy as np
from transformers import AutoModelForSequenceClassification, AutoTokenizer

# Load model and tokenizer
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

model_name = "agentlans/deberta-v3-base-zyda-2-sentiment"
model = AutoModelForSequenceClassification.from_pretrained(model_name, num_labels=1).to(device)
tokenizer = AutoTokenizer.from_pretrained(model_name)

# Function to perform inference
def predict_score(text):
    inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True).to(device)
    with torch.no_grad():
        logits = model(**inputs).logits
    return logits.item()

# Example usage
input_text = "I accidentally the whole thing. Is that bad?"
score = predict_score(input_text)
print(f"Predicted score: {score}")

Example output:

Text Sentiment
Nothing seems to go right, and I'm constantly frustrated. -2.27
Everything is falling apart, and I can't see any way out. -2.11
I feel completely overwhelmed by the challenges I face. -1.43
There are some minor improvements, but overall, things are still tough. -0.76
I can see a glimmer of hope amidst the difficulties I encounter. 0.65
Things are starting to look up, and I’m cautiously optimistic. 1.65
There are many good things happening, and I appreciate them. 2.24
Every day brings new joy and possibilities; I feel truly blessed. 2.31
I’m feeling more positive about my situation than I have in a while. 2.38
Life is full of opportunities, and I'm excited about the future. 2.55

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 64
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 3.0

Training results

Training Loss Epoch Step Validation Loss Mse
0.0449 1.0 3143 0.0538 0.0538
0.0244 2.0 6286 0.0408 0.0408
0.016 3.0 9429 0.0426 0.0426

Framework versions

  • Transformers 4.46.3
  • Pytorch 2.5.1+cu124
  • Datasets 3.1.0
  • Tokenizers 0.20.3
Downloads last month
107
Safetensors
Model size
184M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for agentlans/deberta-v3-base-zyda-2-sentiment

Finetuned
(4)
this model

Dataset used to train agentlans/deberta-v3-base-zyda-2-sentiment