RuBert-tiny2-EmotionsDetected
This model was obtained by fine-tuning the model RuBert-tiny2 on the dataset ru-goemotions containing 28 emotions:
0: admiration (восхищение)
1: amusement (веселье)
2: anger (злость)
3: annoyance (раздражение)
4: approval (одобрение)
5: caring (забота)
6: confusion (непонимание)
7: curiosity (любопытство)
8: desire (желание)
9: disappointment (разочарование)
10: disapproval (неодобрение)
11: disgust (отвращение)
12: embarrassment (смущение)
13: excitement (возбуждение)
14: fear (страх)
15: gratitude (признательность)
16: grief (горе)
17: joy (радость)
18: love (любовь)
19: nervousness (нервозность)
20: optimism (оптимизм)
21: pride (гордость)
22: realization (осознание)
23: relief (облегчение)
24: remorse (раскаяние)
25: sadness (грусть)
26: surprise (удивление)
27: neutral (нейтральность)
At the moment, the model has been trained for 40 epochs with the following hyperparameters:
per_device_train_batch_size=16,
per_device_eval_batch_size=16,
num_train_epochs=1,
weight_decay=0.01,
learning_rate=1e-5,
save_total_limit=2,
load_best_model_at_end=True,
metric_for_best_model="f1",
greater_is_better=True
Test
With the help of a well-known AI, a data set was compiled to verify the veracity of the model's predictions. Here's what happened:
!The data for the test did not undergo a thorough manual inspection!
It can be seen from the above that the model is able to distinguish between basic emotions, but has problems with specific ones (grief (горе), pride (гордость)), and also defines a neutral emotion more often than expected (most likely this happened due to inaccuracies in the dataset, since in ordinary speech you can "use" several emojis at once and most often people use the neutral one)
The model will be updated as improvements are made.
Usage example
import torch
from transformers import AutoTokenizer, AutoModelForSequenceClassification
tokenizer = AutoTokenizer.from_pretrained('AniMAntZeZo/RuBert-tiny2-EmotionsDetected')
model = AutoModelForSequenceClassification.from_pretrained('AniMAntZeZo/RuBert-tiny2-EmotionsDetected')
model.to("cuda" if torch.cuda.is_available() else "cpu",
emotion_columns = [
"admiration", "amusement", "anger", "annoyance", "approval", "caring", "confusion", "curiosity", "desire",
"disappointment", "disapproval", "disgust", "embarrassment", "excitement", "fear", "gratitude", "grief", "joy",
"love", "nervousness", "optimism", "pride", "realization", "relief", "remorse", "sadness", "surprise", "neutral"
]
def predict_emotions(
text,
model,
tokenizer,
emotion_columns,
device="cuda" if torch.cuda.is_available() else "cpu",
threshold=0.1
):
emotion_translations = {
"admiration": "восхищение",
"amusement": "веселье",
"anger": "злость",
"annoyance": "раздражение",
"approval": "одобрение",
"caring": "забота",
"confusion": "непонимание",
"curiosity": "любопытство",
"desire": "желание",
"disappointment": "разочарование",
"disapproval": "неодобрение",
"disgust": "отвращение",
"embarrassment": "смущение",
"excitement": "возбуждение",
"fear": "страх",
"gratitude": "признательность",
"grief": "горе",
"joy": "радость",
"love": "любовь",
"nervousness": "нервозность",
"optimism": "оптимизм",
"pride": "гордость",
"realization": "осознание",
"relief": "облегчение",
"remorse": "раскаяние",
"sadness": "грусть",
"surprise": "удивление",
"neutral": "нейтральность",
}
model.to(device)
model.eval()
inputs = tokenizer(text, return_tensors="pt", padding="max_length", truncation=True, max_length=128).to(device)
with torch.no_grad():
logits = model(**inputs).logits
probabilities = torch.sigmoid(logits).squeeze().cpu().numpy()
predictions = {
f"{emotion} ({emotion_translations[emotion]})": prob
for emotion, prob in zip(emotion_columns, probabilities) if prob > threshold
}
sorted_predictions = dict(sorted(predictions.items(), key=lambda item: item[1], reverse=True))
return sorted_predictions
[INPUT]
example_text = "Как же я рад!"
predictions = predict_emotions(example_text, model, tokenizer, emotion_columns)
print("Emotions:", predictions)
[OUTPUT]
Emotions: {'joy (радость)': 0.6736836, 'excitement (возбуждение)': 0.25723574}
[INPUT]
example_text = "Я обиделся!"
predictions = predict_emotions(example_text, model, tokenizer, emotion_columns)
print("Emotions:", predictions)
[OUTPUT]
Emotions: {'sadness (грусть)': 0.3111033, 'disappointment (разочарование)': 0.2943853, 'annoyance (раздражение)': 0.19748639, 'anger (злость)': 0.16338393}
- Downloads last month
- 5
Model tree for AniMAntZeZo/RuBert-tiny2-EmotionsDetected
Base model
cointegrated/rubert-tiny2