mekes's picture
Update README.md
27f4103 verified
---
widget:
- text: The food was super tasty, I enjoyed every bite.
language:
- en
library_name: transformers
pipeline_tag: text-classification
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This model performs sentiment analysis on user reviews, rating them on a scale from 1 (Very Bad) to 5 (Very Good).
# Model Details
## Model Description
The Bert-User-Review-Rating model is trained on a dataset of 1,300,000 reviews of public places and points of interest. It is capable of classifying user ratings from 1 to 5, where:
5 = Very Good
4 = Good
3 = Neutral
2 = Bad
1 = Very Bad
Model type: BERT-based sentiment analysis model
Language(s) (NLP): English
# Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
The model can be used directly to classify the sentiment of user reviews for public places and points of interest, providing a rating from 1 to 5.
# Bias, Risks and Limitations
The model may reflect biases present in the training data, such as cultural or regional biases, as training data reflects public places in Singapore.
## Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be aware of potential biases and limitations in the model’s performance, particularly when applied to reviews from different regions or contexts.
# How to Get Started with the Model
Use the code below to get started with the model.
## Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-classification", model="mekes/Bert-User-Review-Rating")
result = pipe("The food was super tasty, I enjoyed every bite.")
print(result)
# Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
The subjective performance of the model is better than the metrices, because it was evaluated regarding meeting the exact correct prediction, so if it reflects 4 instead of 5 it was counted as wrong prediciton. Even though the prediction was way better than predicting 1 for a 5 review.
Test Accuracy: 0.714
Test F1 Score: 0.695
Test Loss: 0.698
Test Recall: 0.714
# Environmental Impact
Carbon emissions were estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019)
Calculations were done with Nvidia RTX 3090 instead of the used Nvidia RTX 4090.
For one training run it emmited approximately 1,5 kg CO2