File size: 2,524 Bytes
9ccb1d0
 
6292707
 
 
 
 
9ccb1d0
0049b70
 
 
 
 
403db67
0049b70
403db67
 
0049b70
9bc1150
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
403db67
 
0049b70
403db67
0049b70
403db67
0049b70
403db67
0049b70
 
4e51420
d54ba18
403db67
0049b70
d54ba18
 
 
 
403db67
0049b70
403db67
0049b70
 
403db67
0049b70
4e51420
 
 
 
0049b70
403db67
 
0049b70
403db67
0049b70
403db67
0049b70
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
---
widget:
- text: The food was super tasty, I enjoyed every bite.
language:
- en
library_name: transformers
pipeline_tag: text-classification
---

# Model Card for Model ID

<!-- Provide a quick summary of what the model is/does. -->

This model performs sentiment analysis on user reviews, rating them on a scale from 1 (Very Bad) to 5 (Very Good).

# Model Details
## Model Description

The Bert-User-Review-Rating model is trained on a dataset of 1,300,000 reviews of public places and points of interest. It is capable of classifying user ratings from 1 to 5, where:

5 = Very Good

4 = Good

3 = Neutral

2 = Bad

1 = Very Bad

Model type: BERT-based sentiment analysis model

Language(s) (NLP): English

# Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
The model can be used directly to classify the sentiment of user reviews for public places and points of interest, providing a rating from 1 to 5.

# Bias, Risks and Limitations
The model may reflect biases present in the training data, such as cultural or regional biases, as training data reflects public places in Singapore.

## Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be aware of potential biases and limitations in the model’s performance, particularly when applied to reviews from different regions or contexts.

# How to Get Started with the Model
Use the code below to get started with the model.

## Use a pipeline as a high-level helper

from transformers import pipeline

pipe = pipeline("text-classification", model="mekes/Bert-User-Review-Rating")

result = pipe("The food was super tasty, I enjoyed every bite.")

print(result)

# Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->

The subjective performance of the model is better than the metrices, because it was evaluated regarding meeting the exact correct prediction, so if it reflects 4 instead of 5 it was counted as wrong prediciton. Even though the prediction was way better than predicting 1 for a 5 review. 

Test Accuracy: 0.714
Test F1 Score: 0.695
Test Loss: 0.698
Test Recall: 0.714

# Environmental Impact
Carbon emissions were estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019)

Calculations were done with Nvidia RTX 3090 instead of the used Nvidia RTX 4090. 

For one training run it emmited approximately 1,5 kg CO2