File size: 2,269 Bytes
8e9c323
 
1334058
 
 
 
 
 
8e9c323
 
8c0805e
8e9c323
8c0805e
8e9c323
 
 
8c0805e
8e9c323
8c0805e
 
 
 
 
8e9c323
8c0805e
8e9c323
8c0805e
8e9c323
8c0805e
 
 
 
1334058
8e9c323
8c0805e
8e9c323
8c0805e
 
 
8e9c323
 
 
 
 
8c0805e
8e9c323
8c0805e
8e9c323
8c0805e
8e9c323
 
 
8c0805e
8e9c323
 
 
8c0805e
8e9c323
 
 
8c0805e
8e9c323
 
 
8c0805e
 
8e9c323
c600093
8c0805e
1334058
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
library_name: transformers
tags:
- text-classification
- quantization
- fine-tuning
base_model:
- allenai/scibert_scivocab_uncased
---

# Model Card for Research Paper Annotation Classifier

This model is a fine-tuned version of a pre-trained model for text classification. It is specifically designed to classify sentences from research papers into annotation categories.

## Model Details

### Annotation Categories

- **Methodology (0):** Describes methods or techniques used.
- **None (1):** Content irrelevant for annotation.
- **Novelty (2):** Highlights novel contributions.
- **Past Work (3):** References or compares past research.
- **Result (4):** Discusses experimental results or findings.

### Model Description

This model is part of the 🤗 Transformers library and has been fine-tuned to enable efficient annotation of academic texts. It takes a single sentence as input and predicts one of the five predefined categories to streamline the research annotation process.

- **Developed by:** Ashutosh Adhikari   
- **Model type:** Fine-tuned text classification model  
- **Language(s) (NLP):** English  
- **License:** Apache 2.0  
- **Finetuned from model:** `allenai/scibert_scivocab_uncased`

### Model Sources

- **Repository:** N/A
- **Paper:** N/A  
- **Demo:** N/A

## Uses

### Direct Use

This model can be used as a standalone text classifier to annotate sentences from research papers based on their semantic content.

### Downstream Use

The model can be fine-tuned further for similar tasks, such as classifying academic content in specific domains.

### Out-of-Scope Use

The model is not suitable for multi-paragraph classification or non-academic text.

## Bias, Risks, and Limitations

The model has been trained on specific datasets derived from research papers, so it may not generalize well to other domains or languages.

### Recommendations

Users should evaluate the model’s performance on their specific data and consider fine-tuning for domain-specific tasks.

## How to Get Started with the Model

```python
from transformers import pipeline

classifier = pipeline("text-classification", model="AshutoshAdhikari/rsclf-scibert-improved")
result = classifier("This paper introduces a novel technique for...")
print(result)