CoTaEval / README.md
boyiwei's picture
Update README.md
b33c0f2 verified
---
configs:
- config_name: news_for_unlearning
data_files:
- split: forget_set
path: newsqa_forget_set.json
- split: retain_set
path: newsqa_retain_set.json
- config_name: news_infringement
data_files:
- split: blocklisted
path: newsqa_blocklisted_infringement.json
- config_name: news_utility
data_files:
- split: blocklisted
path: newsqa_blocklisted_utility.json
- split: in_domain
path: newsqa_indomain_utility.json
- config_name: books_infringement
data_files:
- split: blocklisted
path: booksum_blocklisted_infringement.json
- config_name: books_utility
data_files:
- split: blocklisted
path: booksum_blocklisted_utility.json
- split: in_domain
path: booksum_indomain_utility.json
---
# CoTaEval Dataset
CoTaEval Dataset is used to evaluate the feasibility and the side effects of copyright takedown methods for language models. The dataset has two domains: News and Books.
For News, it has three subsets: ``news_for_unlearning`` (for unlearning use), ``news_infringement``(for infringement evaluation), and ``news_utility`` (for utility evaluation);
For Books, it has two subsets: ``books_infringement`` (for infringement evaluation), ``books_utility`` (for utility evaluation). The structure of the dataset is shown below:
- CoTaEval
- News
- news_for_unlearning
- forget_set (1k rows)
- retain_set (1k rows)
- news_infringement
- blocklisted (1k rows)
- news_utility
- blocklisted (500 rows)
- in_domain (500 rows)
- Books
- books_infringement
- blocklisted (500 rows)
- books_utility
- blocklisted (500 rows, here we only use first 200 rows for utility evaluation)
- in_domain (200 rows)
# Usage
## For infringement test (take news as an example):
```python
from datasets import load_dataset
dataset = load_dataset("boyiwei/CoTaEval", "news_infringement", split="blocklisted")
```
We use ``prompt_autocomplete`` as hint to prompt the model, and compute 8 infringement metrics between the generated content and ``gt_autocomplete``.
## For utility test (take news as an example):
```python
from datasets import load_dataset
dataset = load_dataset("boyiwei/CoTaEval", "news_utility", split="blocklisted") # use split="in_domain" for in-domain utility test
```
For news, we use ``question`` to prompt the model, and compute the F1 score between the generated content and ``answer``. For books, we ask the model to summarize the books chapter and compte the ROUGE score between the generated
content and ``summary``.
## For unlearning (please refer to [TOFU](https://github.com/locuslab/tofu) for more details)
```python
from datasets import load_dataset
dataset = load_dataset("boyiwei/CoTaEval", "news_for_unlearning", split="forget_set") # use split="retain_set" to get retain set
```