Spaces:
Sleeping
Sleeping
import pytest | |
from app.model.tokenizer import NigerianLanguageTokenizer | |
from transformers import AutoTokenizer | |
def test_tokenizer(): | |
base_tokenizer = AutoTokenizer.from_pretrained("gpt2") | |
tokenizer = NigerianLanguageTokenizer(base_tokenizer) | |
text = "Sample text" | |
tokens = tokenizer.tokenize_batch([text]) | |
assert tokens is not None |