a b/deidentify/tokenizer/tokenizer_en.py
1
import spacy
2
3
from deidentify.tokenizer import Tokenizer
4
5
NLP = spacy.load('en_core_web_sm')
6
7
8
class TokenizerEN(Tokenizer):
9
10
    def parse_text(self, text: str) -> spacy.tokens.doc.Doc:
11
        return NLP(text)