SentencePieceTokenizer
SentencePiece tokenizer for 'lang'
SentencePieceTokenizer( lang = "en", special_toks = NULL, sp_model = NULL, vocab_sz = NULL, max_vocab_sz = 30000, model_type = "unigram", char_coverage = NULL, cache_dir = "tmp" )
lang |
lang |
special_toks |
special_toks |
sp_model |
sp_model |
vocab_sz |
vocab_sz |
max_vocab_sz |
max_vocab_sz |
model_type |
model_type |
char_coverage |
char_coverage |
cache_dir |
cache_dir |
None
Please choose more modern alternatives, such as Google Chrome or Mozilla Firefox.