This makes it possible to run tokenize on texts that are longer than the number of tokens that fit the context length without having to try to guess how to cut in number of characters beforehand |
||
---|---|---|
.. | ||
__init__.py | ||
bpe_simple_vocab_16e6.txt.gz | ||
clip.py | ||
model.py | ||
simple_tokenizer.py |