{"name":"py-tokenizers","portdir":"python/py-tokenizers","version":"0.22.2","license":"Apache-2","platforms":"any","epoch":0,"replaced_by":null,"homepage":"https://github.com/huggingface/tokenizers","description":"Fast and Customizable Tokenizers","long_description":"Tokenizers provides an implementation of today's most used tokenizers, with a focus on performance and versatility. Includes BPE, WordPiece, and Unigram tokenizer implementations.","active":true,"categories":["python"],"maintainers":[],"variants":[],"dependencies":[{"type":"build","ports":["cargo","legacy-support","rust","clang-20"]},{"type":"lib","ports":["libunwind","py314-tokenizers"]}],"depends_on":[]}