{"name":"py313-tokenizers","portdir":"python/py-tokenizers","version":"0.22.2","license":"Apache-2","platforms":"darwin","epoch":0,"replaced_by":null,"homepage":"https://github.com/huggingface/tokenizers","description":"Fast and Customizable Tokenizers","long_description":"Tokenizers provides an implementation of today's most used tokenizers, with a focus on performance and versatility. Includes BPE, WordPiece, and Unigram tokenizer implementations.","active":true,"categories":["python"],"maintainers":[],"variants":["universal"],"dependencies":[{"type":"build","ports":["clang-20","py313-installer","py313-setuptools-rust","py313-maturin","cargo","legacy-support","rust","py313-build"]},{"type":"lib","ports":["python313","py313-huggingface_hub","libunwind"]}],"depends_on":[{"type":"lib","ports":["py313-transformers"]}]}