{"name":"py312-tokenizers","portdir":"python/py-tokenizers","version":"0.22.2","license":"Apache-2","platforms":"darwin","epoch":0,"replaced_by":null,"homepage":"https://github.com/huggingface/tokenizers","description":"Fast and Customizable Tokenizers","long_description":"Tokenizers provides an implementation of today's most used tokenizers, with a focus on performance and versatility. Includes BPE, WordPiece, and Unigram tokenizer implementations.","active":true,"categories":["python"],"maintainers":[],"variants":["universal"],"dependencies":[{"type":"build","ports":["clang-20","py312-installer","py312-maturin","py312-setuptools-rust","cargo","legacy-support","rust","py312-build"]},{"type":"lib","ports":["python312","py312-huggingface_hub","libunwind"]}],"depends_on":[{"type":"lib","ports":["py312-transformers"]}]}