{"name":"R-tokenizers","portdir":"R/R-tokenizers","version":"0.3.0","license":"MIT","platforms":"darwin","epoch":0,"replaced_by":null,"homepage":"https://docs.ropensci.org/tokenizers","description":"Fast, consistent tokenization of natural language text","long_description":"Fast, consistent tokenization of natural language text","active":true,"categories":["textproc","science","R"],"maintainers":[{"name":"vital.had","github":"barracuda156","ports_count":2571}],"variants":[],"dependencies":[{"type":"build","ports":["clang-16","R"]},{"type":"lib","ports":["R-SnowballC","R-CRAN-recommended","R-Rcpp","R-stringi"]},{"type":"run","ports":["R"]},{"type":"test","ports":["R-knitr","R-testthat","R-rmarkdown","R-covr","R-stopwords"]}],"depends_on":[{"type":"lib","ports":["R-textrecipes","R-tidytext"]},{"type":"test","ports":["R-quanteda"]}]}