About Anaconda Help Download Anaconda
If you were automatically logged out you may need to refresh the page. You're trying to access a page that requires authentication. ×

Convert natural language text into tokens. The tokenizers have a consistent interface and are compatible with Unicode, thanks to being built on the 'stringi' package. Includes tokenizers for shingled n-grams, skip n-grams, words, word stems, sentences, paragraphs, characters, lines, and regular expressions.

Label Latest Version
main 0.1.4
test-label 0.1.4

© 2025 Anaconda, Inc. All Rights Reserved. (v4.0.9) Legal | Privacy Policy