tokenizers and tokenizer.cpp

These are complements: tokenizer.cpp provides a C++ implementation optimized for inference efficiency, while huggingface/tokenizers is the reference Python library that tokenizer.cpp likely wraps or reimplements to achieve production-grade tokenization performance in resource-constrained environments.

tokenizers
90
Verified
tokenizer.cpp
23
Experimental
Maintenance 20/25
Adoption 25/25
Maturity 25/25
Community 20/25
Maintenance 13/25
Adoption 1/25
Maturity 9/25
Community 0/25
Stars: 10,520
Forks: 1,051
Downloads: 129,702,376
Commits (30d): 33
Language: Rust
License: Apache-2.0
Stars: 1
Forks:
Downloads:
Commits (30d): 0
Language: C++
License: Apache-2.0
No risk flags
No Package No Dependents

About tokenizers

huggingface/tokenizers

💥 Fast State-of-the-Art Tokenizers optimized for Research and Production

Implemented in Rust with Python/Node.js/Ruby bindings, it supports BPE, WordPiece, and Unigram tokenization algorithms with integrated normalization that tracks character-level alignment to original text. The library handles full preprocessing pipelines including truncation, padding, and special token injection, enabling both vocabulary training and inference through a unified modular API.

About tokenizer.cpp

Mbeeee111/tokenizer.cpp

📦 Optimize tokenization in C++ for HuggingFace models with a fast, production-ready library supporting BPE, WordPiece, and Unigram methods.

Scores updated daily from GitHub, PyPI, and npm data. How scores work