chonkie and axonode-chunker
These are competitors offering different trade-offs in the document-chunking space: Chonkie prioritizes lightweight efficiency and production-ready RAG pipelines with broad adoption, while axonode-chunker targets semantic coherence and structural preservation for specialized use cases requiring fine-grained control over chunking behavior.
About chonkie
chonkie-inc/chonkie
🦛 CHONK docs with Chonkie ✨ — The lightweight ingestion library for fast, efficient and robust RAG pipelines
Provides pluggable chunking strategies—recursive, semantic, code-aware, and LLM-based—with composable pipeline workflows that chain multiple chunkers and refineries together. Integrates with 32+ tools across tokenizers (GPT-2, BPE), embeddings (OpenAI, Sentence Transformers), vector databases, and LLMs, while supporting 56 languages out-of-the-box through modular dependency installation.
About axonode-chunker
bazilicum/axonode-chunker
Advanced semantic text chunking with custom structural markers, whole-text coherence preservation, and flexible token management. Features async processing, LangChain integration, and dynamic drift detection. Ideal for RAG systems, augmented text processing, and domain-specific document analysis.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work