avilum/minrlm
Token-efficient Recursive Language Model. 3.6x fewer tokens than vanilla LLMs. Data never enters the prompt.
Implements a REPL-based execution model where the LLM generates Python code to query data directly, keeping raw context out of the prompt entirely; uses entropy profiling via zlib compression to identify relevant sections and task-specific routing to optimize code patterns for structured data, search, math, and code retrieval tasks. Wraps execution in a DockerREPL sandbox (seccomp, stdlib-only) and optionally delegates smaller sub-tasks to a secondary LLM on filtered evidence, achieving 30pp accuracy gains over vanilla on frontier models while maintaining flat token cost regardless of document size.
Used by 1 other package. Available on PyPI.
Stars
31
Forks
3
Language
Python
License
MIT
Category
Last pushed
Mar 18, 2026
Monthly downloads
202
Commits (30d)
0
Dependencies
1
Reverse dependents
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/avilum/minrlm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
hassancs91/SimplerLLM
Simplify interactions with Large Language Models
kyegomez/SingLoRA
This repository provides a minimal, single-file implementation of SingLoRA (Single Matrix...
tylerelyt/LLM-Workshop
🌟 Learn Large Language Model development through hands-on projects and real-world implementations
NetEase-Media/grps_trtllm
Higher performance OpenAI LLM service than vLLM serve: A pure C++ high-performance OpenAI LLM...
gtausa197-svg/-Project-Nord-Spiking-Neural-Network-Language-Model
The first pure SNN language model trained from scratch with a fully original architecture. 144M...