volexbpe is a low latency high throughput Byte-Pair encoding derived tokenizer providng exceptional performance & streamline interface.
- o200k_base - used in o3, o1, gpt-4o.
- cl100k_base - used in gpt-4, gpt-3.5 turbo, gpt-3.5, most openai text embedding endpoints.
- r50k_base - majority decreptad.
- gpt-2 - gpt-2, open source
Make sure you have uv
, if not
curl -LsSf https://astral.sh/uv/install.sh | sh
And the Rust toolchain installed, if not
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
then
git clone https://github.com/friedhar/veloxbpe.git
uv run maturin develop
will install veloxbpe
in the local virtual environment.
All benchmarks can be run locally. After you've built from source, run
uv run bench/benchmark_bandwidth_0.py
- Add support for custom BPE training.