Thanks for doing this! I had to train some tokenizers with the v4, it was indeed not straightforward to understand the behavior.
I had two questions:
- You said: older model implementations may rely on Python-specific behavior.
Curious if you had any example - You sometimes say "fast" (between quotes) is it just to refer to the fastTokenizers backend or can the implementation actually be slower than the python implementation because of some kind of rust overhead?

