It's not, it wouldn't make sense to have 100,000 tokens just for the first 100,000 numbers. There's a playground [1] where you can see how LLMs tokenize a string.
12345678987654321 is tokenized on various models like so:
Looks like number string parsing may be of enough importance to warrant token look-ahead recursive sub-parsing, then use the most "promising" token-ization; the one that generates the highest yielding probability tree following that number string.
12345678987654321 is tokenized on various models like so:
[1] https://huggingface.co/spaces/Xenova/the-tokenizer-playgroun...