basically, LLM is not good at math because of its tokenizer and sampling system.
So there are many papers about this,
such as,
So my question is, does OpenAI do this for now? or in the future?
I do not think they do because their tokenizing/sampling system is open source then I think we do not see such a process to decode numbers in special way.
It basically means to retrain the whole model, which is not very economical at all. Plus, adding a separate line (or vector space rather than token embedding space) is still not economical. What about separately adding [img] [video] and all other applications? I believe those will make the specific application more robust. Still, the computation-wise is a bit heavy, and it should be more efficient if OpenAI (and other major LLM providers) could adopt it.
There’s already an approach for this, generally: embedding models. You can remap embeddings (Customizing embeddings | OpenAI Cookbook) if you want, in theory. (I don’t know if anyone still does this though)