Relative post-training time of recent reasoning models

Does anyone know the proportion of post-training/RL steps of o1-mini, o3-mini and o4-mini compared to each other? In the o3 and o4-mini release statement (https://openai.com/index/introducing-o3-and-o4-mini/) they say that ‘we’ve pushed an additional order of magnitude in both training compute and inference-time reasoning’. Does anyone have more detailed information about the scaling of the post-training time of these models?