The rule of thumb for the conversion between tokens used and words represented is 0.75.
If we apply this rule to your document then we have 227 tokens used, which would on average English text encode 170.25 words, with your example being 159 words, a difference of 6.6% more, a value below the margin of error as 0.75 words per token is very approximate.
I’m not sure where your value of 74% comes from, could you explain your calculation method?