I’m using the “content filter” engine to filter out the prompts.
But the question is do I need to run the completions as well, through it?
Warning the user before the prompt is consumed by the API is logical, to save tokens and prevent unsafe content in the first place.
According to the policy, you must not generate unsafe content. Hence, if the completion is unsafe, you would still be violating the policy, even if the prompt itself is safe. In practice however, a well-designed prompt that passes the moderation filter is extremely unlikely to generate an unsafe completion.
Note that while using the moderation filter may be necessary for approval, it’s not a garantuee for policy compliance. The filter isn’t perfect.