What are the latest strategies for prevening prompt leaks?

Hi all,
I am using the OpenAI API to create a health-based chatbot. However, I have some proprietory information in the chatbot’s system message that I want to prevent from getting leaked. Anyone have any sources of information on various tactics or phrases to use to at least block 90% of the attempts by users to steal the system prompts? I don’t expect to prevent 100% of the leaks, but id like to at least prevent more than half. Thank you.

Didn’t put sensitive information into user space.

Anything you give the model can ultimately be retrieved.

1 Like