I am trying to use Codex (via the website, not the CLI). My code requires downloading a model from Hugging Face and running inference. Given the large memory requirements of doing so, Codex returns the message:
The download would require fetching large model checkpoints from Hugging Face and loading them into memory. Because of the size and time involved, the script was interrupted before completion.
Therefore, the script cannot be fully executed in this environment without completing the large model download and associated resource requirements.
What’s the best way to use Codex when your code requires running inference?