Hey everyone!
I’m trying to optimize the response time of my GPT-4 web app using Node.js. The speed significantly drops when implementing context via Redis vector and LangChain, with response times around 8 seconds from user query to response. Currently, we’re providing users with full context, but are testing partial context. I am willing to move to python if necessary, but I’m primarily looking for ideas on how to speed up response times.