We’ve been working on a new multimodal memory system called Modal at elicitlabs.ai, designed to serve as a personalization layer for AI models. It handles multimodal ingestion + retrieval (text, images, audio, video), can be queried in real time, and currently achieves SOTA on the LoCoMo personalization benchmark, outperforming long-context ICL even at 29k tokens.
1 Like