Skip to content

Palo Memory Engines & Performance Benchmarks

Revolutionary episodic memory systems that enhance any LLM, chatbot, or robot with humanlike memory capabilities. Explore our models, their applications, and detailed performance benchmarks.

Reimagining Memory for AI

Mpalo's Palo models represent a fundamental shift in how AI systems remember and understand context. Unlike known RAG approaches that search for relevant information, our episodic memory system mimics human memory patterns, creating more natural, accurate, and meaningful interaction.

Why Episodic Memory Matters

Conventional AI systems struggle with maintaining context over long conversations. Our approach doesn't just retrieve information, it understands experiences and interactions contextually, just like humans do.

  • Maintain memory across different conversations
  • Process both text and visual information naturally
  • Develop truly personalized interactions based on shared history
  • Eliminate repetitive information requests from users

The Mpalo Difference

Our modular approach lets you enhance any AI system with Palo memory capabilities without replacing your existing infrastructure. We validate our unique approach through rigorous benchmarking.

  • Simple integration with any LLM or Robot
  • API-first design for maximum flexibility
  • User friendly browser-extension for Website Chatbots
  • External and Internal memory solutions for optimal performance
  • Close to humanlike memory patterns, not just search functionality
  • Transparent performance evaluation via benchmarking

Keep In Mind: Modes of Operation

All engines come with either a Personalization Mode, which offers humanlike blurry memory and forgetting, or a Research Mode that aims to enhance accuracy, knowledge breadth, and depth while ensuring that important details are not forgotten. These modes have distinct evaluation priorities in our benchmarks.

Models Overview

Lightweight Memory Solutions

Our efficient memory models designed for speed and accessibility, perfect for consumer applications and development environments.

Palo Mini

API: palo-lite

Our entry-level memory model offering essential contextual memory for everyday applications. Perfect for chatbots and lightweight consumer-facing applications where speed matters and resources are limited.

Ideal Use Cases:

  • Customer support chatbots
  • Personal assistant apps
  • Adaptive educational tools
  • E-commerce preference recall

Technical Features:

  • Low latency
  • Basic text memory
  • Cost-effective

Palo

API: palo

A balanced solution offering enhanced memory capabilities with minimal latency impact. Ideal for developer environments and applications requiring more advanced memory features.

Ideal Use Cases:

  • Multilingual customer service
  • Content creation style recall
  • Healthcare interaction history
  • Contextual code understanding

Technical Features:

  • Optimized performance/capability
  • Enhanced text, basic image memory
  • Multilingual support

Premium Memory Solutions

Our advanced memory models delivering the highest accuracy and most sophisticated memory capabilities for enterprise and research applications.

Palo Large

API: palo-large

A powerful memory system designed for enterprise applications requiring high accuracy and advanced memory capabilities. Supports complex use cases and multilingual environments.

Ideal Use Cases:

  • Enterprise knowledge management
  • Advanced CRM recall
  • Regulated financial advisory
  • Complex business assistants
  • Multinational support systems

Technical Features:

  • Superior text/image memory
  • Advanced long context
  • Enhanced recall accuracy

Palo Deep

API: palo-deep

Our Research Grade Engine representing the cutting edge of artificial memory research. Designed for applications requiring the most sophisticated memory capabilities available.

Ideal Use Cases:

  • Research synthesis assistants
  • Long-term therapeutic apps
  • Complex multimodal integrations
  • Scientific discovery platforms

Technical Features:

  • State-of-the-art architecture
  • Unlimited history support
  • Text, code, image integration
  • Advanced multimodal memory
  • Highest recall accuracy

Research Papers

Mpalo is dedicated to advancing the field of AI memory and cognitive architectures. On our research page, you'll find our publications, pre-prints, and articles detailing our innovations. We also share a curated list of influential papers from the broader research community that align with our mission and inspire our work.

Performance Benchmarking

To ensure transparency and validate the capabilities of our Palo engines, we employ a comprehensive benchmarking framework. We evaluate performance across key dimensions relevant to episodic memory systems, comparing against state-of-the-art models where applicable.

Our evaluation focuses on three core areas:

  • Embedding Accuracy: How well the engine understands the meaning and relationships within the data it stores.
  • Reconstruction Accuracy: The engine's ability to precisely recall past information and events.
  • Similarity Recall: How effectively the engine retrieves relevant memories based on semantic similarity to a query.

Evaluation strategies are tailored for both Personalization and Research modes, reflecting their different design goals. For detailed methodologies and datasets, explore the sections below. Our related research papers, which provide deeper insights into the technologies and theories, can be found on our Research page.

Benchmark: Embedding Accuracy

Embedding accuracy measures how well Palo engines capture the semantic meaning of text, images, and code in their internal representations. High-quality embeddings are crucial for understanding context and retrieving relevant memories.

We evaluate this using standard benchmarks like MTEB and STS, measuring performance with metrics such as Cosine Similarity and Spearman Correlation against human judgments.

Graph Placeholder: Embedding Quality Comparison

(Illustrative graph showing Palo models' performance on semantic similarity tasks vs. SoTA embedding models like Sentence Transformers, OpenAI Embeddings)

Performance varies by engine (Lite, Palo Bloom, Large, Deep), with premium engines demonstrating superior embedding quality for complex data types.

Benchmark: Reconstruction Accuracy

This benchmark assesses how accurately Palo engines can retrieve and regenerate specific details from past interactions or stored data. This is vital for episodic memory, especially in applications requiring faithful recall like healthcare or financial advisory (Research Mode), or plausible recall (Personalization Mode).

We use datasets involving long contexts and episodic narratives, employing metrics like ROUGE, BLEU, Exact Match, and specialized scores for Simple Recall and Chronological Awareness.

Graph Placeholder: Recall Accuracy Over Time / Context Length

(Illustrative graph showing Palo models' ability to recall specific facts from long documents or dialogues compared to long-context LLMs like Gemini 1.5 Pro or Llama 3.1. Could show difference between Personalization/Research modes).

Palo Large and Deep excel in long-context recall accuracy, crucial for enterprise and research applications. Personalization mode metrics focus more on plausibility and gist recall.

Benchmark: Similarity Recall (RAG-like Functionality)

Similarity recall measures the effectiveness of Palo engines in retrieving relevant memories based on semantic similarity to a given query. This is key for augmenting LLMs or chatbots with contextual information, similar to Retrieval-Augmented Generation (RAG).

Evaluation uses standard information retrieval benchmarks (e.g., BEIR, NQ, MS MARCO) and metrics like NDCG, Recall@k, Precision@k, and MRR.

Graph Placeholder: Retrieval Relevance (NDCG@k)

(Illustrative graph comparing Palo models' performance on information retrieval tasks against standard RAG pipelines or vector databases using metrics like NDCG or Recall@k).

The quality of similarity recall depends heavily on the underlying embedding accuracy. Premium Palo engines offer state-of-the-art retrieval performance for augmenting external systems.

Memory Management & Ethics

Mpalo is committed to responsible memory management. All Palo models include advanced mechanisms to ensure ethical, efficient, and privacy-preserving memory operations.

Intelligent Forgetting

Our models implement cognitive science-based forgetting mechanisms to prevent memory overload while preserving essential information. This is particularly nuanced in Personalization Mode.

  • Importance-weighted retention
  • Temporal decay patterns
  • Automatic summarization
  • Memory consolidation

Privacy & Ethics Controls

Our commitment to ethical AI means comprehensive controls for managing memory with user privacy and consent at the forefront.

  • Selective memory purging
  • Explicit consent management
  • Configurable retention policies
  • Privacy-preserving encryption

Learn more about our Trust & Ethics principles

Use Cases Summary

Palo engines enable a wide range of applications where contextual understanding and memory persistence are crucial. From simple chatbot recall to complex, long-term multimodal interactions, Mpalo provides the memory foundation for next-generation AI experiences, validated by strong benchmark performance.

Was this page helpful?

Your feedback helps us improve our documentation.

Main Navigation


API Sections


On This Page