Palo Memory Engines
Welcome to the detailed documentation for Mpalo's Palo AI Memory Engines. Each engine is designed to provide unique capabilities for integrating persistent, context-aware memory into your applications. Below, you'll find comprehensive information on Palo Mini, Palo Bloom, Palo DEEP, and Palo DEEP-Research, including their features, API names, technical specifications, and operational modes.
Please note: The previously announced "Palo Large" model is currently postponed as we focus on refining and enhancing our current suite of offerings.
Operating Modes
Palo AI Memory Engines offer two distinct operational modes that you can select based on your application's needs. This choice defines how Palo stores and recalls information, giving you powerful control over your AI's behavior.
1. Personalization Mode
Focus: Adaptive, "humanlike" memory.
How it Works: This mode uses vector reconstruction to recall memories. This can result in "blurry" recall, where the core patterns and context are remembered, but the exact wording might shift, similar to human memory. It allows for creative connections and emergent behavior.
Best For: Conversational chatbots, personal assistants, and creative applications where a humanlike feel is more important than perfect factual recall.
Available on: Palo Mini, Palo Bloom, Palo DEEP.
2. Research Mode
Focus: 100% accurate, factual recall with zero hallucinations.
How it Works: This mode stores the original text as metadata alongside the vector. When recalling a memory, Palo retrieves this exact, unaltered metadata, bypassing reconstruction entirely. This guarantees that what you put in is exactly what you get out.
Best For: Enterprise knowledge bases, legal or medical Q&A bots, technical documentation search—any application where factual precision is non-negotiable.
Available on: Palo DEEP, Palo DEEP-Research.
Featured AI Memory Engines at a Glance
All Palo engines are exceptionally fast and cost-effective, designed to provide a powerful memory layer for external LLMs.
Palo Mini
Ultra-fast AI memory engine for LLMs, enabling quick contextual recall for interactions supporting ~4096+ token context windows.
Learn more »Palo Bloom
Versatile AI memory engine for LLMs, offering a balance of deeper memory and performance for interactions supporting ~8192+ token context windows.
Learn more »Palo DEEP
Advanced AI memory engine for LLMs, providing highly reliable, accurate, and deep recall for complex applications with significantly larger context interactions.
Learn more »Palo DEEP-Research
Specialized for 100% accurate, factual recall. Ideal for enterprise knowledge bases, legal/medical docs, and applications where accuracy is non-negotiable.
Learn more »Memory Engine Comparison
Feature | Palo Mini | Palo Bloom | Palo DEEP | Palo DEEP-Research |
---|---|---|---|---|
API Name | palo-lite |
palo |
palo-deep |
palo-deep-research |
Primary Function | Fast contextual memory for LLMs | Balanced memory & performance | Advanced, deep memory for complex applications | Specialized memory for 100% accurate factual recall |
Supported Context (per Input) | ~4096+ tokens | ~8192+ tokens | Significantly Larger | Significantly Larger |
Key Memory Features | Episodic Recall, Semantic Search | Enhanced Recall & Search, Basic Relationship Linking | Memory Mapping, Advanced Accurate Recall, Traversal | All DEEP Features + Specialized Fine-tuning |
Primary Use Cases | Simple chatbots, CLI tools, basic personalization. | Personal assistants, mobile apps, educational tools. | Enterprise knowledge bases, complex robotics, advanced support. | Legal/Medical Q&A, compliance checks, technical lookups. |
Operational Mode(s) | Personalization | Personalization | Personalization & Research | Personalization & Research |
Performance | Exceptionally Fast | Very Fast | Fast, optimized for depth | Fast, optimized for accuracy |
Learn More | Details » | Details » | Details » | Details » |
Palo Mini
API Name: palo-lite
Palo Mini is an exceptionally fast and cost-effective AI memory engine designed to augment external LLMs. It provides essential episodic memory, enabling quick contextual recall for LLM-driven applications supporting interaction context windows of approximately 4096 tokens or more. Ideal for scenarios requiring rapid, memory-enhanced responses with minimal latency and resource usage. Operates primarily in "Personalization Mode."
Key Features & Specifications:
Palo Bloom
API Name: palo
Palo is a versatile and exceptionally fast, cost-effective AI memory engine that enhances external LLMs. It offers a balance of deeper memory capabilities and high performance, optimized for LLM applications on mobile/edge devices or those requiring robust memory for interaction context windows of approximately 8192 tokens or more. Operates primarily in "Personalization Mode."
Key Features & Specifications:
Palo DEEP
API Name: palo-deep
Palo DEEP is an advanced AI memory engine for external LLMs. Engineered for complex applications, it provides highly reliable and profound memory recall. It integrates sophisticated features like Memory Mapping and Memory Traversal for comprehensive semantic network building, ensuring nuanced and dependable context for the external LLM.
Key Features & Specifications:
Palo DEEP-Research
API Name: palo-deep-research
Palo DEEP-Research is our premier memory engine, specialized for applications where 100% factual accuracy and verifiable recall are non-negotiable. It leverages the full power of the DEEP engine and is fine-tuned for understanding and retrieving information from dense, specialized documents. It is the definitive choice for building mission-critical AI systems.
Key Features & Specifications:
Keep in Mind
All engines come with either a Personalization Mode, which offers humanlike blurry memory and forgetting, or a Research Mode that aims to enhance accuracy, knowledge breadth, and depth while ensuring that important details are not forgotten.
Our Mission, in short
At Mpalo, we stand against profit-over-people capitalism. The majority of profit is reinvested into research to ensure our technology remains consumer-friendly and transparent. We deliver modular, humanlike memory solutions that safeguard user data, prevent bias, and foster long-term, reliable storage of experiences.
Our commitment is to create technology that serves businesses, developers, and consumers alike—building trust, enhancing engagement, and igniting nostalgia through memory-driven AI that truly resonates.
Get Started
If you're new to Palo Bloom, start here to learn the essentials and make your first API call.
Was this page helpful?
Your feedback helps us improve our documentation.