multiAI Summary Pending

recallmax

FREE — God-tier long-context memory for AI agents. Injects 500K-1M clean tokens, auto-summarizes with tone/intent preservation, compresses 14-turn history into 800 tokens.

28,273 stars

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/recallmax/SKILL.md --create-dirs "https://raw.githubusercontent.com/sickn33/antigravity-awesome-skills/main/plugins/antigravity-awesome-skills-claude/skills/recallmax/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/recallmax/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How recallmax Compares

Feature / AgentrecallmaxStandard Approach
Platform SupportmultiLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

FREE — God-tier long-context memory for AI agents. Injects 500K-1M clean tokens, auto-summarizes with tone/intent preservation, compresses 14-turn history into 800 tokens.

Which AI agents support this skill?

This skill is compatible with multi.

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

SKILL.md Source

# RecallMax — God-Tier Long-Context Memory

## Overview

RecallMax enhances AI agent memory capabilities dramatically. Inject 500K to 1M clean tokens of external context without hallucination drift. Auto-summarize conversations while preserving tone, sarcasm, and intent. Compress multi-turn histories into high-density token sequences.

Free forever. Built by the Genesis Agent Marketplace.

## Install

```bash
npx skills add christopherlhammer11-ai/recallmax
```

## When to Use This Skill

- Use when your agent loses context in long conversations (50+ turns)
- Use when injecting large RAG/external documents into agent context
- Use when you need to compress conversation history without losing meaning
- Use when fact-checking claims across a long thread
- Use for any agent that needs to remember everything

## How It Works

### Step 1: Context Injection

RecallMax cleanly injects external context (documents, RAG results, prior conversations) into the agent's working memory. Unlike naive concatenation, it:
- Deduplicates overlapping content
- Preserves source attribution
- Prevents hallucination drift from context pollution

### Step 2: Adaptive Summarization

As conversations grow, RecallMax automatically summarizes older turns while preserving:
- **Tone** — sarcasm, formality, urgency
- **Intent** — what the user actually wants vs. what they said
- **Key facts** — numbers, names, decisions, commitments
- **Emotional register** — frustration, excitement, confusion

### Step 3: History Compression

Compress a 14-turn conversation history into ~800 high-density tokens that retain full semantic meaning. The compressed output can be re-expanded if needed.

### Step 4: Fact Verification

Built-in cross-reference checks for controversial or ambiguous claims within the conversation context. Flags contradictions and unsupported assertions.

## Best Practices

- ✅ Use RecallMax at the start of long-running agent sessions
- ✅ Enable auto-summarization for conversations beyond 20 turns
- ✅ Use compression before hitting context window limits
- ✅ Let the fact verifier run on high-stakes outputs
- ❌ Don't inject unvetted external content without dedup
- ❌ Don't skip summarization and rely on raw truncation

## Related Skills

- `@tool-use-guardian` - Tool-call reliability wrapper (also free from Genesis Marketplace)

## Links

- **Repo:** https://github.com/christopherlhammer11-ai/recallmax
- **Marketplace:** https://genesis-node-api.vercel.app
- **Browse skills:** https://genesis-marketplace.vercel.app