Automatic Learning
Hebbrix learns and improves completely automatically. No thumbs up/down required. It just works.
Zero User Effort
Unlike other systems that pester you for feedback, Hebbrix uses automatic reinforcement learning. It evaluates its own answers, detects hallucinations, and learns from every interaction - all behind the scenes.
How It Learns Automatically
Every time Hebbrix answers a question, it runs 6 automatic quality checks:
Self-Consistency Check
Generates the same answer multiple times. If answers match → high confidence → positive reward
Hallucination Detection (NLI)
Uses Natural Language Inference to verify answer is grounded in actual memories, not made up
LLM-as-Judge
The AI judges its own answer quality: "Was I accurate? Complete? Grounded in facts?"
Answer Quality Heuristics
Checks length, specificity, confidence. Contains numbers/dates? Good. Vague "maybe" answers? Bad.
Memory Attribution
Tracks which memories actually contributed to the answer. Good memories get rewarded.
Retrieval Quality
Measures how well retrieved memories matched the query (vector similarity scores)
What Happens Behind the Scenes
- 1User asks a question
- 2Hebbrix searches memories and generates answer
- 3Automatic reward calculator evaluates answer quality (-1 to +1)
- 4LLM-as-Judge determines which memories contributed
- 5Useful memories get their access_count increased
- 6Next time, better memories appear first!
Research-Backed
This approach is based on cutting-edge 2024-2025 research:
- Memory-R1 (arXiv:2508.19828) - Memory-based RL
- RLSR (2025) - Self-Reward for LLMs
- Self-RAG (ICLR 2024) - Self-Reflective RAG
- DeepSeek-R1 (Nature 2025) - Reinforcement Learning
The Bottom Line
You don't need to do anything. Just use Hebbrix normally, and it gets smarter every day. The system learns from its own performance, not from annoying feedback prompts.
