Research & Citations
shodh-memory is grounded in decades of cognitive psychology and neuroscience research. Every constant in our codebase has a citation. Here are the papers that shaped our architecture.
📄 Read the Paper (PDF)Cite shodh-memory
If you use shodh-memory in your research, please cite it using one of the formats below.
@software{sharma2026shodh,
author = {Sharma, Varun},
title = {shodh-memory: Cognitive Memory for AI Agents},
year = {2026},
url = {https://github.com/varun29ankuS/shodh-memory},
doi = {10.5281/zenodo.18668709},
version = {0.1.90},
license = {Apache-2.0},
note = {Hebbian learning, 3-tier architecture (Cowan 2001), hybrid decay (Wixted 2004), spreading activation (Anderson 1984)},
}Sharma, V. (2026). shodh-memory: Cognitive memory for AI agents (Version 0.1.90) [Computer software]. https://doi.org/10.5281/zenodo.18668709
V. Sharma, "shodh-memory: Cognitive memory for AI agents," version 0.1.90, 2026. doi: 10.5281/zenodo.18668709
Memory Decay & Forgetting
How memories fade over time and why power-law decay matters
strength
100% |████
80% |██████
60% | █████▓▓
40% | ███▓▓▓▒▒
20% | ▓▓▒▒▒░░░░
10% | ▒░░░░░░░░░
+─────────────────────
0 1d 3d 7d 30d
exponential → power-lawOn the Form of Forgetting
Psychological Science, 2(6), 409-415
The Psychology and Neuroscience of Forgetting
Annual Review of Psychology, 55, 235-269
Reflections of the Environment in Memory
Psychological Science, 2(6), 396-408
Hebbian Learning & Synaptic Plasticity
Fire together, wire together—the basis of associative memory
before: after: ○ ─── ○ ● ═══ ● | | | | ○ ─── ○ ● ═══ ● weak edges strengthened (0.10) (+0.025/co-access) co-activation → stronger bonds
Synaptic Modifications in Cultured Hippocampal Neurons
Journal of Neuroscience, 18(24), 10464-10472
The Organization of Behavior
New York: Wiley
Spreading Activation
How activation spreads through associative networks
○
/ \
○ ○ ← hop 2 (0.49)
/ \ \
● ● ○ ← hop 1 (0.70)
\ /
★ ← query node (1.00)
activation = weight × 0.7^hops
surfaces related contextSpread of Activation
Journal of Experimental Psychology: Learning, Memory, and Cognition, 10(4), 791-798
spreadr: An R package for simulating spreading activation in a network
Behavior Research Methods, 51, 910-929
Sleep & Memory Consolidation
How replay during rest strengthens memory traces
awake consolidation stable ░░░▒░░▒░ → ▒▓▓▓▓▒▒░ → ████▓▒░░ fragile replay durable traces cycles engrams hippocampal replay → cortical storage (our maintenance cycles mirror this)
About Sleep's Role in Memory
Physiological Reviews, 93(2), 681-766
Cognitive Neuroscience of Emotional Memory
Nature Reviews Neuroscience, 7, 54-64
Interference & Competition
How similar memories compete and interfere
memory A [███████▓▓▓] sim=0.92
memory B [████████▓▒] sim=0.95
│
▼
retrieval competition: B wins
A suppressed by interference
threshold > 0.85 → inhibition
(prevents redundant recall)Critical Issues in Interference Theory
Memory & Cognition, 1, 19-40
Interference and Inhibition in Memory Retrieval
In E.L. Bjork & R.A. Bjork (Eds.), Memory (pp. 237-313). Academic Press
Cognitive Architecture
Computational models of human memory
+---------------------------+
| +---------------------+ |
| | +---------------+ | |
| | | Working | | | <- seconds
| | +---------------+ | |
| | Session | | <- hours
| +---------------------+ |
| Long-Term Memory | <- permanent
+---------------------------+
Cowan's embedded processesEvolving Conceptions of Memory Storage, Selective Attention, and Their Mutual Constraints
Psychological Bulletin, 104(2), 163-191
The Magical Number 4 in Short-Term Memory
Behavioral and Brain Sciences, 24(1), 87-114
How Can the Human Mind Occur in the Physical Universe?
Oxford University Press
Memory—a Century of Consolidation
Science, 287(5451), 248-251
The Neurobiology of Consolidations, or, How Stable Is the Engram?
Annual Review of Psychology, 55, 51-86
Graph-Enhanced Retrieval
Combining knowledge graphs with vector search
query → [semantic] ──┐
[keyword ] ──┤ RRF
[graph ] ──┤ fusion → results
[temporal] ──┘
multi-signal retrieval:
vector + BM25 + graph + recency
beats any single signal aloneFrom Local to Global: A Graph RAG Approach to Query-Focused Summarization
arXiv:2404.16130
A Syntactically-Based Query Reformulation Technique for Information Retrieval
Information Processing & Management, 42(5), 1332-1363
Associative Memory & Hopfield Networks
Content-addressable recall and energy-based memory retrieval—the foundation that transformers rediscovered
Energy landscape:
╲ ╱╲ ╱╲ ╱
╲ ╱ ╲ ╱ ╲ ╱
╲ ╱ ╲ ╱ ╲ ╱
╲╱ ╲╱ ╲╱
mem_A mem_B mem_C
input rolls to nearest valley
partial query → complete recallNeural Networks and Physical Systems with Emergent Collective Computational Abilities
Proceedings of the National Academy of Sciences, 79(8), 2554-2558
Hopfield Networks is All You Need
International Conference on Learning Representations (ICLR)
Dense Associative Memories for Pattern Recognition
Advances in Neural Information Processing Systems (NeurIPS), 29
Biologically Plausible Learning
Local learning rules that don't need backpropagation—how brains actually learn
backpropagation: shodh/biological: global loss → chain local co-activation rule across all → strengthen edge layers (non-bio) (Hebbian, local) needs full graph needs only neighbors O(params) O(1) per update brain can't do this brain does exactly this
The Forward-Forward Algorithm: Some Preliminary Investigations
arXiv:2212.13345
Synaptic Plasticity Forms and Functions
Annual Review of Neuroscience, 43, 95-117
A History of Spike-Timing-Dependent Plasticity
Frontiers in Synaptic Neuroscience, 3, 4
Vector Search & Approximate Nearest Neighbors
Efficient similarity search at scale without brute force
Vamana graph (DiskANN):
●───●───● medoid
|\ | / | (entry point)
| ●─●─● | |
|/ | \ | v
●───●───● greedy walk to
nearest neighbor
<100k: Vamana | >100k: SPANN+PQDiskANN: Fast Accurate Billion-point Nearest Neighbor Search on a Single Node
Advances in Neural Information Processing Systems (NeurIPS), 32
SPANN: Highly-efficient Billion-scale Approximate Nearest Neighborhood Search
Advances in Neural Information Processing Systems (NeurIPS), 34
Product Quantization for Nearest Neighbor Search
IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(1), 117-128
Open Science, Open Source
All 27 citations and 200+ tunable constants are documented in src/constants.rs with full justification. We believe AI memory systems should be grounded in science, not magic numbers.