by pacoWebConsult on 10/8/2025, 3:16:15 PM
by ffsm8 on 10/8/2025, 7:36:17 PM
The memory feature I'd like to have would need built-in support from anthropic
It'd be essentially
1. Language server support for lookups & keeping track of the code
2. Being able to "pin" memories to functions, classes, properties etc via the language server support/providing this context whenever changes are made in this function/class/properties etc, but not kept, so all following changes outside of that will no longer include this context (basically, changes that touch code with which memories will be done by agents with additional context, and only the results are synced back, not the way to achieve it)
3. Provide a ide integration for this context so you can easily keep track of what's available just by moving the cursor to the point the memory is pinned at
Sadly impossible to achieve via MCP.
by DenisM on 10/8/2025, 7:56:08 PM
I think everyone concluded at this point that we need to improve models memory capabilities, but different people take different approach.
My experience is that ChatGPT can engage in a very thoughtful conversations but if I ask for a summary it makes something very generic, useful to an outsider, but it does not catch salient points which were the most important outcomes.
Did you notice the same problem?
by datadrivenangel on 10/8/2025, 4:26:10 PM
How does Claude know when to try and remember?
Often memory works too well and crowds out new things, so how are you balancing that?
by bryanhogan on 10/8/2025, 3:11:41 PM
Why would you not use context files in form of .md? E.g. how the SpecKit project does it.
by iambateman on 10/8/2025, 3:53:56 PM
I’ve started asking Claude to write tutorials that live in a _docs folder alongside my code.
Then it can reference those tutorials for specific things.
Interested in giving this a shot but it feels like a lot of infrastructure.
by DenisM on 10/8/2025, 5:32:42 PM
Imho you would have an easier sell if you separate knowledge into tiers: 1)overall design 2) coding standards 3) reasoning that lead to design 4) components and their individual structure 5) your current issue 6) etc
Your project becomes progressively more valuable the further you go down the list. The overall design should be documented and curated to onboard new hires. Documenting current issues is a waste of time compared to capturing live discussion, so Recall is super useful here.
by daxfohl on 10/8/2025, 4:32:54 PM
I'm surprised Anthropic doesn't offer something like this server-side, with an API to control it. Seems like it'd be a lot more efficient than having client manually reworking the context and uploading the whole thing.
by jumski on 10/8/2025, 6:09:36 PM
Memory is hard! I'm very curious how the version history approach is working for you? Have you considered an age when retrieving? Is model supposed to manage the version history on its own? Is the semantic search used to help with that?
by aktuel on 10/8/2025, 7:12:07 PM
Wouldn't the cache over time also be filled up with irrelevant and redundant information?
by tarun_anand on 10/8/2025, 3:22:56 PM
Claude introduced it's own memories api.. have you had a look?
by the_arun on 10/8/2025, 3:50:09 PM
I wish there was a way to send compressed context to LLMs instead of plain text. This will reduce token size, performance & operational costs.
by asdev on 10/8/2025, 3:58:32 PM
The problem is you need to tell prompt Claude to "Store" or "Remember", if you don't it will never call the MCP server. Ideally, Claude would have some mechanism to store memories without any explicit prompting but I don't think that's currently possible today.
by warthog on 10/8/2025, 3:20:01 PM
imo it would be better to carry the whole memory outside of the inference time where you could use an LLM as a judge to track the output of the chat and the prompts submitted
it would sort of work like grammarly itself and you can use it to metaprompt
i find all the memory tooling, even native ones on claude and chatgpt to be too intrusive
by h1fra on 10/8/2025, 3:38:36 PM
I'm not super familiar with context and "memory", but adding context manually or via memory doesn't end up consuming context length either way?
by gmerc on 10/8/2025, 4:43:28 PM
Every single persistent memory feature is a persistence vector for prompt injection.
by alecco on 10/8/2025, 3:48:50 PM
Why not just ask CC to write a prompt or Markdown file to re-start the conversation in a new chat?
by jcmontx on 10/8/2025, 3:07:20 PM
If this delivers can be 100% game changer, I will try it out and give some feedback
by mannyv on 10/8/2025, 3:32:27 PM
This is excellent for those of us who are building local AIs.
by aiisthefiture on 10/8/2025, 7:35:35 PM
With redis? Why?
by otterley on 10/8/2025, 3:36:05 PM
Does it work with Valkey as well?
by bananapub on 10/8/2025, 4:09:35 PM
how did you benchmark this against much less convoluted solutions, like "a text file"?
how much better was this to justify all that extra complexity?
by iamleppert on 10/8/2025, 4:02:57 PM
I'm not seeing how this is any different than a standard vector database MCP tool. It's not like Claude is going to know about any of the things you told it to "remember" unless you explicitly tell it to use its memory tool like shown in the demo, to remember something you've stored.
by jMyles on 10/8/2025, 3:59:53 PM
Heh, I'm building the same thing this week (albeit with postgres rather than redis). I bet like 15% of the people here are.
Why would you bloat the (already crowded) context window with 27 tools instead of the 2 simplest ones: Save Memory & Search Memory? Or even just search, handling the save process through a listener on a directory of markdown memory files that Claude Code can natively edit?