Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I’ve used memory in Claude desktop for a while after MCP was supported. At first I liked it and was excited to see the new memories being created. Over time it suggests storing strange things to memories (an immaterial part of a prompt) and if I didn’t watch it like a hawk, it just gets really noisy and messy and made prompts less successful to accomplish my tasks so I ended up just disabling it.

It’s also worth mentioning that some folks attributed ChatGPT’s bout of extreme sycophancy to its memory feature. Not saying it isn’t useful, but it’s not a magical solution and will definitely affect Claude’s performance and not guaranteed that it’ll be for the better.





I have also created a MCP memory tool, it has both RAG over past chats and a graph based read/write space. But I tend not to use it much since I feel it dials the LLM into past context to the detriment of fresh ideation. It is just less creative the more context you put in.

Then I also made an anti-memory MCP tool - it implements calling a LLM with a prompt, it has no context except what is precisely disclosed. I found that controlling the amount of information disclosed in a prompt can reactivate the creative side of the model.

For example I would take a project description and remove half the details, let the LLM fill it back in. Do this a number of times, and then analyze the outputs to extract new insights. Creativity has a sweet spot - if you disclose too much the model will just give up creative answers, if you disclose too little it will not be on target. Memory exposure should be like a sexy dress, not too short, not too long.

I kind of like the implementation for chat history search from Claude, it will use this tool when instructed, but normally not use it. This is a good approach. ChatGPT memory is stupid, it will recall things from past chats in an uncontrolled way.


With ChatGPT the memory feature, particularly in combination with RLHF sampling from user chats with memory, led to an amplification problem which in that case amplified sycophancy.

In Anthropic's case, it's probably also going to lead to an amplification problem, but due to the amount of overcorrection for sycophancy I suspect it's going to amplify more of a aggressiveness and paranoia towards the user (which we've already started to see with the 4.5 models due to the amount of adversarial training).




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: