
Are you like me? Tired of opening the same notes before Monday planning and wishing assistant remembered which weeks I'd already tried a new habit. A small, recurring friction, re-reading the same context every few days, pushed me to try GLM-5 with its "deep memory" abilities.
I'm Anna. This piece is my field report: what surprised me, where the friction lingered, and whether GLM-5 deep memory actually made daily life a little lighter. I'll show how it behaved in three real workflows, explain the setup I used, note privacy trade-offs I double-checked, and close with a practical before-and-after comparison against GLM-4.7.

The first thing I noticed was that GLM-5 accepts and keeps more context without my frantic summary-ing. Where older assistants forced me to paste a paragraph and hope it stuck, this felt like handing over a shoebox of past notes and having the assistant pull out the right postcard when asked.
Concretely: I fed it three weeks of my weekly plans, including items I'd moved from week to week. When I asked, "What did I try last week that I should repeat?" the model retrieved not just the most recent entry but the pattern, three half-hearted attempts to build a 10-minute writing habit and one small tweak that actually stuck (writing upon waking, not at night). That pattern-level recall was the meaningful thing. It wasn't just verbatim recall: it identified the trend.
This matches what DataCamp calls "long-term memory" in LLMs, the kind that persists across sessions through external databases or vector stores rather than staying in the model's context window.
That said, I wasn't handing it a perfect archive. I deliberately left some notes messy, one-line items, typos, and shorthand. GLM-5's deep memory still pulled useful context from that mess more often than I expected. But occasionally it missed a detail I'd assumed it would keep (a cancelled appointment that was renamed), which reminded me not to treat deep memory as flawless archival software.

What surprised me more was how follow-ups changed. With GLM-4.7, I'd ask a question, get a decent answer, and then have to rewrite context into the next prompt. With GLM-5, follow-ups felt like half a conversation: I'd ask about a plan, tweak one parameter in a sentence, and the assistant would re-evaluate the whole plan with that tweak in mind.
For example, I asked it to adapt my weekly plan when I said, "Make it work for a day when I have a dentist appointment at 10 a.m." Instead of only moving one task, GLM-5 considered downstream effects: it suggested swapping a deep-focus writing block for an afternoon time slot and gave a short rationale (commute and mental energy). That kind of reasoning, propagating a single change through related tasks, made edits feel less manual.

How I tested it: each Sunday I dumped a rough bullet list of the past three weeks and a short intention list for the week ahead. I asked GLM-5 to synthesize what to keep, what to drop, and what to try differently.
What changed: instead of saying, "Here's a new plan," it said things like, "You tried A and B for three weeks: A stuck when you did it at 7 a.m., B never passed 1–2 attempts." That context made it easier to decide what to keep without re-reading everything. It reduced the mental load: I didn't have to remember whether something was a fresh experiment or an ongoing habit.
Friction: getting the past weeks in there required either a quick paste or using an app that syncs notes into the memory store. That felt like a small setup cost for the payoff.
I used deep memory gently here: short notes about people, how someone likes coffee, a curious detail they mentioned. GLM-5 linked those details to recent exchanges, so when I asked for a suggested icebreaker for a check-in message, the lines it proposed referenced the correct context without me retyping it.
What I liked: the phrasing felt natural and specific. Rather than a generic "ask about their project," it suggested, "Ask about the prototype they mentioned last month," and even referenced the exact phrase they used. That small specificity saved me the tiny awkwardness of vague messages.
I tested a 10-minute micro-habit: daily sketching. After four attempts, GLM-5 noticed a pattern: I skipped weekends. It suggested an adaptive cadence, shorter weekday check-ins and a low-effort weekend prompt, rather than insisting on daily consistency.
Why that mattered: the suggestion matched my real-life energy curve. I felt less guilty and more likely to continue. The model's adaptive nudge removed an all-or-nothing framing that had derailed earlier attempts.
Where it faltered: it sometimes misattributed causality, suggesting changes tied to factors I hadn't recorded (like assuming I'd stopped because of time rather than motivation). That's on me for not feeding it more structured signals (tags, reasons for skipping).
I'm keeping this setup note brief because most readers I know don't want a deep configuration session.
Full disclosure: I used a Macaron client during testing because it offered a practical way to manage persistent context with GLM-5.
What Macaron store (based on the docs and my confirmations): short persistent memory entries you choose to save, metadata about when entries were added, and usage signals that help the model surface relevant memories. These are the pieces that let GLM-5 say, "You mentioned this last month."
What Macaron don't store: raw local device data unrelated to the memory entries you explicitly save, and transient session context that isn't saved.
If you're like me, storing lightweight habit notes and small relationship facts, this felt reasonable. If you plan to store medical or legal details, treat the memory store like any other third-party service: minimize exposure, use deletion tools, and read the retention policy.
Macaron had an explicit memory review flow: that made it easier to audit what the assistant knew.

Before (GLM-4.7):
After (GLM-5 with deep memory):
Measured impact (what I actually noticed):
Lingering limits: GLM-5 sometimes overgeneralized or filled gaps with plausible but unverified inferences. I learned to confirm critical facts and to keep sensitive information out of persistent memory until I trusted the workflow.
We create Macaron because these recurring minor frustrations are so realistic: weekly plans to review old notes, habits of trying something but then forgetting the outcome, and always having to re-explain the context.
Macaron helps you organize these lightweight but persistent memories, allowing the assistant to "remember that you have tried something" at the right time.
Try Macaron here!

A small lingering observation
I'll keep using GLM-5 deep memory for those tiny, repetitive frictions, weekly planning and habit nudges, mostly. It didn't revolutionize my life, but it quietly took away a few little frictions that added up. Your mileage will depend on how much you care about shaving off those five-to-ten-minute annoyances and how comfortable you are with a third-party memory store. I'm curious whether the same patterns hold for people who share accounts or use the memory store across devices, worth testing, if you're the curious type.