(no title)
jzapletal | 13 days ago
What surprised us:
- Cost: $0.0002/screenshot (we budgeted 100x more), guess cloud vision APIs got cheap fast
- CPU: 5% (exp. 50%) and laptop stays cool
- Quality: night and day vs local models, we tried running vision locally first and it was mediocre
It works by triggering a screenshot on activity, sending it to a cloud vision model for summarization, then deleting the screenshot and storing only the text in local SQLite. You query it via MCP – "what was I working on before lunch?" and Claude actually knows.
quinncom|12 days ago
Which local models did you try? GLM-OCR seems like it would excel at this: https://huggingface.co/zai-org/GLM-OCR
quinncom|12 days ago
Screenshots are summarized in ~28 seconds. Here's the last one:
> "The user switched to the Hacker News tab, displaying item 47049307 with a “Gave Claude photographic memory for $0.0002/screenshot” headline. The chat now shows “Sonnet 4.6” and a message asking “What have I been doing in the past 10 minutes?” profile, replacing prior Signal content. The satellite map background remains unchanged."
The satellite map background remains unchanged message appears in every summary (my desktop background is a random Google Maps satellite image that rotates every hour).
I would like to experiment with custom model instructions – for example, to ignore desktop background images.
Earlier in my testing it was sending screenshots for both of my displays at the same time, which was much slower, but now it's only sending screenshots of my main screen. Does MemoryLane only send screenshots for displays that have active windows?
Here's the first test of the MCP server in Claude – https://ss.strco.de/SCR-20260217-onbp.png – it works!
BloondAndDoom|13 days ago
fidorka|13 days ago