(no title)
mark_l_watson | 3 days ago
At least some of us in HN talk about limiting the data we give to Facebook, Google, Microsoft, etc. Isn’t it just as important to limit what we share with non-privacy preserving AIs?
Note: tech friends have asked me how I can use slightly weaker AI models and be happy about it: I still use Gemini Plus (and Anthropic via AntiGravity) for technical work: everything I do as a software developer is open source and all of my writing (20+ books) is Open Content so I don’t care about privacy and being direct-marketed based on my tech work. To me it makes sense to use the best AI just for tech work and a private AI for everything else. Think about this if a family member has a serious health problem, or something else private: do you want to use open web searches and open AI chats, or do you want to use private web search and private AI access? Why not make privacy your default, except in special situations?
ignoramous|3 days ago
Duck.ai's Privacy Policy goes:
This is not much different to the BigLabs, tbh.Otoh, privatemode.ai, confer.to, trymaple.ai are at least attempting Apple AI-like confidentiality.
y-c-o-m-b|3 days ago
adamtaylor_13|3 days ago
I'm not worried about the privacy aspect though many suggest that I should be. The power the dossier has given them to navigate the medical industry in the United States has been absolutely incredible. They don't have to be stuck when a random doctor who has never heard of their illness suggests that they might be overreacting. They can simply find someone who will help them. They can talk, in medical lingo, about their test results and discuss them with the doctor on equal footing.
I'm not sure this would've been nearly as successful without Opus 4.5/4.6 driving the harness. I'm not also not sure what real privacy risk there is here; it all sounds very theoretical.
janejanejanejan|3 days ago