This just weird Twitter rage. The user seems to think the private information in their tax return will leak, just because Gemini tried to summarize it.
To back this up, they reference the story about Gemini chats "leaking" on the public internet. What actually happened there was that people were creating publicly viewable chats via gemini.google.com/share/ links, and some of those got posted on social media and were indexed.
Title is sensationalized. According to the thread, Gemini is enabled only after opening Gemini. Ingestion is enabled until Gemini is closed.
> OK, more testing and I think I've figured it out (and it's still bad!). It seems that if you've ever clicked the Gemini button for a type of document then it remains open whenever you open another of that type--and therefore automatically ingests and summarizes it. So, e.g...
What is the problem here? Google already has your document, all they are doing with Gemini is running it through a bunch of transformers and spitting back some information. The act of sending input into an LLM does not itself train the LLM on that data (these systems don't "learn" via execution of normal input/output).
Is the concern that Google would consider your use of Gemini on this document as consent to use it for future training?
1. Some people (e.g. some artists) don't like generative AI as a matter of principle, seeing it as soulless, corporate, and entirely trained on stolen data.
2. Many people resent Clippy-style popup features. They appear at the most inconvenient times, and everyone knows they're mostly for the benefit of the product manager with a user count KPI. And the harder they get pushed, the more people resent them.
3. The distinction between things-known-from-training-data, things-known-from-context and things-known-from-RAG and so on is pretty opaque to most users - and not clearly guaranteed by the documentation. If it's an assistant that can schedule reminders, can find things in your e-mails and google drive, and it promises "Personal Results" where "your communication requests will be used to improve your experience with Gemini" the distinctions are pretty ambiguous.
4. The LLM industry norm is to play fast and loose with training data.
Unless explicitly promised otherwise, the assumption is always that anything being passed to a LLM API may be retained for various reasons. That’s the expectation that’s been set by the industry.
You should have the control of your information per use case / purpose. The ad networks are also Google's property - would you be fine with the Adsense looking at all your private documents? GDPR got that right - you consent to a specific usage, not to a free-for-all.
"Gemini is your always-on AI assistant across Google Workspace"
So user buys a product without even reading the product page, then complains about product? m'kay
NO! Don't put your private documents on someone else's computer, that's easy.
But if you do, don't put it on the hard drive of a company that makes money from your personal data for advertising, it sounds almost childish to explain.
It's not clear who you're replying to, but this isn't a conversation with a person. What's being discussed - away from the person who published this - are the truth claims in the post.
[+] [-] wcoenen|1 year ago|reply
To back this up, they reference the story about Gemini chats "leaking" on the public internet. What actually happened there was that people were creating publicly viewable chats via gemini.google.com/share/ links, and some of those got posted on social media and were indexed.
[+] [-] nitinreddy88|1 year ago|reply
[+] [-] Sayrus|1 year ago|reply
> OK, more testing and I think I've figured it out (and it's still bad!). It seems that if you've ever clicked the Gemini button for a type of document then it remains open whenever you open another of that type--and therefore automatically ingests and summarizes it. So, e.g...
[+] [-] rezonant|1 year ago|reply
Is the concern that Google would consider your use of Gemini on this document as consent to use it for future training?
[+] [-] michaelt|1 year ago|reply
1. Some people (e.g. some artists) don't like generative AI as a matter of principle, seeing it as soulless, corporate, and entirely trained on stolen data.
2. Many people resent Clippy-style popup features. They appear at the most inconvenient times, and everyone knows they're mostly for the benefit of the product manager with a user count KPI. And the harder they get pushed, the more people resent them.
3. The distinction between things-known-from-training-data, things-known-from-context and things-known-from-RAG and so on is pretty opaque to most users - and not clearly guaranteed by the documentation. If it's an assistant that can schedule reminders, can find things in your e-mails and google drive, and it promises "Personal Results" where "your communication requests will be used to improve your experience with Gemini" the distinctions are pretty ambiguous.
4. The LLM industry norm is to play fast and loose with training data.
[+] [-] zarzavat|1 year ago|reply
[+] [-] viraptor|1 year ago|reply
[+] [-] aniviacat|1 year ago|reply
[+] [-] baud147258|1 year ago|reply
[+] [-] fragmede|1 year ago|reply
[+] [-] egorfine|1 year ago|reply
Emotionally it makes my blood boil though. I am absolutely with the author.
I had to leave Sentry for crash reports because they did report things to me I have never asked them to watch for in the first place. Same vibe.
[+] [-] whoitwas|1 year ago|reply
[+] [-] michaelmior|1 year ago|reply
[+] [-] viraptor|1 year ago|reply
[+] [-] unknown|1 year ago|reply
[deleted]
[+] [-] kkfx|1 year ago|reply
[+] [-] 31337Logic|1 year ago|reply
[+] [-] hexage1814|1 year ago|reply
[+] [-] cynicalsecurity|1 year ago|reply
[+] [-] Gant1|1 year ago|reply
[+] [-] rvnx|1 year ago|reply
[+] [-] hiddencost|1 year ago|reply
[+] [-] BSDobelix|1 year ago|reply
NO! Don't put your private documents on someone else's computer, that's easy.
But if you do, don't put it on the hard drive of a company that makes money from your personal data for advertising, it sounds almost childish to explain.
[+] [-] robertlagrant|1 year ago|reply