this post was submitted on 15 Jul 2024
276 points (98.9% liked)
Privacy
31790 readers
207 users here now
A place to discuss privacy and freedom in the digital world.
Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.
In this community everyone is welcome to post links and discuss topics related to privacy.
Some Rules
- Posting a link to a website containing tracking isn't great, if contents of the website are behind a paywall maybe copy them into the post
- Don't promote proprietary software
- Try to keep things on topic
- If you have a question, please try searching for previous discussions, maybe it has already been answered
- Reposts are fine, but should have at least a couple of weeks in between so that the post can reach a new audience
- Be nice :)
Related communities
Chat rooms
-
[Matrix/Element]Dead
much thanks to @gary_host_laptop for the logo design :)
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
For the people who didn't read the article. Read this TLDR: When you open a Google Doc. A Gemini sidebar appears, so you can ask questions about the document. Here, it summarized a document without the user asking.
The article title makes it seem like they are using your files to train AI which no proof exists for that(yet)
At least the data is sent to Gemini servers. This alone can be illegal but I'm not sure. What I'm more sure about is that they do use the data to train the models.
Since it is Google Docs, the data is already on Google servers. But yeah, it doesn't exactly instill confidence into the confidentiality of documents on Google Docs.
Thank you for the service!
I see your point re training, but aint the entire point why they want peasants using their models is to train them more?
Generative AI doesn't get any training in use. The explosion in public AI offerings falls into three categories:
To make a good model you need two things:
User data might meet need 2, but it fails at need 1. Running random data through neural networks to make it more exploitable (more accurate interest extraction, etc) makes sense, but training on that data doesn't.
This is clearly demonstrated by Google's search AI, which learned lots of useful info from Reddit but also learned absurd lies with the same weight. Not just overtuned-for-confidence lies, straight up glue-the-cheese-on lies.
Thank you for explaining this.
Ok so what is ChatGPT angle here providing this services for "free"
What do they get out of it? or is this just a google play to get you in the door, then data mine?
They have two avenues to make money:
Probably market dominance