this post was submitted on 08 Aug 2023
317 points (93.4% liked)

AssholeDesign

7645 readers
10 users here now

This is a community for designs specifically crafted to make the experience worse for the user. This can be due to greed, apathy, laziness or just downright scumbaggery.

founded 2 years ago
MODERATORS
 

Source: https://front-end.social/@fox/110846484782705013

Text in the screenshot from Grammarly says:

We develop data sets to train our algorithms so that we can improve the services we provide to customers like you. We have devoted significant time and resources to developing methods to ensure that these data sets are anonymized and de-identified.

To develop these data sets, we sample snippets of text at random, disassociate them from a user's account, and then use a variety of different methods to strip the text of identifying information (such as identifiers, contact details, addresses, etc.). Only then do we use the snippets to train our algorithms-and the original text is deleted. In other words, we don't store any text in a manner that can be associated with your account or used to identify you or anyone else.

We currently offer a feature that permits customers to opt out of this use for Grammarly Business teams of 500 users or more. Please let me know if you might be interested in a license of this size, and I'II forward your request to the corresponding team.

you are viewing a single comment's thread
view the rest of the comments
[–] Adalast 2 points 1 year ago

Good assessment. I am leery of companies using my data for AI training in some ways, but overall I understand that data is data to the AI model and it neither knows nor cares who I am and what I say. Also, after anonymization, sanitation, and cleaning, most data sets look like noise to the casual observer, even someone who knows what they are looking at usually has to take some time to get their head around a format, so it isn't like I'm really worried about some human looking at my data in the dataset.

My issue is, and will always be, data brokers. The instant that that data set is being sold to some broker who can rather trivially de-anonomize it by cross-referencing it with other data in their possession and turn around to sell it to god only knows who (fuck politicians using data brokers to target specific people), I have some serious problems. So I guess the collection and internal utilization of data is not a problem for me, it is what some greedy little shit looking to make his Q2 report bottom line decides to do with it that worries me.