It kind of is. Normal users have no idea of just how much spam and abuse they’re being shielded from by the very algorithms you want to ban. If it was let through, basically every single user-facing site and service would become unusable. (I work on abuse and account security, so I’ve got a decent idea on what the raw stream of sewage looks like). And that’s in the current situation, where the economics of spam at least are kept in check by those algorithms. It’d get even worse the moment that filtering spam becomes regulated.
Second, it’s just not possible for people to handle the spam and moderation problem themselves. 99.9% of the people do not have the expertise for handling their own spam. The few that remain do not have the data for it. I guess you might be imagining something like a “this is spam, don’t show me things like it” button. But that is just the kind of algorithmic content curation that you want to ban, since “things like this” is not objective.
What would such a transparent opt-in dialogue would say, and why would we expect it to discourage the majority of people from opting in?
I think it would say “click here if you want to see more videos like this”. It would be true, and basically every single user would click on it, because the reason they went to YouTube or TikTok or whatever was too see that video, they liked it, and they would like to see more.
What benefit is there to the average user to know just what features are fed into the model? I honestly can’t think of any. But the benefit to the malicious actors is massive. Abuse is an iterative and adversarial process, where the attacker has most of the advantages. If you additionally make give them the benefit of information assymmetry, by forcing the defender to reveal their data and their methods, the task becomes impossible.
And believe it or not, none of these companies actually want to be hosting Nazis. If nothing else, nobody will want to advertise if there’s a risk of the ads being shown in that kind of context. That’s why they have thousands, maybe tens of thousands, of people working on trust and safety. Yes, even Facebook.
If you want to get a feel for what this kind of work entails, and the challenges, the leaked Facebook TnS report about January 6th is well worth a read. I don’t love how Buzzfeed framed the story, but the report which is quoted in full at the end of the article is a great example of anti-abuse work, and one of the very few public examples:
Right. So that is not what you originally asked for, which was “as much specificity as possible”. Steam it as most “give me some plausible explanation” for why it was recommended rather than really allowing for you to understand why the decision was made. For example: why is the game you’re looking at similar to GTA V and Project Cars? As-is, they could justify literally every recommendation as “this is similar to GTA V”.
Ok, so that’s a good change in scope, because it’s now implementable and could be understood by a normal person.
Would it actually achieve what you want? I think your concern is the gradual radicalization via recommendations, right? The recommendations will ultimately just tell you that the recommendation is for something very similar to what you’ve already liked. You’re basically suggesting some kind of an intervention “hey, do you understand that these people whose content you’re watching are basically white supremacists”, but timed to happen early enough that the victim is shocked by that.
You won’t get that level of introspection or insight just by reporting the raw facts of why a recommendation was made. You’d need some very different system, that’s basically completely orthogonal to the recommendation system itself. An interesting research problem, but seems very hard to regulate into existence.
What you were proposing is going to apply to every search engine, no matter how it is monetized. What is the uncurated ranking of search results that they’d use? It simply does not exist. The closest you could get to is every search just returning 10 random links, whether they are related to your search terms or not. But even that wouldn’t work since there are subjective decisions going into just what pages make it into the search index in the first place.
Interesting idea. I guess I’d be a bit worried that the incentive for the researchers is to find something they can sensationalize. An “everything is basically OK” result might be publishable, but would sink into obscurity.