January 27, 2019
Our beloved YouTube benefactors are working hard to keep dangerous information from spreading. The infection must not spread, you understand? This is actually good for you. It will make you enjoy YouTube more. Believe us.
From YouTube’s Official Blog:
When recommendations are at their best, they help users find a new song to fall in love with, discover their next favorite creator, or learn that great paella recipe. That’s why we update our recommendations system all the time—we want to make sure we’re suggesting videos that people actually want to watch.
We’ll continue that work this year, including taking a closer look at how we can reduce the spread of content that comes close to—but doesn’t quite cross the line of—violating our Community Guidelines. To that end, we’ll begin reducing recommendations of borderline content and content that could misinform users in harmful ways—such as videos promoting a phony miracle cure for a serious illness, claiming the earth is flat, or making blatantly false claims about historic events like 9/11.
We’re at a point in our civilization where further discussion is not really needed. We have our absolute truths brought by our cool fellow whites who just happen to suffer from really cold hands hence the constant rubbing, and frankly we’re fine as we are now. We don’t need a marketplace of ideas, do we? That’s just going to generate more division and hate, and what we need right now is unity, because diversity is our strength.
While this shift will apply to less than one percent of the content on YouTube, we believe that limiting the recommendation of these types of videos will mean a better experience for the YouTube community. To be clear, this will only affect recommendations of what videos to watch, not whether a video is available on YouTube. As always, people can still access all videos that comply with our Community Guidelines and, when relevant, these videos may appear in recommendations for channel subscribers and in search results. We think this change strikes a balance between maintaining a platform for free speech and living up to our responsibility to users.
That must be a very dangerous one percent of content produced by some very dangerous bad goys.
This change relies on a combination of machine learning and real people. We work with human evaluators and experts from all over the United States to help train the machine learning systems that generate recommendations. These evaluators are trained using public guidelines and provide critical input on the quality of a video.
Flat Earth, 9/11 videos, miracle cures… excuses. These “trained human evaluators and experts” are going to be vetting political content to keep the Goyim from knowing.
Trust us, goy.
This new law is to keep criminals such as pedophiles, rapists, and murderers out of the workforce, and sure, we’ll mainly use it to keep wrong-thinkers out of the workplace, but those are also criminals, you know? They produce intangible hate crimes.
Did I say workforce? I may have skipped a couple of steps there. I meant YouTube.