June 26, 2016
2016 is the year of the shitlord but also the year of the shutdown.
The Shuttening continues, as both Facebook and Google have quietly begun programs to auto-censor wrongthink.
Neither company has made a public announcement on this issue.
Some of the web’s biggest destinations for watching videos have quietly started using automation to remove extremist content from their sites, according to two people familiar with the process.
The move is a major step forward for internet companies that are eager to eradicate violent propaganda from their sites and are under pressure to do so from governments around the world as attacks by extremists proliferate, from Syria to Belgium and the United States.
YouTube and Facebook are among the sites deploying systems to block or rapidly take down Islamic State videos and other similar material, the sources said.
Yes, this is always their explanation – it’s to stop Moslem terrorists.
However, these Moslem terrorists always seem to continue using these services, while we in the Alt-Right are shut down.
The technology was originally developed to identify and remove copyright-protected content on video sites. It looks for “hashes,” a type of unique digital fingerprint that internet companies automatically assign to specific videos, allowing all content with matching fingerprints to be removed rapidly.
Such a system would catch attempts to repost content already identified as unacceptable, but would not automatically block videos that have not been seen before.
So, when Evalion got shut down, I told people to reupload her videos. What this technology will do, when fully implemented, is make that impossible.
They are preparing to literally exterminate our presence on these video sharing sites.
The companies would not confirm that they are using the method or talk about how it might be employed, but numerous people familiar with the technology said that posted videos could be checked against a database of banned content to identify new postings of, say, a beheading or a lecture inciting violence.
The two sources would not discuss how much human work goes into reviewing videos identified as matches or near-matches by the technology. They also would not say how videos in the databases were initially identified as extremist.
The videos are flagged by users then reviewed by people in India.
Oh, yes, yes. Looking very much like they are doing a very nasty against the Jews, yes, we are seeing it. Yes, shutting down, thank you very much.
The process is not mysterious.
In late April, amid pressure from U.S. President Barack Obama and other U.S. and European leaders concerned about online radicalization, internet companies including Alphabet Inc’s YouTube, Twitter Inc, Facebook Inc and CloudFlare held a call to discuss options, including a content-blocking system put forward by the private Counter Extremism Project, according to one person on the call and three who were briefed on what was discussed.
CEP was co-founded by ((Joseph Lieberman))).
The discussions underscored the central but difficult role some of the world’s most influential companies now play in addressing issues such as terrorism, free speech and the lines between government and corporate authority.
None of the companies at this point has embraced the anti-extremist group’s system, and they have typically been wary of outside intervention in how their sites should be policed.
“It’s a little bit different than copyright or child pornography, where things are very clearly illegal,” said Seamus Hughes, deputy director of George Washington University’s Program on Extremism.
The companies now using automation are not publicly discussing it, two sources said, in part out of concern that terrorists might learn how to manipulate their systems or that repressive regimes might insist the technology be used to censor opponents.
“There’s no upside in these companies talking about it,” said Matthew Prince, chief executive of content distribution company CloudFlare. “Why would they brag about censorship?”
The two people familiar with the still-evolving industry practice confirmed it to Reuters after the Counter Extremism Project publicly described its content-blocking system for the first time last week and urged the big internet companies to adopt it.
This obviously has nothing at all to do with terrorism. The government doesn’t care about that. Right now, you can go on Twitter and find hundreds of ISIS accounts. You won’t find an Andrew Anglin account, however.
They will begin implementing this technology under the guise of blocking ISIS and they will quickly more into blocking other “extremist” content.
This is why it is so important for us to support new independent and free speech outlets, such as the video sharing site BitTube.
The way censorship works is that it gets progressively more extreme. Eventually, they will be censoring cuckservative videos, and then those people will have to migrate to our free speech systems, where they will be exposed to our hardcore realities.