Big Tech VS Free Speech ?
The end of Section 230 may be the key!
Please Donate to https://fundly.com/stopthebias
Together we can bring attention to the social media censorship and hold these monopolies to the exemption they have hid behind.
It’s no longer a question of whether the Giant Social Media Companies – Google, Twitter, Facebook, Instagram, etc. – have become too powerful. They’ve matured to the point that they can actually affect what people see, read, listen to and even what … See More they think. To make matters worse, they’ve decided that they will use these powers to change voting patterns and to Censor speech that opposes their political beliefs.
It’s time to stop them before all is lost. Harmeet Dhillon (Attorney Suing Google and Republican Party Official) has been on Tucker Carlson’s show frequently of late and she warns,
“Trump won’t win in 2020 and we will never win another election if we don’t stop this!”
One of the most likely ways for Congress to stop them would be to revise Section 230 of the Communications and Decency Act (CDA) that provides a special exemption from liability for content that is posted on their platforms. This exemption was initially extended to them because they claimed that their platforms would be a place for people from all points of view to post their ideas. Given their current Censorship actions, we all know that is no longer the case.
Consequently, the Social Media Platforms should be subjected to the possibility that they be responsible for all content that is posted on their sites since they selectively publish just as the New York Times or Washington Post do. In fairness, then, the Social Media Platforms should bear the same risk of liability for their content as other publishers.
This move would, of course, destroy their business model so they would be likely to change the Censorship tactics they use against Conservatives in order to avoid any changes to Section 230 of the CDA.
Alternatively, the threat of Antitrust Litigation is another avenue that may get their attention. The government should apply the same techniques against these Social Media Giants as they used to bring Microsoft to heel.
Our goal is to see our leaders pursue these remedies before it’s too late!
The 1996 law that made the web is in the crosshairs
Internet companies have long been shielded from legal responsibility for toxic user content by the Section 230 statute. Now that they’re huge, rich, and behaving badly, that gift could be taken away.
In the face of that toxic content’s intractability and the futility of the tech giants’ attempts to deal with it, it’s become a mainstream belief in Washington, D.C.–and a growing realization in Silicon Valley–that it’s no longer a question of whether to, but how to, regulate companies like Google, Twitter, and Facebook to hold them accountable for the content on their platforms. One of the most likely ways for Congress to do that would be to revise Section 230.
UNDERSTANDING SECTION 230
Section 230 remains a misunderstood part of the law. As Wyden explained it to me, the statute provides both a “shield” and a “sword” to internet companies. The “shield” protects tech companies from liability for harmful content posted on their platforms by users. To wit:
(c) (1) No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.
Specifically, it relieves web platform operators of liability when their users post content that violates state law by defaming another person or group, or painting someone or something in a false light, or publicly disclosing private facts. Section 230 does not protect tech companies from federal criminal liability or from intellectual property claims.
“Because content is posted on their platforms so rapidly there’s just no way they can possibly police everything,” Senator Wyden told me.
The “sword” refers to the 230’s “good samaritan” clause, which gives tech companies legal cover for choices they make when moderating user content. Before § 230, tech companies were hesitant to moderate content for fear of being branded “publishers” and thus made liable for toxic user content on their sites. Per the clause:
(c) (2) (a) No provider or user of an interactive computer service shall be held liable on account of any action voluntarily taken in good faith to restrict access to or availability of material that the provider or user considers to be obscene, lewd, lascivious, filthy, excessively violent, harassing, or otherwise objectionable, whether or not such material is constitutionally protected
“I wanted to make sure that internet companies could moderate their websites without getting clobbered by lawsuits,” Wyden said on the House floor back in March. “I think everybody can agree that’s a better scenario than the alternative, which means websites hiding their heads in the sand out of fear of being weighed down with liability.”
Many lawmakers, including Wyden, feel the tech giants have been slow to detect and remove harmful user content, that they’ve used the legal cover provided by § 230 to avoid taking active responsibility for user content on their platforms.
And by 2016 the harmful content wasn’t just hurting individuals or businesses, but whole societies. Social sites like YouTube became unwitting recruiting platforms for violent terrorist groups. Russian hackers weaponized Facebook to spread disinformation, which caused division and rancor among voters, and eroded confidence in the outcome of the 2016 U.S. presidential election.
As Wyden pointed out on the floor of the Senate in March, the tech giants have even profited from the toxic content.
“Section 230 means they [tech companies] are not required to fact-check or scrub every video, post, or tweet,” Wyden said. “But there have been far too many alarming examples of algorithms driving vile, hateful, or conspiratorial content to the top of the sites millions of people click onto every day –companies seeming to aid in the spread of this content as a direct function of their business models.”
And the harm may get a lot worse. Future bad actors may use machine learning, natural language, and computer vision technology to create convincing video or audio footage depicting a person doing or saying something provocative that they didn’t really do or say. Such “Deepfake” content, skillfully created and deployed with the right subject matter at the right time, could cause serious harm to individuals, or even calamitous damage to whole nations. Imagine a deep-faked president taking to Twitter to declare war on North Korea.
It’s a growing belief in Washington in 2018 that tech companies might become more focused on keeping such harmful user content off of their platforms if the legal protections provided in § 230 were taken away.
There’s a real question over whether Wyden’s “shield” still fits. Section 230 says web companies won’t be treated as publishers, but they look a lot more like publishers in 2018 than they did in 1996.
In 1996 websites and services often looked like digital versions of real-world things. Craigslist was essentially a digital version of the classifieds. Prodigy offered an internet on-ramp and some bulletin boards. GeoCities let “homesteaders” build pages that were organized (by content type) in “neighborhoods” or “cities.”
Over time the dominant business models changed. Many internet businesses and publishers came to rely on interactive advertising for income, a business model that relied on browser tracking and the collection of users’ personal data to target ads.
To increase engagement, internet companies began “personalizing” their sites so that each user would have a different and unique experience, tailor-made to their interests. Websites became highly curated experiences served up by algorithms. And the algorithms were fed by the personal data and browsing histories of users.
Facebook came along in 2004 and soon took user data collection to the next level. The company provided a free social network, but harvested users’ personal data to target ads to them on Facebook and elsewhere on the web. And the data was very good. Not only could Facebook capture all kinds of data about a user’s tastes, but it could capture the user’s friends’ tastes too. This was catnip to advertisers because the social data proved to be a powerful indicator of what sorts of ads the user might click on.
Facebook also leveraged its copious user data, including that on the user’s clicks, likes, and shares, to inform the complex algorithms that curate the content in users’ news feeds. It began showing users the posts, news, and other content that the user–based on their personal tastes–was most likely to respond to. This put more attention-grabbing stuff in front of its users’ eyeballs, which pumped up engagement and created more opportunities to show ads.
This sounds a lot like the work of a publisher. “Our goal is to build the perfect personalized newspaper for every person in the world,” Facebook CEO Mark Zuckerberg said in 2014.
But Facebook has always been quick to insist that it’s not a publisher, just a neutral technology platform. There’s a very good reason for that: Publishers are liable for the content
Follow @PeterBoykin on Social Media
Cash App: https://cash.me/app/CJBHWPS
Cash ID: $peterboykin1
Showdown over Barr contempt charges brewing; Court battle over Trump tax records developing Dems ready to hold Barr in contempt, Starr blasts leak of ‘whiny’ Mueller letter As Democrats scheduled a vote this Wednesday on whether to hold Attorney General William Barr in contempt of Congress, former independent counsel Ken Starr sharply … See More criticized the leak of Special Counsel Robert Mueller’s letter following Barr’s summary of the Russia report. Details of the March 24 letter went public shortly before Barr testified before the Senate Judiciary Committee last week. Multiple news outlets reported that Mueller raised concerns about how Barr’s conclusions from the investigation were being portrayed. In an interview on “The Story with Martha MacCallum” on Monday, Starr called the release of the letter an “unforgivable sin” and said Mueller’s complaints were “whiny.” Wednesday’s scheduled contempt vote on Barr is based on the Justice Department’s failure to provide the full text of Mueller’s report by a Monday morning deadline.
Court battle over Trump’s tax records developing Treasury Secretary Steven Mnuchin has denied House Ways and Means Committee Chairman Richard Neal’s request for President Trump’s tax returns, setting up a likely court battle. In a letter on Monday, Mnuchin said he’d relied on the advice of the Justice Department and concluded the request lacked a “legitimate legislative purpose.” Mnuchin’s decision is sure to set in motion a legal battle over Trump’s tax returns. The likely options available to Democrats would be to subpoena the Internal Revenue Service for the returns or to file a lawsuit.
Trump pardons former Army lieutenant convicted of killing suspected al Qaeda terrorist President Trump has pardoned a former Army lieutenant who was convicted in 2009 of killing an Iraqi prisoner suspected of being an Al Qaeda terrorist, the White House announced Monday evening. White House Press Secretary Sarah Sanders cited “broad support” for Michael Behenna, of Edmond, Okla., “from the military, Oklahoma elected officials, and the public” — including 37 generals and admirals, along with a former Pentagon inspector general — as the reason for Trump’s clemency grant. Sanders also said Behenna had been a “model prisoner” while serving his sentence. A military court originally sentenced Behenna to 25 years for unpremeditated murder in a combat zone. However, the Army’s highest appellate court noted concern about how the trial court had handled Behenna’s claim of self-defense, Sanders said. The Army Clemency and Parole Board reduced his sentence to 15 years and paroled him in 2014, as soon as he was eligible.
Ingraham gets an up-close look at the border crisis On Monday’s edition of “The Ingraham Angle, host Laura Ingraham joined the U.S. Customs and Border Protection to get an exclusive lookat the border crossing in Del Rio, Texas.
Awaiting the royal reveal Britain’s Prince Harry and Meghan Markle welcomed a bouncing baby boy early Monday morning, and now the world is waiting to find out the royal newborn’s name and what he looks like. On Monday, a jubilant Harry said he and Markle still hadn’t decided on his name. Images of the baby boy were not released. A few details were released in statement on Instagram. “We are pleased to announce that Their Royal Highnesses The Duke and Duchess of Sussex welcomed their firstborn child in the early morning on May 6th, 2019. Their Royal Highnesses’ son weighs 7lbs. 3oz.,” the couple announced. “The Duchess and baby are both healthy and well, and the couple thank members of the public for their shared excitement and support during this very special time in their lives. More details will be shared in the forthcoming days.”