Finding that President Trump’s Twitter feed constitutes a public forum, a federal judge in New York City held that it’s a First Amendment violation when the President or one of his assistants blocks a Twitter user from viewing or responding to one of the President’s tweets. As the New York Times points out, the decision
With the effective date of the EU’s General Data Protection Regulation (GDPR) less than one month away, companies subject to the GDPR are racing to comply with the regulation’s data privacy laws. But, for those companies, May 25 doesn’t represent a finish line as much as it does a starting gate.
In the coming…
Based on copyright infringement, emotional distress and other claims, a federal district court in California awarded $6.4 million to a victim of revenge porn, the posting of explicit material without the subject’s consent. The judgment is believed to be one of the largest awards relating to revenge porn. A Socially Aware post that we wrote…
In a decision that has generated considerable controversy, a federal court in New York has held that the popular practice of embedding tweets into websites and blogs can result in copyright infringement. Plaintiff Justin Goldman had taken a photo of NFL quarterback Tom Brady, which Goldman posted to Snapchat. Snapchat users “screengrabbed” the image…
As we have noted previously, YouTube users sometimes object when the online video giant removes their videos based on terms-of-use violations, such as artificially inflated view counts. In a recent California case, Bartholomew v. YouTube, LLC, the court rejected a user’s claim that the statement YouTube posted after it removed her video, which allegedly gave the impression that the video contained offensive content, was defamatory.
Joyce Bartholomew is a musician who creates what she calls “original Christian ministry music.” Ms. Bartholomew produced a video for the song “What Was Your Name” and posted the video on YouTube in January 2014. YouTube assigned a URL to the video, which Ms. Bartholomew began sharing with her listeners and viewers. By April 2014, she claims that the video had amassed over 30,000 views.
Shortly afterwards, however, YouTube removed the video and replaced it with the image of a “distressed face” and the following removal statement: “This video has been removed because its content violated YouTube’s Terms of Service.” The removal statement also provided a hyperlink to YouTube’s “Community Guideline Tips,” which identifies 10 categories of prohibited content: “Sex and Nudity,” “Hate Speech,” “Shocking and Disgusting,” “Dangerous Illegal Acts,” “Children,” “Copyright,” “Privacy,” “Harassment,” “Impersonation” and “Threats.”…
Last year we covered a wide range of online legal and business subjects intended for readers ranging from Internet entrepreneurs to social media marketers, from online shoppers to e-tailers, from networkers to influencers (and the brands that pay them).
The topics of our blog posts covered a myriad of cutting-edge subjects, including a new federal…
In order to comply with a new German law requiring social media sites to take down hate speech, Twitter and Facebook removed anti-Islamic social media posts authored by a German far-right political party.
The Obama administration’s screening of social media accounts of aspiring immigrants from majority-Muslim nations yielded little actionable intelligence, but the Trump…
Often derided as clickbait, listicles get a bum rap. They can be light on substantive content, sure, but sometimes that’s a good thing, especially for the busy readers of legal blogs, who would do well to treat themselves to some easily browsable reading material once in a while.
And so, at Socially Aware, we’ve made an annual tradition of curating a “List of Lists”—an inventory of the predictions, retrospectives and roundups that we think will be of most interest to our readership.
We’ll update this page throughout the month as additional pertinent content is published.
Technology & Social Media Law
Social Media (General)
Happy 2018 to our readers! It has become a Socially Aware tradition to start the New Year with some predictions from our editors and contributors. With smart contracts on the horizon, the Internet of Things and cryptocurrencies in the spotlight, and a number of closely watched lawsuits moving toward resolution, 2018 promises to be an exciting year in the world of emerging technology and Internet law.
Here are some of our predictions regarding tech-related legal developments over the next twelve months. As always, the views expressed are not to be attributed to Morrison & Foerster or its clients.
From John Delaney, Co-Founder and Co-Editor, Socially Aware, and Partner at Morrison & Foerster:
Regarding Web Scraping
Web scraping is an increasingly common activity among businesses (by one estimate, web-scraping bots account for as much as 46% of Internet traffic), and is helping to fuel the “Big Data” revolution. Despite the growing popularity of web scraping, courts have been generally unsympathetic to web scrapers. Last August, however, web scrapers finally received a huge victory, as the U.S. District Court for the Northern District of California enjoined LinkedIn from blocking hiQ Labs’ scraping of publicly available user profiles from the LinkedIn website in the hiQ Labs, Inc. v. LinkedIn Corp. litigation. The case is now on appeal to the Ninth Circuit; although my sense is that the Ninth Circuit will reject the broad scope and rationale of the lower court’s ruling, if the Ninth Circuit nevertheless ultimately sides with hiQ Labs, the web scraper, the decision could be a game changer, bringing online scraping out of the shadows and perhaps spurring more aggressive uses of scraping tools and scraped data. On the other hand, if the Ninth Circuit reverses, we may see companies reexamining and perhaps curtailing their scraping initiatives. Either way, 2018 promises to bring greater clarity to this murky area of the law.
Regarding the Growing Challenges for Social Media Platforms
2017 was a tough year for social media platforms. After years of positive press, immense consumer goodwill and a generally “hands off” attitude from regulators, last year saw a growing backlash against social media due to a number of reasons: the continued rise of trolling creating an ever-more toxic online environment; criticism of social media’s role in the dissemination of fake news; the growing concern over social media “filter bubbles” and “echo chambers”; and worries about the potential societal impact of social media’s algorithm-driven effectiveness in attracting and keeping a grip on our attention. Expect to see in 2018 further efforts by social media companies to get out ahead of most if not all of these issues, in the hopes of winning over critics and discouraging greater governmental regulation.
Regarding the DMCA Safe Harbor for Hosting of User-Generated Content
The backlash against social media noted in my prior item may also be reflected to some extent in several 2017 court decisions regarding the DMCA safe harbor shielding website operators and other online service providers from copyright damages in connection with user-generated content (and perhaps in the CDA Section 230 case law discussed by Aaron Rubin below). After nearly two decades of court decisions generally taking an ever more expansive approach to this particular DMCA safe harbor, the pendulum begun to swing in the other direction in 2016, and this trend picked up steam in 2017, culminating in the Ninth Circuit’s Mavrix decision, which found an social media platform provider’s use of volunteer curators to review user posts to deprive the provider of DMCA safe harbor protection. Expect to see the pendulum continue to swing in favor of copyright owners in DMCA safe harbor decisions over the coming year.
Regarding Smart Contracts
Expect to see broader, mainstream adoption of “smart contracts,” especially in the B2B context—and perhaps litigation over smart contracts in 2019 . . . .
From Aaron Rubin, Co-Editor, Socially Aware, and Partner at Morrison & Foerster:
Regarding the CDA Section 230 Safe Harbor
We noted previously that 2016 was a particularly rough year for Section 230 of the Communications Decency Act and the immunity that the statute provides website operators against liability arising from third-party or user-generated content. Now that 2017 is in the rear view mirror, Section 230 is still standing but its future remains imperiled. We have seen evidence of Section 230’s resiliency in recent cases where courts rejected plaintiffs’ creative attempts to find chinks in the immunity’s armor by arguing, for example, that websites lose immunity when they use data analytics to direct users to content, or when they fail to warn users of potential dangers, or when they share ad revenue with content developers. Nonetheless, it is clear that the knives are still out for Section 230, including in Congress, where a number of bills are under consideration that would significantly limit the safe harbor in the name of combatting sex trafficking. I predict that 2018 will only see these efforts to rein in Section 230 increase. …
In an effort to deter hate groups from tweeting sanitized versions of their messages, Twitter has began considering account holders’ off platform behavior when the platform evaluates whether potentially harmful tweets should be removed and account holders should be suspended or permanently banned.
In connection with Congressional efforts to deter online sex trafficking by narrowing…