Skip to main content
https://www.highperformancecpmgate.com/rgeesizw1?key=a9d7b2ab045c91688419e8e18a006621

Facebook partially documents its content recommendation system

Algorithmic recommendation systems on social media sites like YouTube, Facebook and Twitter, have shouldered much of the blame for the spread of misinformation, propaganda, hate speech, conspiracy theories and other harmful content. Facebook, in particular, has come under fire in recent days for allowing QAnon conspiracy groups to thrive on its platform and for helping militia groups to scale membership. Today, Facebook is attempting to combat claims that its recommendation systems are at any way at fault for how people are exposed to troubling, objectionable, dangerous, misleading, and untruthful content.

The company has, for the first time, made public how its content recommendation guidelines work.

In new documentation available in Facebook’s Help Center and Instagram’s Help Center, the company details how Facebook and Instagram’s algorithms work to filter out content, accounts, Pages, Groups and Events from its recommendations.

Currently, Facebook’s Suggestions may appear as Pages You May Like, “Suggested For You” posts in News Feed, People You May Know, or Groups You Should Join. Instagram’s suggestions are found within Instagram Explore, Accounts You May Like, and IGTV Discover.

The company says Facebook’s existing guidelines have been in place since 2016 under a strategy it references as “remove, reduce, and inform.” This strategy focuses on removing content that violates Facebook’s Community Standards, reducing the spread of problematic content that does not violate its standards, and informing people with additional information so they can choose what to click, read or share, Facebook explains.

The Recommendation Guidelines typically fall under Facebook’s efforts in the “reduce” area, and are designed to maintain a higher standard than Facebook’s Community Standards, because they push users to follow new accounts, groups, Pages and the like.

Facebook, in the new documentation, details five key categories that are not eligible for recommendations. Instagram’s guidelines are similar. However, the documentation offers no deep insight into how Facebook actually chooses how it chooses what to recommend to a given user. That’s a key piece to understanding recommendation technology, and one Facebook intentionally left out.

One obvious category of content that many not be eligible for recommendation includes those that would impede Facebook’s “ability to foster a safe community,” such as content focused on self-harm, suicide, eating disorders, violence, sexually explicit, regulated content like tobacco or drugs, content shared by non-recommendable accounts or entities.

Facebook also claims to not recommend sensitive or low-quality content, content users frequently say they dislike, and content associated with low-quality publishings. These further categories include things like clickbait, deceptive business models, payday loans, products making exaggerated health claims or offering “miracle cures,” content promoting cosmetic procedures, contest, giveaways, engagement bait, unoriginal content stolen from another source, content from websites that get a disproportionate number of clicks from Facebook versus other places on the web, news that doesn’t include transparent information about the authorship or staff.

In addition, Facebook claims it won’t recommend fake or misleading content, like those making claims found false by independent fact checkers, vaccine-related misinformation, and content promoting the use of fraudulent documents.

It says it will also “try” not to recommend accounts or entities that recently violated Community Standards, shared content Facebook tries to not recommend, posts vaccine-related misinformation, has engaged in purchasing “Likes,” has been banned from running ads, posted false information, or are associated with movements tied to violence.

The latter claim, of course, follows recent news that a Kenosha militia Facebook Event remained on the platform after being flagged 455 times after its creation, and had been cleared by 4 moderators as non-violating content. The associated Page had issued a “calls to arms” and hosted comments about people asking what types of weapons to bring. Ultimately, two people were killed and a third was injured at protests in Kenosha, Wisconsin when a 17-year old armed with an AR-15-style rifle broke curfew, crossed state lines, and shot at protestors.

Given Facebook’s track record, it’s worth considering how well Facebook is capable of abiding by its own stated guidelines. Plenty of people have found their way to what should be ineligible content, like conspiracy theories, dangerous health content, COVID-19 misinformation and more by clicking through on suggestions at times when the guidelines failed. QAnon grew through Facebook recommendations, it’s been reported.

It’s also worth noting, there are many gray areas that guidelines like these fail to cover.

Militia groups and conspiracy theories are only a couple examples. Amid the pandemic, U.S. users who disagreed with government guidelines on business closures can easily find themselves pointed towards various “reopen” groups where members don’t just discuss politics, but openly brag about not wearing masks in public or even when required to do so at their workplace. They offer tips on how to get away with not wearing masks, and celebrate their successes with selfies. These groups may not technically break rules by their description alone, but encourage behavior that constitutes a threat to public health.

Meanwhile, even if Facebook doesn’t directly recommend a group, a quick search for a topic will direct you to what would otherwise be ineligible content within Facebook’s recommendation system.

For instance, a quick search for the word “vaccines,” currently suggests a number of groups focused on vaccine injuries, alternative cures, and general anti-vax content. These even outnumber the pro-vax content. At a time when the world’s scientists are trying to develop protection against the novel coronavirus in the form of a vaccine, allowing anti-vaxxers a massive public forum to spread their ideas is just one example of how Facebook is enabling the spread of ideas that may ultimately become a global public health threat.

The more complicated question, however, is where does Facebook draw the line in terms of policing users having these discussions versus favoring an environment that supports free speech? With few government regulations in place, Facebook ultimately gets to make this decision for itself.

Recommendations are only a part of Facebook’s overall engagement system, and one that’s often blamed for directing users to harmful content. But much of the harmful content that users find could be those groups and Pages that show up at top of Facebook search results when users turn to Facebook for general information on a topic. Facebook’s search engine favors engagement and activity — like how many members a group has or how often users post — not how close its content aligns with accepted truths or medical guidelines.

Facebook’s search algorithms aren’t being similarly documented in as much detail.

 

 

Comments

Popular posts from this blog

Uber co-founder Garrett Camp steps back from board director role

Uber co-founder Garrett Camp is relinquishing his role as a board director and switching to board observer — where he says he’ll focus on product strategy for the ride hailing giant. Camp made the announcement in a short Medium post in which he writes of his decade at Uber: “I’ve learned a lot, and realized that I’m most helpful when focused on product strategy & design, and this is where I’d like to focus going forward.” “I will continue to work with Dara [Khosrowshahi, Uber CEO] and the product and technology leadership teams to brainstorm new ideas, iterate on plans and designs, and continue to innovate at scale,” he adds. “We have a strong and diverse team in place, and I’m confident everyone will navigate well during these turbulent times.” The Canadian billionaire entrepreneur signs off by saying he’s looking forward to helping Uber “brainstorm the next big idea”. Camp hasn’t been short of ideas over his career in tech. He’s the co-founder of the web 2.0 recommendatio

Drone crash near kids leads Swiss Post and Matternet to suspend autonomous deliveries

A serious crash by a delivery drone in Switzerland have grounded the fleet and put a partnership on ice. Within a stone’s throw of a school, the incident raised grim possibilities for the possibilities of catastrophic failure of payload-bearing autonomous aerial vehicles. The drones were operated by Matternet as part of a partnership with the Swiss Post (i.e. the postal service), which was using the craft to dispatch lab samples from one medical center for priority cases. As far as potential applications of drone delivery, it’s a home run — but twice now the craft have crashed, first with a soft landing and the second time a very hard one. The first incident, in January, was the result of a GPS hardware error; the drone entered a planned failback state and deployed its emergency parachute, falling slowly to the ground. Measures were taken to improve the GPS systems. The second failure in May, however, led to the drone attempting to deploy its parachute again, only to sever the line

How the world’s largest cannabis dispensary avoids social media restrictions

Planet 13 is the world’s largest cannabis dispensary. Located in Las Vegas, blocks off the Strip, the facility is the size of a small Walmart. By design, it’s hard to miss. Planet 13 is upending the dispensary model. It’s big, loud and visitors are encouraged to photograph everything. As part of the cannabis industry, Planet 13 is heavily restricted on the type of content it can publish on Instagram, Facebook and other social media platforms. It’s not allowed to post pictures of buds or vapes on some sites. It can’t talk about pricing or product selection on others.   View this post on Instagram   A post shared by Morgan Celeste SF Blogger (@bayareabeautyblogger) on Jan 25, 2020 at 7:54pm PST Instead, Planet 13 encourages its thousands of visitors to take photos and videos. Starting with the entrance, the facility is full of surprises tailored for the ‘gram. As a business, Planet 13’s social media content is heavily restricted and monito