Moderating a Facebook gardening group in western New York is not without challenges. There are complaints of wooly bugs, inclement weather and the novice members who insist on using dish detergent on their plants.
And then there’s the word “hoe.”
Facebook’s algorithms sometimes flag this particular word as “violating community standards,” apparently referring to a different word, one without an “e” at the end that is nonetheless often misspelled as the garden tool.
Normally, Facebook’s automated systems will flag posts with offending material and delete them. But if a group’s members — or worse, administrators — violate the rules too many times, the entire group can get shut down.
Elizabeth Licata, one of the group’s moderators, was worried about this. After all, the group, WNY Gardeners, has more than 7,500 members who use it to get gardening tips and advice. It’s been especially popular during the pandemic when many homebound people took up gardening for the first time.
A hoe by any other name could be a rake, a harrow or a rototill. But Licata was not about to ban the word from the group, or try to delete each instance. When a group member commented “Push pull hoe!” on a post asking for “your most loved & indispensable weeding tool,” Facebook sent a notification that said “We reviewed this comment and found it goes against our standards for harassment and bullying.”
Facebook uses both human moderators and artificial intelligence to root out material that goes against its rules. In this case, a human likely would have known that a hoe in a gardening group is likely not an instance of harassment or bullying. But AI is not always good at context and the nuances of language.
Users often complain that they report violent or abusive language and Facebook rules that it’s not in violation of its community standards. Misinformation on vaccines and elections has been a long-running and well-documented problem for the social media company. On the flip side are groups like Licata’s that get caught up in overly zealous algorithms.
Licata said she never heard from a person and Facebook, and found navigating the social network’s system of surveys and ways to try to set the record straight was futile.
A Facebook representative said in an email this week the company found the group and corrected the mistaken enforcements. It also put an extra check in place, meaning that someone — an actual person — will check offending posts before the group is considered for deletion. The company would not say if other gardening groups had similar problems.