Bumble’s algorithm will now report you for body-shaming

Bumble has stepped up its efforts to provide a safe dating space for users by banning body-shaming on the app. The app will use an algorithm to flag derogatory comments about someone’s appearance. “This includes language that can be deemed fat-phobic, ableist, racist, colorist, homophobic or transphobic,” the company said in a statement. Human moderators will then review the offending account. People who’ve use the language in their profile or on Bumble’s chat function will first receive a warning for their behavior. Repeated incidents or particularly harmful comments will result in a permanent ban. If the derogatory language has slipped through…

This story continues at The Next Web

Related Articles

How Bumble’s clever design helped the app go public

Lara Mendonça built Bumble’s product design team during the pandemic. Her strategy contributed to a wildly successful IPO.
Bumble went public on Thursday, and investors were very much interested in swiping right. The dating app, which soared in its first day of trading, is well-known for letting women make the first move. And in fact much of its success is due to this fundamental design decision.Read Full Story

Bumble files to go public

The dating and networking service Bumble has filed to go public. The company, launched by a former co-founder of the IAC-owned Tinder, plans to list its share on the Nasdaq stock exchange, using the ticker symbol “BMBL.” Bumble’s planned IPO was first reported in December. Whitney Wolfe Herd doesn’t care what she’s supposed to do Bumble […]

Responses

Your email address will not be published. Required fields are marked *

Receive the latest news

Subscribe To Our Weekly Newsletter

Get notified about chronicles from TreatMyBrand directly in your inbox