LOUNGE all new asksnapzu ideasforsnapzu newtribes interesting pics videos funny technology science technews gaming health history worldnews business web research entertainment food living internet socialmedia mobile space sports photography nature animals movies culture travel television finance music celebrities gadgets environment usa crime politics law money justice psychology security cars wtf art google books lifetips bigbrother women apple kids recipes whoa military privacy education facebook medicine computing wildlife design war drugs middleeast diet toplists economy fail violence humor africa microsoft parenting dogs canada neuroscience architecture religion advertising infographics sex journalism disaster software aviation relationships energy booze life japan ukraine newmovies nsa cannabis name Name of the tribe humanrights nasa cute weather gifs discoveries cops futurism football earth dataviz pets guns entrepreneurship fitness android extremeweather fashion insects india northamerica
  • ohtwenty

    The tools used to screen such [sexually explicit] comments, volunteer moderators told the BBC, haven’t been working properly for over a year, allowing between 50,000 and 100,000 “predatory” accounts to remain on YouTube.

    That's not disturbing at all. I mean it's not like Youtube is part of one of the biggest companies in the world, and could easily afford to fix an issue like this, right?

    Besides, as highlighted elsewhere, Youtube quickly claimed this was a tiny problem, whereas even just a quick search shows that this is an issue across a ton of accounts, and with a looooot more videos than the 150,000 they've deleted so far. If these accounts are able to programmatically make dozens of new videos a week (or a day), this feels like mopping up the ocean. Of shit.