"We always insist that our media providers adhere to the most stringent of precautions to ensure our brands do not appear next to inappropriate content". The restrictions and banishment of videos are very specific for religious extremists and Russian Federation or comedy skits that appear to show children being forcibly drowned.
Advertising experts noted that no major brand wants to be associated with pedophilia on YouTube no matter how small the audience for that content may be.
YouTube has terminated hundreds of accounts and removed more than 150,000 videos from its platform after multiple big-name advertisers pulled ads over disturbing content involving children.
HTC U11 Android 8.0 Oreo update starts rolling out
He didn't give any details on timing for the non-U11 devices, only going as far as saying that " It's coming soon ". If you're someone who doesn't live in the Americas, you should expect a separate update in the next few weeks.
But they appear to have allegedly drawn comments from people viewing the videos with sexual intent, creating a freaky situation where ad revenue is being pulled not because of the specific content of the videos, but because of who they were accidentally appealing to.
It said: "There shouldn't be any ads running on this content and we are working urgently to fix this". User speculations, which dubbed the event "ElsaGate", ranged from the content being procedurally generated by AI in an effort to manipulate view counts and thus rake in more ad revenue, to more complex claims linking the videos to government-funded mind control projects.
Within the span of 6 months, YouTube has eliminated over 50 user channels and stopped running ads on over 3.5 million videos. YouTube Go, a new app created to broaden the accessibility of the behemoth video-sharing service. The Guardian reported that 300 hours of video are uploaded every minute. They find it increasingly hard to tell the difference between inappropriate and appropriate content. Companies have three choices when placing their advertisements, according to the Wall Street Journal.
Campaigners have warned that pedophiles were targeting the videos posted on YouTube. The videos of young girls that attracted sexualized comments were not, on their face, sexual. "If we really want to block all content that violates the platform rules, then we would have to move to a model where platform users submit content they want to publish to an editor for approval, as we do when publishing in journals".