I saw this article, which made me think about it…

Kids under 16 to be banned from social media after Senate passes world-first laws


Seeing what kind of brainrot kids are watching, makes me think it’s a good idea. I wouldn’t say all content is bad, but most kids will get hooked on trash content that is intentionally designed to grab their attention.

What would be an effective way to enforce a restriction with the fewest possible side effects? And who should be the one enforcing that restriction in your opinion?

  • @Dave@lemmy.nz
    link
    fedilink
    19
    edit-2
    8 hours ago

    I can’t remember which article I was reading, probably one on Lemmy, but it said that we know social media algorithms are bad for people and their mental and physical health, that they are divisive, drive extremism, and just in general are not safe for society.

    Drugs are regulated to ensure they are safe, so why aren’t social media algorithms regulated the same way? Politicians not understanding the technical details of algorithms is not an excuse - politicians also don’t understand the technical details of drugs, so they have a process involving experts that ensures they are safe.

    I think I’m on the side of that article. Social media algorithms are demonstrably unsafe in a range of ways, and it’s not just for under 16s. So I think we should be regulating the algorithms, requiring companies wishing to use them to prove they are safe before they do so. You could pre-approve certain basic ones (rank by date, rank by upvotes minus downvotes with time decay like lemmy, etc). You could issue patents to them like we do with drugs. But all in all, I think I am on the side of fixing the problem rather than pretending to care in the name of saving the kids.