Uganda Lawmakers Should Act As USA Congress Says Kids Aren’t Safe On Social Media

. USA Tech CEOs are back in DC to pledge (again) that they’ll handle it

By Our International


USA Congress grilled the chief executives of several big USA tech companies this week, including Meta CEO Mark Zuckerberg, about potential harms from their products on teens. Until now, the social platforms have largely had the same response: We’ll help teens and families make smart decisions themselves.

In Africa its even worse as there are no deliberate measures and regulations barring children from accessing adult content on their phones or learning gadgets like cheap tablet computers. Child safety advocates in Uganda have previously reported that children are exposed to unwanted nudity , outright pornography and criminal advances by adults for sex .

But now, in USA there is growing claims that social media can hurt young users, including worries that it risks driving them to depression or even suicide, online safety advocates say that response falls far short. And with a presidential election looming — and state lawmakers stealing the spotlight from their federal counterparts — Congress is set to press tech companies to go beyond the tools they’ve rolled out in the past.

The chief executives of TikTok, Snap, Discord and X are set to testified alongside Zuckerberg at Wednesday’s Senate Judiciary Committee hearing. For some, including X CEO Linda Yaccarino, Snap CEO Evan Spiegel and Discord CEO Jason Citron, Wednesday’s hearing marked their first-ever testimony in front of Congress.

Many of the tech CEOs are likely to use Wednesday’s hearing to tout tools and policies to protect children and give parents more control over their kids’ online experiences.

Some companies, such as Snap and Discord, told the media they plan to distance themselves from the likes of Meta by emphasizing they do not focus on serving users algorithmically recommended content in potentially addictive or harmful ways.

However, parents and online safety advocacy groups say many of the tools released by social media platforms don’t go far enough — largely leaving the job of protecting teens up to parents and, in some cases, the young users themselves — and that tech platforms can no longer be left to self-regulate.

“What the committee needs to do is to push these executives to commit to major changes, especially to disconnect their advertising and marketing systems from services that are known to attract and target youth,” said Jeff Chester, executive director of online consumer protection nonprofit the Center for Digital Democracy.

And the proliferation of generative artificial intelligence tools — which can give bad actors new ways to create and spread malicious content on social media — only raises the stakes for ensuring tech platforms have safety features built in by default.

Are the kids alright?

Several major platforms — including Meta, Snapchat, Discord and TikTok — have rolled out oversight tools that allow parents to link their accounts to their teens’ to get information about how they’re using the platforms and have some control over their experience.

Some platforms, such as Instagram and TikTok, also introduced “take a break” reminders or screentime limits for teens and tweaked their algorithms to avoid sending teens down rabbit holes of harmful content, such as self harm or eating disorder media.

This month Meta announced a proposed blueprint for federal legislation calling for app stores, not social media companies, to verify users’ ages and enforce an age minimum.

Meta also unveiled a slew of new youth safety efforts that included hiding “age-inappropriate content” such as posts discussing self-harm and eating disorders from teens’ Instagram feeds and stories; prompting teens to turn on more restrictive security settings on its apps; a “nighttime nudge” that encourages teen users to stop scrolling on Instagram late at night; and changing teens’ default privacy settings to restrict people they don’t follow or aren’t connected to from sending them direct.

Online safety experts say that some of the new updates, such as restrictions on adult strangers messaging teens, are welcome changes, but that others still put too much pressure on parents to keep their kids safe.

Some also say the fact that it has taken platforms years, in some cases, to make relatively basic safety updates is a sign the companies can no longer be trusted to regulate themselves.

“It shouldn’t have taken a decade of predators grooming children on Instagram, it shouldn’t have taken massively embarrassing … lawsuits, it shouldn’t have taken Mark Zuckerberg being hauled before Congress next week,” for Meta and other platforms to make such changes, said Josh Golin, executive director of nonprofit children’s safety group Fairplay.

For their part, Meta and other platforms have said they’re aiming to walk a fine line: trying to keep young users safe without too strongly imposing views about what content is or isn’t appropriate for them to view, and instead aiming to empower parents to make those judgment calls.

As efforts to rein in tech platforms have ground to a standstill on Capitol Hill, much of the momentum for regulating social media has picked up outside the halls of Congress.

In recent years, ArkansasLouisianaOhioUtah, and others have passed laws restricting social media for teens, in many cases by establishing a minimum age for social media use or by requiring a tech platform to obtain parental consent before creating accounts for minors.

Leave a Reply

Your email address will not be published. Required fields are marked *