Starbucks CEO Howard Schultz revealed that they have been working on a very exciting new digital initiative that builds on our existing industry-leading digital platform in innovative ways.
New ‘Call of Duty: Modern Warfare 2’ information will be revealed at an esports event later this week
While Unilever is keen to embrace these new virtual worlds, it wants to mitigate the risks and concerns around web3 and the metaverse.
WhatsApp will block screenshots for View Once messages and will let users silently leave group chats
The Bundesliga champions aim to create personalized digital experiences in managing relationships with supporters globally to secure their long-term loyalty.
Apple has reportedly started production for its September product launch event for the next iPhone and Apple Watch
Every virtual NBA game is driven by actual NBA data, which Sportradar obtained as part of a US$1 billion streaming and betting data agreement with the league.
Facebook's own internal documents offer a blueprint for making social media safer for teens
Right at the time social media became popular, teen mental health began to falter. Between 2010 and 2019, rates of depression and loneliness doubled in the U.S. and globally, suicide rates soared for teens in the U.S. and emergency room admissions for self-harm tripled among U.S. 10- to 14-year-old girls. Social scientists like myself have been warning for years that the ubiquity of social media might be at the root of the growing mental health crisis for teens.
Yet when Facebook CEO Mark Zuckerberg was asked during a congressional hearing in March to acknowledge the connection between social media and these troubling mental health trends, he replied, “I don’t think that the research is conclusive on that.”
Just six months later, The Wall Street Journal reported that Facebook had been doing its own research for years on the negative effects of Instagram, the company’s photo-sharing app popular with teens and young adults. Six internal documents summarizing the research, leaked by a whistle-blower, were posted in full on Sept. 29, 2021.
The details in the 209 pages are revealing. They suggest not only that Facebook knew how Instagram could be harmful, but that the company also was aware of possible solutions to mitigate those harms. Facebook’s own research strongly suggests that social media should be subject to more stringent regulation and include more guardrails to protect the mental health of its users.
There are two primary ways the company can do this: enforcing time limits and increasing the minimum age of users.
A ticking time bomb for mental health
That’s important because many teens, especially girls, spend large amounts of time on social media.
One study in the U.K. found that one-quarter of 15-year-old girls spent more than five hours a day using social media – and 38% of those girls were clinically depressed. Comparatively, among girls who used social media less than one hour a day, only 15% were depressed.
Although the internal Facebook research didn’t examine links between time on Instagram and mental health, they did ask teens about what were, in their view, the worst aspects of Instagram. One of the things teens disliked the most about the app was how much time they spent on it.
Teens, the report said, had “an addict’s narrative about their use. … They wish they could spend less time caring about it but they can’t help themselves.”
They knew they were spending too much time online, but had a hard time controlling how much time they spent. One-third of teens suggested Instagram should remind them to take a break or encourage them to get off the app.
That would be a step in the right direction, but simple nudges might not be enough to get teens to close the app and keep it closed. And while parents can already set time limits using the parental controls included on most smartphones, many of them don’t know how to use these controls or are unaware how much time teens are spending on social media.
So better regulations might need to put teeth into time limits, such as limiting the number of hours teens under 18 can spend on social media apps. A blackout period overnight might also be useful, as many teens use their smartphones at night when they should be sleeping.
One internal Facebook study of more than 50,000 people from 10 countries found that half of teen girls compare their appearance to others’ on Instagram. Those appearance-based comparisons, the study found, peaked when users were 13 to 18 and were much less common among adult women.
This is key, as body image issues seem to be one of the biggest reasons why social media use is linked to depression among teen girls. It also dovetails with research I reported in my book, “iGen,” finding that social media use is more strongly linked to unhappiness among younger teens than older ones.
This suggests another avenue for regulation: age minimums. A 1998 law called the Children’s Online Privacy Protection Rule already sets the age minimum for social media accounts at 13. That limit is problematic for two reasons. First, 13 is a developmentally challenging time, right as boys and girls are going through puberty and bullying is at its peak.
Second, the age minimum is not regularly enforced. Kids 12 and under can simply lie about their age to sign up for an account, and they’re rarely kicked off the platform for being underage. During a Facebook event with Instagram head Adam Mosseri, the young celebrity JoJo Siwa noted she had been using Instagram since she was 8 years old, forcing Mosseri to acknowledge that it’s easy to lie about your age.
The problem is how to enforce an age limit online for a population that is too young for IDs. Raising the minimum age to create a social media account to 16, 17 or 18 could solve two problems at once: It would prevent kids from signing up until they’re a bit more developed and mature, and it would be easier to enforce. For example, potential users might be asked to submit a photo of their state-issued ID, which most teens have by 16.
Verifying age would also make it easier to construct a safer app for younger users that might, say, hide follower counts or restrict access to celebrity accounts, both of which Facebook’s research found negatively impacted girls’ body images.
Curtailing that fear of missing out
It’s tempting to think regulations like these would cause teens to riot in the streets – after all, they love keeping up with their friends on social media. But the teens interviewed by Facebook for its internal research were well aware of social media’s downsides.
“The reason why our generation is so messed up and has higher anxiety and depression than our parents’ is because we have to deal with social media. Everyone feels like they have to be perfect,” one teen girl told the researchers. Other teens have spoken publicly about the negative effects of social media.
More stringent regulation would help with another issue teens know all too well: the unwritten mandate to use social media or be left out.
“Young people are acutely aware that Instagram can be bad for their mental health yet are compelled to spend time on the app for fear of missing out,” Facebook’s internal research concluded.
If age limits were enforced, the peer pressure of being on social media would vanish; no or few classmates would be there. Regulating time on the app could also help if teens knew their friends wouldn’t constantly be online.
Until they do, it will be up to parents and teens themselves to set limits. That won’t be easy, but teens will be safer for it.