Protecting Children

Kids and Social Media


Jhonatan Jimenez


A primary issue regarding protecting children on social media is exposure to

inappropriate content.  Privacy risks are also critical, since minors may unknowingly

share personal information that can be exploited by people with bad intentions.

Furthermore, algorithm-driven content recommendations can sometimes amplify

harmful or age-inappropriate material.

A way we can help mitigate these issues is by having platforms implement stronger

age verification systems and enforce stricter content moderation policies. Parental

controls and default privacy settings for minors can help limit exposure to harmful

interactions. Finally, increased transparency and accountability from social media

companies can help improve safer online environments for young users.


Mehnaz Barsha


One of the biggest threats to children online today is the ease at which predators

can reach them through voice and chat features on social media and gaming platforms.

Unlike text messages that can be reviewed later, live voice conversations disappear the

moment they end, giving parents almost no visibility into who their child is speaking

with or what is being said. Predators exploit this by pretending to be other kids and

being friendly to gradually build trust before pushing boundaries.


While some platforms like Roblox have made changes to better protect younger users,

they do not go far enough. Most platforms were built around keeping users engaged,

not keeping them safe, and updating a few features does nothing to change that problem.


To help address this, children need to be taught what manipulation actually looks like

in practice, things like an adult who seems overly interested in them, pushes to keep

conversations private, or tries to move the conversation to another app. Teaching kids

to recognize grooming early can go a long way in keeping them safe. The Kids

Online Safety Act is a step in the right direction because it holds tech companies

accountable rather than leaving the responsibility on parents and children to figure

it out on their own.


Dylan Moutinho


A major concern in protecting children on social media is their exposure to

cyberbullying, inappropriate content, online predators, and privacy risks.

Many children also often share personal information without understanding

the long-term consequences, which can impact their safety and mental health.

This has especially been a problem for ROBLOX, which in recent years has

been exposed for having hundreds of predators on its platform and there are

little to no safety features to stop it.


To reduce these risks, parents should monitor online activity and encourage

open conversations about digital safety. Social media platforms should strengthen

age verification, privacy settings, and content moderation. Schools can also support

digital literacy education to help children recognize and respond to online threats.



Nadia Brown


The largest issue affecting children’s social media usages is cyberbullying.

According to a national survey conducted by the Cyberbullying Research Center

of 3,466 students ages 13-17 58% reported experiencing cyberbullying.

Cyberbullying can happen by text, email, instant messaging, and social media;

however, social media is the largest contributor. Unlike traditional forms of bullying,

cyberbullying can happen anywhere. Cyberbullying occurs everywhere, but Instagram

is the largest contributor.

 

  

 

The most common forms of social bullying reported by adolescents include:

  • 56% hurtful comments

  • 53% exclusion

  • 53% online rumors  

  • 50% embarrassment or humiliation

  • 42% repeated unwanted contact via text or online

  • 38% direct threats through text message or direct messages

 

Cyberbullying is an international public health concern because of its impact on the

mental health of adolescence. Targets of cyberbullying report increased depressive

thoughts, anxiety, loneliness, and suicidal behavior. Perpetrators of cyberbullying are

more likely to report increased substance abuse, aggression, and delinquent behavior.

 

To reduce cyberbullying, social media companies can implement AI to detect bullying,

harassment, and offensive content or add new features to reduce unwanted attention or

comments.  For example, Facebook uses AI to moderate content. If a post, comment,

or story goes against their Community Standards, then it is removed from Facebook.

Content that does not violate the Community Standards but is still questionable goes

on to a team of human reviewers.

 

 

Another example is Instagram’s Comment Warning feature. When a user tries to post

a potentially offensive comment. The user is reminded of Instagram’s community

guidelines. This helps because it lets the user know that their comment may be removed

or hidden if they attempt to proceed. Additionally, Hidden Words provides users the

ability to create a custom word list to reduce the chance of unwanted comments on their

page. Moreover, if those features are not enough, users have the option to restrict

someone, limit unwanted interaction from another user for a period, or block them

entirely.  


Charles Murphy


 A few issues to consider regarding the subject of protecting children in social media,

are inadequate age verification protocols, no mandated parental control infrastructure,

and lack of a technical roadmap for enforcement and oversight. These are facets of the

social media algorithm that don’t exist, making the algorithm inherently suited to your

likes, habits and rituals without the need for boundaries or regulation. The algorithm

of social media does not handle the age of the user well and will autonomously suggest

anything that may get a positive reaction. Within this ‘normal’ content, an occasional

inappropriate item may be shown to a minor user. Furthermore, a minor can lie about

their age easily with any sufficient device level solutions such as age verification in place.

All of this, of course, presents itself as a problem.

 

Some proposed solutions suggest there should be child-specific versions of social media

that host an array of child-friendly features, and an algorithm tailored to children.

A device level solution would be to have a child-mode smartphone with mandatory

parental control dashboards, activity / content monitoring, and emergency lockdown

features. Some of these solutions can then be encapsulated into laws to govern the moral

imperative of protecting children online, which in turn changes the way we think about

and utilize social media.


Jun Li Lin


Child protection online has always been quite the huge topic, the most recent well

known one I know is the Roblox Schlep controversy. Where the Youtuber Schlep a

youtube channel known for conduction sting operations against child predators on the

platform was sent a cease and desist letter threatening to take legal action against him

if he continues. After the whole controversy finishes Roblox began implementing age

verification onto the platform, which does not help at all to protect the children since

they would just take their parents ID cards and use that to have access to Roblox chat.

The problem with age verification is that on day one of it being implemented there was

already sellers Ebay selling accounts with pre verified accounts, that combined with the

fact that anyone can fake how old they are on the platform does not help with the

problem, instead this made it easier for predators to message children since kids who

faked their age is now being placed with people who are 18+ and also for predators to

buy fake accounts are now being placed with kids. 


Child protection online is quite difficult to tackle, the main way I can think of protecting

kids is to teach them in school about online safety and also implementing chat

moderators for kids games, every method implemented has a way for bad actors to get

around it so teaching both parents and their children awareness is the best way to protect

them online. 


Elsa Shaikh

One big issue regarding protecting children on social media is their exposure to harmful

content, cyberbullying, and addictive platform features that can negatively impact their

mental health. The algorithms used in social media apps tend to prioritize engagement

over safety, which further increases these risks.

To mitigate these issues, platforms should improve content moderation as well as age

verification. Parents can also use parental controls and limit screen time. Lastly, stronger

policies focused on child safety are also essential.

Soo Hee Min


 One of the biggest challenges in protecting children from social media is that excessive

use can lead to decreased academic engagement and negative mental health outcomes.

Indeed, a large-scale study of Finnish adolescents found that students who used the

internet excessively were at a higher risk of school absence, demonstrating that online

use can impact academic engagement.


https://www.theguardian.com/education/2024/apr/16/teenagers-who-use-internet-to-excess-more-likely-to-skip-school


    To address this issue, I believe realistic adjustments to the usage environment are

needed, rather than relying solely on individual self-control. For example, some

American schools require students to keep their phones in locked pouches or classroom

lockers during class hours to improve concentration. Social media platforms could also

implement practical features, such as automatic pauses after a certain amount of

continuous use or reduced notification frequency, for minors' accounts.


    I believe that a combination of school-level mobile phone management and

improved platform design is a more realistic solution than relying solely on individuals.



David Lin


The most important difference between an adult and children consuming content online

is the fact that children are oftentimes exposed to the internet at a young age when their

brains are still undergoing a period of critical development. A question I thought of

when researching was whether there are any hidden benefits of screen time we can

leverage to benefit the development of children. It turns out, there aren't many pros of

giving children access to the internet. According to an OSF Healthcare article,

pediatricians recommend 0 screen time to those under 2, no more than 1 hour for those

from 2-5, and no more than 2 hours for those 5-17. It shocked me to find the

recommendation to be so low because we know that the reality is that exposure to the

internet is much greater for most children. I would argue the benefits of being online act

as a double-edged sword. During a period of time between the ages of 9-18 when

children develop and have a longing for social interaction and belonging, social media

fills this void in the worst ways. As we discussed last week, algorithms are designed to

be addictive, and the first topic we explored taught us that misinformation and echo

chambers are all too common in the internet age. Nowadays, many teenagers and even

some younger children face a bigger threat than dangerous people online looking to

manipulate minors, or even the harmful effects of too much screen time. A.I. Related

suicides is a topic that has become increasingly popular for people of all ages but

especially children. It may seem unusual for those of us that use A.I.  as a tool for work

to see responses that encourage suicides but as Sanford states in a Stanford Medicine

article, "One key difference is that the large language models that form the backbone

of these companions tend to be sycophantic, giving users their preferred answers.

The chatbot learns more about the user’s preferences with each interaction and responds

accordingly. This, of course, is because companies have a profit motive to see that you

return again and again to their AI companions. The chatbots are designed to be really

good at forming a bond with the user". The advancement of the amount of context a

particular chat bot can hold in their memory has led to an increase in the number of

people who seek these chat bots for comfort and companionship. As Sanford suggests,

chat bots are designed to tell you what you want to hear, whether that be a correct

response to a homework question or to sooth you when going through a rough time.

This is dangerous for children who are developing their prefrontal cortex with certain

mental health issues exacerbated by being chronically online to begin with. In terms of

solutions to these problems, in an article in Stanford Medicine titled Screen time:

The good, the healthy and the mind-numbing , Armitage says:

    "We recently published a study that followed kids from as young as 7 to as old as

15 as they received their first phone. We found, on average, getting a phone at a younger

age was no better or worse than getting it at an older age, in terms of depressive

symptoms, sleep and school grades. (...) I recommend parents wait to give their child a

phone until they are mature enough to regulate their own use and not allow it to distract

them from sleep, homework, family time, and playing and socializing with other kids in

the real world"

At the end of the day, maturity of the child as well as the content that is being consumed

in question creates unique cases for each child that needs to be addressed individually.

Although parents may not hit the mark perfectly each time, a good rule of thumb is

decreasing screen time overall which will benefit children in the long run. 



Samuel Emile


After reviewing the readings on protecting children online, one major issue that stood

out to me is how social media use is linked to mental health problems among young

people. Research shows that heavy use of these platforms is associated with higher rates

of anxiety, depression, and even self-harm, especially among younger teens. One concern

is that algorithms often push emotionally intense or harmful content because it increases

engagement, even if it negatively affects users. Another issue is that current laws and

protections are inconsistent. Some states and countries require parental consent or age

limits, but enforcement is often weak because platforms do not have strong verification

systems. This makes it easy for younger children to access content that may not be

appropriate or safe. To help address these problems, I think several best practices could

help. First, platforms should be required to design youth-specific versions with safer

features, such as time limits, better reporting tools, and reduced algorithmic

recommendations. Second, stronger age verification methods combined with parental

authorization could prevent younger children from creating accounts without supervision.

Finally, device-level parental controls, like monitoring dashboards and usage limits,

could give parents more practical tools to guide their children’s online activity.

Overall, protecting children on social media will require cooperation between lawmakers,

technology companies, and families. Clear regulations combined with safer design

practices can help reduce risks while still allowing young people to benefit from digital

communication.


Zahra Qureshi


Social media platforms present serious risks to children, including exposure to harmful

or inappropriate content, cyberbullying, online predators, and addictive algorithm-driven

feeds that can negatively impact mental health. Many platforms rely on weak age

verification systems, allowing underage users to bypass restrictions easily. Additionally,

children often lack awareness of privacy risks, leading them to overshare personal

information that can be exploited. These factors collectively create an online

environment that can be unsafe and psychologically damaging for young users.


To mitigate these issues, stronger age verification systems and safer default privacy

settings should be implemented by platforms. Parents and guardians should use available

parental controls and maintain open communication with children about their online

experiences. Schools can also promote digital literacy education to teach children how

to recognize harmful content, protect their privacy, and report suspicious behavior.

A combined effort from platforms, policymakers, educators, and families is essential

to ensure a safer digital environment for children.


Kevin Dias


Protecting children is a difficult thing to do on social media, but I don't think the ways

we are going about it currently are good for the long term. The requirement that everyone

provides identification and facial confirmation when signing up for social media is a

major privacy violation, and it actually makes it worse for children and people in general

because these third party confirmation services typically get hacked. For example, the

company that did Discord's ID verification was just hacked, and all of these photos of

people and their IDs were leaked onto the internet for anyone to download. This is not

the way to do things. I don't even think children should be on social media anyways,

I think it's something that could seriously stunt a child's social development.


Viktor Hreskiv


I believe protecting children on social media is very important. I am concerned about

cyberbullying, harmful content, online predators, mental health problems, and

gambling-like features in some video games. I think many platforms do not do enough

to check ages or reduce addictive features.


In my opinion, companies should use stronger age verification, better content

moderation, and stricter privacy settings for minors. I also believe there should be

clearer rules about gambling-style game features in games that are primarily played

by young people. I think parents and schools should teach children more about online

safety and monitor their activity.


I believe protecting children online requires companies, parents, schools, and

policymakers to work together in order to combat this.


Ahmed Abdulghany


After reviewing the materials, I realized how complicated the issue of protecting

children on social media really is. There are clear mental health concerns.

The Columbia Undergraduate Law Review article discusses research showing a

connection between heavy social media use and increased depression and anxiety

among teens. That trend is concerning, especially given how much time young people

spend online today.

At the same time, broad state bans such as those passed in Utah and Florida raise

constitutional concerns. Cases like Brown v. Entertainment Merchants Association and

Pierce v. Society of Sisters suggest that minors do have First Amendment protections

and that parents, not the government, generally have the primary role in directing their

children’s upbringing. Completely restricting access may conflict with those principles.

Another issue is that many laws focus more on limiting access than on improving

platform safety. Connecticut’s SB00003 strengthens privacy protections for minors

and addresses online safety, but it does not fully solve practical enforcement issues

such as reliable age verification or consistent parental control systems across platforms.


In my view, a more realistic solution would focus on safety measures rather than outright

bans. Social media platforms could develop youth specific versions with time limits,

simplified content feeds, and stronger reporting tools. Stronger age verification combined with clear parental consent systems could also improve accountability. In addition, device level parental controls, like those available through Google Play, give parents practical tools to manage content without removing access entirely.

Overall, protecting children online requires cooperation between lawmakers, technology

companies, and families. A balanced approach that strengthens safeguards while

respecting parental authority and constitutional rights seems more effective than

broad prohibitions.


Fabricio Miranda


The biggest issue I see regarding children's safety is online harassment/ cyberbullying.

I see it very often with my niece and her playing Roblox. Sometime ago she would

consistently come up to me and tell me that a person on Roblox was being mean to her

and saying mean things, sometimes the things would be hash tagged so you couldn't see

the actual message but from context you can tell how bad it was. Luckily she wouldn't

talk back to the people she would just ignore it. This can get increasingly bad if the

person being cyberbullied doesn't know how to block or remove someone, the person

can continue to join the games you are in and continue to harass you, this can create

stress in a place where you should be having fun playing with friends or by yourself.

This is also prevalent in any other type of social media platforms a child may have access

to. A child can innocently make a comment on a short they watched then people can

reply hateful stuff to it. The best practice I have found is a simple one that has been

around for a while which is to simply block the person harassing you. I showed my niece

how to do so and now she doesn't ask me to do it for her anymore instead she'll tell me

that she herself blocked someone who was being mean to her. It is simple yet effective.


Matthew Fletcher


I believe there are several major concerns when it comes to protecting children online.

One of the most serious issues is protecting children from predators. A big part of the

problem is that many children are not fully taught how unsafe the internet can be.

In many cases, parents may not completely understand how online predators operate.

People often talk about teaching kids to avoid the “creepy” stranger on the internet, and

while that advice is important, it oversimplifies the issue. Most online predators do not

appear creepy at first. Instead, they present themselves as kind, supportive, and

trustworthy individuals in order to build relationships with children. This process, known

as grooming, allows them to gain a child’s trust before manipulating them.

Another major concern is the impact of the internet on children’s mental health. There

has been a noticeable rise in mental health struggles among young people, and social

media plays a significant role. Many children experience cyberbullying simply because

they are young or vulnerable. This constant exposure to negativity and comparison can

contribute to increased rates of depression and anxiety among children and teenagers.

A third issue involves the information that children share online and, more importantly,

what parents share about their children. Some parents post pictures or videos of their

children for entertainment, attention, or views, without fully considering the long-term

consequences. In doing so, they may unintentionally contribute to the risks children face

online. A major example of this is the rise of “family vloggers” on platforms like

YouTube. These parents record and post detailed videos about their children’s daily lives.

For example, a video titled something like “My Daughter Had Her First Kiss” may be

partially staged for views, yet it still exposes deeply personal moments. Content like this

can attract the wrong kind of attention and create material that predators may misuse.

In my opinion, better protection for children online requires stronger education and

possibly new regulations. There should be laws or programs in place that focus on

teaching both parents and children about internet safety, with particular emphasis on

educating parents. Many of the problems mentioned stem from a lack of awareness and

responsible decision-making by adults. I also believe there should be stronger age

verification systems online. However, implementing such systems may be difficult

because many people are concerned about increased data tracking and privacy issues.

Finally, platforms such as YouTube and Facebook should implement more advanced

screening processes for videos that feature children. These systems should ensure that

no inappropriate or exploitative content involving children is being posted.

Overall, protecting children online requires awareness, responsibility, and cooperation

between parents, platforms, and policymakers.


David Flores


One of the issues when it comes to Protecting Children in social media would be the

use of collecting data to verify one’s age. There are two main examples, Discord and

Roblox, both pushing out age verification methods. First, this might be a great step

for protecting children till we understand how they are implementing this security

feature. For instance, Discord’s age verification is raising privacy concerns about

how one verifies one's own age. Doing a face scan, uploading an ID, or letting their Age

Inference Model determine the user's age.

    The problem here is that Discord had recently had a data breach in September 2025,

showcasing how providing sensitive data to them might lead to it being leaked or used

as ransomware. Another concern would be surveillance where their use of an "age

inference model" would determine the user's age based on user's behavior patterns

which showcases that Discord is already collecting and analyzing personal data; with

this in mind, there could be AI bias within the model. 



Jude Duperval 



There seems to be many issues with regards to protecting children in social media,

and this in part can be explained by their susceptibility towards novelty-seeking, and

anticipation with poor-assessment of long-term consequences. Despite a child possibly

being lectured on what should and shouldn't be done, many are prone to naivety due to

a lack of life experience, and discernment. So unfortunately, some messages may only

be understood by the child when it’s ‘too late’. This applies not only to social media and

its possible effects, but to life more broadly.  

 

On social media, it’s not unusual for one to enter a state of continuous comparison, join

certain spaces which may narrow one’s way of thinking, constantly chase emotionally

charged and exhilarating content, and much more (as discussed last class). When you

apply this understanding to the context of a child, it's of no surprise that we see data

reports stating that children are showing higher rates of self-hatred, self-harm,

depression, and even suicide. Their expected naivety has an affinity towards these

avenues both because of the way it adheres to their psyche, but also because of the

existence of strangers trying to capitalize on their vulnerability. 

 

Mitigating these issues is much more complicated than merely giving a lecture to

children on the risks behind the use of social media. Of course, one could promote

more parental control tools, variations of social media specifically designed for children,

and much more, but this is a temporary solution. A long-term solution would require

stricter legislative initiatives alongside tech and its continued ‘advancements’,

for this is a recurring theme that is a result of morally grey areas existing. Nevertheless,

the promotion of proactiveness instead of reactiveness to parents is a good step forward,

for children need guidance, not mere instructions. 


William Socci


Children should be the most protected against technology. They are just learning how to

function as humans and how to properly behave when it comes to social and personal

interactions. Studies have shown that technology has negatively impacted these parts of

teenagers' lives. So if teenagers and adults aren't safe how do we think an even younger

mind will react? When children are exposed to technology for long periods of time they

are subject to the mind numbing, anxious, and depressive feelings that no one should be

dealing with. We have created a generation of iPad kids even though all of our best

memories come from playing outside and socializing. However, since it calms the kids

down to a point they are barely functioning and become addicted, parents utilize this as

like it's a reward for good behavior even though we're just feeding the addiction as well

as treating it like a reward for the parents because they don't have to deal with the

changing emotions of a child.

It is difficult to mitigate this because we are in a world of ever evolving technology so

not exposing the children to it early is almost like holding them back as other children

will have a better understanding of its working at a younger age.  However, because

it's not being used properly as a teaching device I believe there should almost be

negligence laws and strict time restrictions on all devices especially when a child is

sitting in front of the screen.


Dante Prutt


This page talks about how social media over usage ties directly to depression and

anxiety among teens and that most platforms lack and sort of protection against these

problems. Solutions for this kind of problem include child version social media

platforms forcing more parental guidelines and adding further education on how to

assist a teen facing depression due to these apps


Olivier Jean Pierre



Protecting children on social media is a real challenge, as children are vulnerable to

harmful content online, cyberbullying, online predators, and social comparison. Social

media platforms use addictive algorithms that encourage endless scrolling and constant

notifications, which can make it difficult for children to control their screen time on

social media. Children are also exposed to privacy risks, as they can unintentionally

disclose personal information which may endanger them or someone else. Furthermore,

current age verification mechanisms are poor and face id raises privacy concerns for the

user and as a result make it easy for children to register on  any platforms. Finally,

companies are more motivated by maximizing engagement than by ensuring safety since

that's where the revenue comes from.


Some ways to combat these problems is to limit notifications, place screen time

reminders, perhaps parental controls and perhaps online safety education (which I

think is most important since there are many schemes people play online) and don't

reveal personal information (if you do let it be fake)


Brent Aguilar 


From what I have learned, protecting children online comes from a mixture of education,

monitoring, and control. Children face risk on social media platforms that can include

instances like Cyberbullying, Predatory behavior, over exposure to harmful or

endangering content and many more. 


A step in the right direction and possibly the most important would-be education.

Teaching children about the dangers of social media and how to not allow themselves to

be put at risk is vital for protecting kids. Informing them about scams, encouraging them

to report uncomfortable behavior, and setting clear rules on what can and cannot be posted

online are just a few examples of what can be included in this education. 


Another way to protect children would be to monitor and control the media that is being

posted and consumed. Certain software's such as BARK or Kid guard are examples of

ways to monitor media and control what is being allowed into the children's platforms.

It also helps monitor messages received on social media websites by forwarding the

messages to an administrator before it even reaches the child. 


Overall, from what I have read, I believe that education is the most effective practice

to mitigate these issues. Although monitoring and controlling is a viable option, it is

almost impossible, in this day and age, to completely control the media someone else

is consuming. So, it would be better to educate children on these topics, so they are

prepared as much as possible. 

Filipe Rodrigues

One issue with protecting children on social media is that apps like TikTok do not have

strong age verification. It is very easy for someone underage to lie about their age and

create an account. This gives kids access to content and features that may not be

appropriate for them, and it also allows minors to post things publicly that could put them

at risk.


I think platforms should have much better age verification, possibly using some form of

official ID or a government supported system. Right now, the current system relies too

much on honesty, which does not really work. Stronger verification would help enforce

age restrictions and make these platforms safer for younger users. Social media companies

should take more responsibility in enforcing their own rules and limiting access where

appropriate. 


Marlon Yaucan

Issues that I find regarding protecting children in social media are addictive algorithms.

Social Media platforms are designed to keep children glued to their screen and maximize

their engagement. This persistent stimulation can lead to major mental health issues.

Some include increased anxiety, depression, body image issues, and chronic sleep

disruption. Another issue I have is safety and exploitation. Social media also provides

a landscape where predators can use various tactics to lure or exploit minors. 

The best practices to mitigate these issues are strict privacy and parental controls.

By setting profiles to private by default, disable location tracking, and using platform

family pairing features, you can protect children's physical safety and allow parents to

manage their accounts. You can also establish clear, agreed upon rules. You can have

device-free zones like the dinner table or bedrooms and setting daily time limits to

prevent doom scrolling.

Misael Mercado


When you're working with these kids, it’s helpful to frame the "no social media until high

school" approach as a developmental bridge rather than a punishment. Middle school is

already a chaotic period for brain development and self-esteem; adding an algorithmic

highlight reel into that mix is often just too much for them to process at once. If a student

is ready for some access, think of restrictions and time limits as digital "training wheels it’s about giving them a safe, small space to practice before they hit the open road. For the

kids who are already deep into it, the strategy shifts: you must sit down and have real,

eye-level conversations. Instead of lecturing, talk to them about the "why" behind the

threats explain how algorithms are designed to keep them scrolling even when they’re

unhappy, and how predators or scammers look for specific vulnerabilities. It’s not about

scaring them; it’s about giving them the "intel" they need to protect themselves in a

world that isn't always looking out for them.


From Gemini


This collection of student responses highlights the multifaceted dangers children face on social media and proposes various strategies for mitigation. The consensus is that current protections are insufficient, requiring a collaborative effort between tech companies, parents, and lawmakers.

Below is a categorized summary of the core issues and proposed solutions identified across the contributions.


1. Primary Risks and Issues

The students identified several recurring threats to minors in digital spaces:

  • Predatory Behavior & Grooming: Many students (Barsha, Moutinho, Fletcher) emphasized how predators use voice/chat features to build trust and "groom" children, often moving conversations to private apps.

  • Mental Health & Development: Excessive screen time and addictive algorithms are linked to anxiety, depression, self-harm, and decreased academic engagement (Min, Emile, Duperval).

  • Cyberbullying: Brown highlights that 58% of students have experienced cyberbullying, with Instagram being a primary source of hurtful comments and exclusion.

  • Content Exposure: Algorithms prioritize engagement over safety, often recommending inappropriate, harmful, or "gambling-like" content to minors (Jimenez, Hreskiv).

  • Privacy & Data Risks: Children often unknowingly share personal data. Paradoxically, some solutions like ID-based age verification create new risks, as third-party verification services are vulnerable to data breaches (Flores, Dias).


2. The "Roblox" Case Study

Multiple students (Moutinho, Lin, Miranda) cited Roblox as a specific example of these challenges:

  • It has been criticized for harboring predators.

  • In-game harassment is common and can be difficult for children to manage without blocking tools.

  • Attempts at age verification have been circumvented by children using parents' IDs or buying pre-verified accounts on eBay.


3. Proposed Solutions

The responses suggest a "layered" approach to safety:

Technical & Platform Changes

  • Youth-Specific Versions: Creating versions of apps with restricted algorithms, no "infinite scroll," and reduced notifications.

  • Enhanced Moderation: Using AI to detect bullying and "Hidden Words" filters to block harassment (Brown).

  • Default Privacy: Setting all minor accounts to "private" by default and limiting who can contact them.

Educational & Parental Strategies

  • Digital Literacy: Teaching children to recognize the signs of grooming and manipulation (Barsha, Aguilar).

  • Parental Involvement: Using monitoring tools (like Bark or Kid Guard) and maintaining open communication rather than just "lecturing."

  • School Policies: Implementing phone-free zones or "locked pouches" during school hours to improve focus (Min).

Regulatory Action

  • Legislative Oversight: Support for the Kids Online Safety Act (KOSA) to hold tech companies legally accountable for safety (Barsha).

  • Mandatory Standards: Establishing laws that require consistent parental control dashboards and stricter age verification across all platforms.


Comparison of Perspectives

Perspective

Focus Area

Key Argument

Strict/Regulatory

Legal Accountability

Government must step in because platforms prioritize profit over safety (Barsha, Murphy).

Educational

Digital Literacy

Tech changes daily; the only permanent solution is teaching kids to recognize threats (Aguilar, Lin).

Privacy-Conscious

Data Security

Age verification via IDs is a privacy nightmare and can lead to identity theft (Dias, Flores).

Hands-On Parental

Practical Monitoring

Parents should use locking pouches, monitor dashboards, and teach "blocking" skills (Min, Miranda).


Based on the diverse concerns and solutions raised by the contributors, here is a comprehensive policy recommendation designed to balance safety, privacy, and the developmental needs of young users.


Policy Recommendation: The "Safety by Design" Framework

To effectively protect children, we must move away from a "reactive" model and toward a proactive, layered infrastructure that shares responsibility between platforms, parents, and the state.

1. Platform Responsibility & Algorithmic Reform

Social media companies must be legally required to alter the fundamental architecture of accounts held by minors.

  • Safety-First Algorithms: Disable "engagement-based" recommendations for users under 18. Instead, feeds should be chronological or limited to verified "safe" educational/age-appropriate content.

  • Friction by Design: Implement "automatic pauses." After 60 minutes of continuous use, the app should lock for a mandatory 15-minute "cool-down" period to combat addiction.

  • Enhanced AI Moderation: Mandatory integration of AI sentiment analysis to detect and flag grooming patterns or cyberbullying in real-time, providing immediate prompts to the child (e.g., "Is this person making you uncomfortable?").

2. Privacy-Preserving Age Verification

To address the valid privacy concerns regarding ID leaks, platforms should move away from centralizing sensitive documents.

  • Zero-Knowledge Proofs (ZKP): Adopt verification technologies that confirm a user is over/under a certain age without actually storing or even "seeing" the underlying ID document.

  • Device-Level Verification: Shift the burden of age verification to the device/operating system level (Apple/Google) rather than individual apps, reducing the number of companies holding a minor's data.

3. Empowerment Through Education (Digital Literacy)

Legislative funding should be allocated to integrate "Digital Self-Defense" into K-12 curricula.

  • Grooming Awareness: Teach children specifically how predators build trust and the "red flags" of manipulation (e.g., moving to a private app, asking for secrets).

  • The "Block & Report" Culture: Normalize the use of safety tools so children feel empowered to terminate uncomfortable interactions without fear of social repercussion.

4. Transparent Parental Oversight

Platforms must provide a standardized "Parental Dashboard" that is intuitive and universal.

  • Visibility Without Intrusion: Provide parents with high-level data (who the child is talking to, time spent) without necessarily infringing on the child's private text content, unless flagged by safety AI.

  • Control over "Loot Boxes": Strictly regulate or ban gambling-like features (loot boxes, virtual currency) in games frequented by minors to prevent early-onset financial manipulation.

5. Accountability & Enforcement

  • The Kids Online Safety Act (KOSA) Support: Create a legal "Duty of Care," making platforms liable if their product design is found to knowingly contribute to self-harm, eating disorders, or exploitation.

  • Third-Party Audits: Require annual, independent safety audits of major platforms to ensure their moderation systems are actually functioning as claimed.


Conclusion

The consensus among the contributors is clear: Reliance on individual self-control is a losing strategy. A successful policy must treat online safety as a public health issue, requiring the same level of rigorous engineering and regulation we apply to physical toys, food, and medicine.


No comments:

Post a Comment

Assignment #5 due 3/23/26 Security Risks in Corporate America

  Briefly outline some security risks in corporate America and steps that companies take to combat them. They can be cyber or internal like ...