Plain Language Summary
# Sammy's Law Summary **What It Would Do:** Sammy's Law would require large social media platforms to allow parental monitoring software to access and manage children's accounts and activity. Under this bill, parents or guardians could authorize safety software providers to view their child's data, manage interactions and settings, and monitor their social media use.
The software providers would be required to register with a governing body and could only share a child's data with parents/guardians in limited cases—specifically when the child faces or may face serious harms like suicide risk, eating disorders, sexual abuse, harassment, or academic dishonesty. **Who It Affects:** The bill primarily affects children under 17, their parents and guardians, social media companies (especially large platforms), and makers of parental monitoring software. Social media platforms would be required to create mechanisms allowing this oversight, while software companies would gain access to monitor children's accounts on these platforms. **Current Status:** The bill (HR 2657) was introduced by Representative Debbie Wasserman Schultz (D-FL) in the 119th Congress and is currently in committee, meaning it has not yet been debated or voted on by the full House.
CRS Official Summary
Sammy’s LawThis bill requires large social media platforms to permit certain providers of safety software to monitor and manage the activity of children under the age of 17 on such platforms.Specifically, large social media platforms must make available a mechanism by which a child or their parent or guardian may permit a provider of safety software to (1) manage the child’s interactions, content, and account settings on the platform; and (2) regularly access the child’s user data.A software provider may only disclose a child’s data under limited circumstances, including to the child’s parent or guardian if the child is experiencing or is at foreseeable risk of experiencing specified harms. Such harms include suicide, eating disorders, sexual abuse, harassment, and academic dishonesty. The provider may only share data necessary for a reasonable parent or caregiver to understand that the child is experiencing or is at risk of harm.To participate, a software provider must register with the Federal Trade Commission, undergo a security review, and demonstrate that, among other requirements, the provider is based in the United States and will use a child's data solely to protect them from harm.Under the bill, a large social media platform is generally a service that enables a child to share content through the internet with other users that the child has become aware of solely through the platform, and which has more than 100 million monthly global active users or generates more than $1 billion in gross annual revenue.
Latest Action
Forwarded by Subcommittee to Full Committee by Voice Vote.