Do Not Turn Child Protection into Internet Access Control
542 points
7 hours ago
| 52 comments
| news.dyne.org
| HN
yalogin
4 hours ago
[-]
The big issue isn’t even age verification. The end goal is verified user identification. They want every transaction on the internet to be associated with the exact identity of the user. No more anonymity.

In the short term the way it will be implemented is this — age verification will not be a binary, it will also want to push your DoB, name, location etc and they say “the choice is with the user” but the default will be to send everything. Very soon there will be services that require DoB or name or something else to gate new or existing functionality. That is the slippery slope it will be built as and that is how they win the game

reply
totetsu
1 hour ago
[-]
It’s not very soon, it’s already the case that if one wants to enable the latest models in the OpenAI api you have to submit your details to their “identity provider”.
reply
abracadaniel
1 hour ago
[-]
Which is why it’s important to be able to run models locally. Which also might explain the strategy behind buying all of the memory that is or will exist for at least a year out. Maybe we’ll eventually see AI safety be used to prevent people from running local models.
reply
gruez
47 minutes ago
[-]
Given the recent mexican telecom hacks were allegedly done with significant help from openai/anthropic's chatbots, it seems at least somewhat prudent to require some sort of identity verification for API access? I'm struggling to see how this isn't the tech community's version of "no background checks for gun purchases" or "no KYC for bank accounts".
reply
paradox460
38 minutes ago
[-]
They were also likely done with keyboards and mice. Should we require id at point of purchase for those?
reply
gruez
33 minutes ago
[-]
Alright, so does that mean we don't need KYC for gun purchases or bank accounts either?

Of course you're probably going to say something about how guns and bank accounts are crucial components to crime, in which case the same holds for AI in the mexican telecoms hack.

reply
BobbyJo
17 minutes ago
[-]
What happens when everyone needs to use AI for their job? Genuine question that I think gets at the heart of the debate.

Once a common technology that everyone has access to becomes powerful enough to alter the lives of others on command, do we as a society just need to do away with the concept of anonymity? We are all just too powerful in isolation, and too much of a threat to the collective, that we cannot reasonably expect not to have some governing body watching at all times?

Today, you can buy parts/print a completely untraceable firearm, so do we license sales of steel tubing and 3D printers?

reply
gruez
4 minutes ago
[-]
>What happens when everyone needs to use AI for their job? Genuine question that I think gets at the heart of the debate.

Considering most places does direct deposit and that requires a bank account (so KYC), I don't see what's particularly new here. Many places also do background and/or work eligibility checks, which again is a form of KYC.

>Today, you can buy parts/print a completely untraceable firearm, so do we license sales of steel tubing and 3D printers?

Fortunately 3d printed guns are bad enough that it's not really an issue, although the bigger threat is probably CNC machines. However that's probably will get a pass, because they're eye-wateringly expensive compared to black market guns that nobody would bother.

reply
Buttons840
1 hour ago
[-]
Somehow they will eliminate anonymity for real people, but bots will still be pushing Russian or... some other country's interests with massive bot farms.
reply
owisd
4 hours ago
[-]
If the end goal was user identification then the digital ID + zero knowledge proof age verification methods would be disallowed, which they aren't. https://blog.google/products-and-platforms/platforms/google-...
reply
mindslight
3 hours ago
[-]
You got suckered by the marketing. Google's "zero knowledge" approach requires devices locked down with remote attestation, which prohibits end users from running their own code (when interacting with websites that prevent it, which as time goes on under this plan will be everywhere). The only actual difference here is that this is Google's desired approach to destroying anonymity and personal computing.
reply
remcob
2 hours ago
[-]
Why is that required? The whole point of zero knowledge proofs is that it can run on untrusted devices.
reply
Aurornis
2 hours ago
[-]
Because true “zero knowledge” proofs are actually useless for age gating purposes.

Conceptually, if a proof was truly zero knowledge and there were no restrictions on generating it, there would also be nothing stopping someone from launching a website where you clicked a button and were given a free token generated from their ID. If it was truly a zero knowledge proof it would be impossible to revoke the ID that generated it, so there is no disincentive to freely share IDs.

So every real world “zero knowledge” proof eventually restricts something. Some require you to request your tokens from a government entity. Others try to do hardware attention chains so theoretically you can’t generate them outside of the approved means.

But the hacker fantasy of truly zero knowledge proofs is impossible because 1 hour after launch there would be a dozen “Show HN” posts with vibe coded websites that dispense zero knowledge tokens.

reply
gbear605
2 hours ago
[-]
You’d have to ask Google
reply
hei-lima
4 hours ago
[-]
I was a kid with unrestricted, unsupervised internet access, and it definitely affected many things in my life. If I happen to have a child in the future, they won't go through that.

The Brazilian government passed a law requiring age verification for every site categorized as 16+. It can't be self-declared, so companies usually resort to facial scans and ID verification. I DO NOT want photos of our Brazilian children going to foreign agents who are PROVEN to profit from and do God-knows-what with our biometric data. And the funniest part? The same law says 'regulation shall not, under any circumstances, authorize or result in the implementation of mass surveillance mechanisms,' but also mandates that these measures must be 'AUDITABLE.' In other words, someone needs access to that data. It’s all so stupid and incoherent.

People who are less tech-literate FIERCELY support the measure, and whenever someone opposes it, they claim that person supports digital child abuse...

Anyway... the responsibility of protection should come from the parents, not from companies that profit off your biometric data.

reply
poly2it
3 hours ago
[-]
I guess the opposite case might not be as interesting to many, but I achieved basically unfiltered internet access as a child, and it has been immensely helpful for me as a person. Everything I am today -- a programmer, technically literate, a founder of a startup with momentum, I am because I had freedom and autonomy as a child (which was not granted to me, rather achieved by me). Many of the people of my age who grew up with strict controls and supervisory parents seem kind of lost and uninformed to me, now that they are turning into adults. I feel this narrative is surprisingly rarely heard on HN, but I cannot be the only one?
reply
cedws
1 hour ago
[-]
I think the same for me, I’m pretty sure I wouldn’t be in my career if I had been restricted to an hour a day on a filtered iPad.

But I also think the internet has more potential for harm now. Widespread social media makes it easy for predators. YouTube actively incentivises content creators to produce brain numbing shit instead of the more amateur and educational content I was exposed to. Instagram creates vicious dopamine hooks that children have no mental defense against.

Also sorry to sound egotistical but I think I was an outlier that drifted into doing educational things, many or most kids will spend every moment they get just playing video games.

That being said, I’m in favour of parents doing the parenting, not the government.

reply
deadbabe
4 hours ago
[-]
What did it affect in your life? Ultimately something with affect a kid’s life.
reply
hei-lima
4 hours ago
[-]
I mean... access to adult content at that age is really, really bad. It really messed up my brain. Gore videos, chatting with adults, etc. But I learned many good things, too. It's a double-edged sword.
reply
ivanjermakov
4 hours ago
[-]
I don't see how this "child protection" enforcement would help in case of small obscure websites with porn and gore? No way their admins gonna comply. I doubt ISPs would go that far to DNS whitelist compliant websites only.
reply
gzread
2 hours ago
[-]
Does the admin of the small website hypothetically agree that they don't want to show gore to children?
reply
Aurornis
2 hours ago
[-]
The admins of sites like that DGAF about anything or anyone. They enjoy the chaos and shock.

If you expect admins of edgelord websites to respect the laws of different countries or even care about kids, I suggest checking out 4Chan’s response to various attempts to regulate them.

reply
hei-lima
3 hours ago
[-]
I never said this would help... in fact, I’m against this kind of measure, at least the way it’s being done. But I wouldn’t be surprised if Brazilian ISPs are forced to block this sort of thing (just look at what happened with Twitter (X) the year before last).
reply
amatecha
3 hours ago
[-]
For me, it didn't mess up my brain at all, it showed me a much broader range of what humanity really is, which is exactly what I wanted to understand at that time. I understood the depravity humans will exact upon others, or those they see as lesser (such as the treatment of animals, or prisoners, "the enemy" whoever/whatever that may be). I also saw unfiltered sharing of valuable knowledge, science, tech stuff, software, games, music, culture...

The uncensored internet taught me more than I could ever have been taught in school, and I'll be forever grateful for that. It didn't take me long to understand that I could generally hate no ethnicity or people or country, and the people who do are manipulated by their government or other powerful figures in their life (or disproportionately swayed by experiences in their life). Humans are pretty much all the same, we all have far far more in common than we do differences. I have a stronger perspective of this than my immediate ancestors (demonstrated over and over throughout my life) and I do credit my exposure to the open internet for a huge amount of that.

There is one huge and problematic difference now, though: the uncensored internet of the 90's is nothing like the disinformation-saturated internet of today.

reply
udhottuhao
4 hours ago
[-]
As a kid, I know that it is pretty easy to avoid those websites(because I do).
reply
hei-lima
3 hours ago
[-]
Congrats. Keep it that way.
reply
sneak
3 hours ago
[-]
What did it do to mess up your brain? What were the lasting negative effects?
reply
grvdrm
4 hours ago
[-]
Messed up how?
reply
bilekas
6 hours ago
[-]
It's too late and never about children, simply deeper forms of data harvesting and surveillance.

What makes me extremely sad and concerned is that more recent generations simply have no idea or expectation of privacy online anymore. There will never be more of a fight against all this Orwellian behavior.

reply
smartmic
6 hours ago
[-]
It’s only too late when we stop fighting back and accept it as a given. Don’t underestimate civil disobedience and the hacker spirit.
reply
GeoAtreides
4 hours ago
[-]
UK showed how to deal with civil disobedience (fast tracked judicial process). Hardware attestation will deal with the hacker spirit.

Above all, the LLM panopticon will watch us all.

Technology will not save us. Nothing will save us but ourselves and we're busy making rent and doomscrolling.

reply
Velocifyer
4 hours ago
[-]
[flagged]
reply
dang
2 hours ago
[-]
Hey, could you please review the site guidelines and stick to them when posting here? We'd appreciate it.

https://news.ycombinator.com/newsguidelines.html

reply
bilekas
5 hours ago
[-]
While I agree with you, my worry is that younger generations have been conditioned to just expect privacy invasions, and I hear the same "Well I have nothing to hide" more and more with my younger family at least.
reply
girvo
5 hours ago
[-]
> and I hear the same "Well I have nothing to hide" more and more with my younger family at least.

Which is funny as thats what I heard from my older family growing up. Except it's a lie and they have plenty to hide!

reply
ori_b
4 hours ago
[-]
"Pass me your phone, I want to screenshot a few things and post on social media".
reply
catlifeonmars
5 hours ago
[-]
This. Fatigue and despair are by far the most effective way to control a population. You don’t need to convince people you’re doing the right thing, you just have to convince them that it’s too late.
reply
drnick1
4 hours ago
[-]
Absolutely, but this can only happen if we refuse to run nonfree software on our machines. Even if the maintainers of a Linux distro decided to somehow implement some anti user feature like age attestation, it would be trivial to patch that out from the source or to remove it from a running system with root access. The real danger here is devices that are not fully owned by the user, such as iPhones.
reply
nandomrumber
2 hours ago
[-]
And the overwhelming majority of anything running Android.
reply
bigyabai
5 hours ago
[-]
I do underestimate the hacker spirit. HN's response to Client Side Scanning was disheartening, barely anyone could condemn Apple despite the obvious red-line being crossed: https://news.ycombinator.com/item?id=28068741

And once you step outside HN, forget it. You can save yourself, but there are thousands of people that do respond to the "think of the children!" nonsense and will call you a creep for objecting to it. It's game over now, you will fight against this for the rest of your life.

reply
gzread
2 hours ago
[-]
HN is mostly for people with entrepreneur spirit, not hacker spirit - entrepreneurs who want to be known as hackers. The difference is vast.
reply
chinabot
2 hours ago
[-]
big opportunity here for therealhackernews.com
reply
gzread
2 hours ago
[-]
Not really. You couldn't get the network effect. Real hackers hang out on IRC.
reply
Cider9986
2 hours ago
[-]
That was almost 5 years ago. Lately, though, I see more people have stopped tolerating these attacks on freedom. See pewdiepie, louis rossman, deflock, piracy ressurection. Uk petition against digital ID becomes one of the largest petitions in history.

https://petition.parliament.uk/petitions/730194

reply
tqi
6 hours ago
[-]
I think it would be helpful to engage with the possibility that they are neither stupid nor ignorant, rather that they simply have different values and priorities than the early internet users.
reply
Levitz
5 hours ago
[-]
And what would those values and priorities be? Because it doesn't seem to me that they align with what they actually do.

For example, it seems to me there is a whole lot of worry around megacorporations, often related to capitalism and the inequalities it brings.

In that context, if you don't place privacy as a priority, how are you not either stupid or ignorant? Is my premise just wrong?

reply
ndriscoll
4 hours ago
[-]
You can be in favor of privacy while simultaneously thinking porn, gambling, and advertisers shouldn't be targeting children. The age verification bills I've read have steep penalties for retaining information, so that seems fine since that's literally more protection than you get in person.

It's really more just concluding that those corporations should be liable for their behavior. It also has nothing to do with "the Internet" which is largely unaffected. Except of course ideas for forcing OS behavior coming out of California which are obviously bad.

I actually think things could be a lot simpler if we just made the laws like alcohol: it's illegal (with criminal liability) for a non-parent adult to provide <restricted thing> to a child. Simple enough. Seems to work fine as-is for Internet alcohol purchases. Businesses dealing in restricted industries can figure out how to avoid that liability. That's entirely compatible with making it illegal for businesses to stalk everyone, which we should also do!

reply
fc417fc802
1 hour ago
[-]
If you implemented that simple solution the expected outcome is businesses collecting ID at the door. But unlike the age verification bills there'd be no prohibition of or penalty for misuse of the collected information. It's a strictly worse outcome.

You can make intentional targeting illegal without criminalizing the accidental. And mandating self categorization of content by service providers would enable standardized filtering that was broadly effective.

The above won't get kids off of social media and it won't serve the purposes of the surveillance state but it will meet the stated goals of those pushing these measures.

Keeping children off of social media is a much trickier problem. I think we'd be better served by banning certain sorts of algorithmic feeds.

reply
ndriscoll
45 minutes ago
[-]
Okay, so make it illegal for them to record any information which is what the actual laws do (or better, explicitly criminalize all the other current stalking). The point is you don't need to be prescriptive about how to prevent children from accessing the sites. Just make it so you can face massive fines and be arrested if you don't. They can figure out how to comply with the law, and they can be effective or be shut down.

They're not actually owed a solution for how to make their business model work. They can just be told that what they're doing is unacceptable, and they can figure out what they'd like to do next. If you're worried they might react with some other unacceptable thing, we can clarify that that's not okay either.

reply
fc417fc802
19 minutes ago
[-]
I agree that open ended requirements are better than the imposition of prescriptive solutions. But I don't want online ID verification and that's where your proposal logically leads so I am equally opposed to it.

> They're not actually owed a solution for how to make their business model work. They can just be told that what they're doing is unacceptable,

You listed a few different things previously. Which one are we talking about here?

I think the rest of us are owed a solution where we can still do what we want without having our privacy violated. Regulations need to take the end user into account.

I already proposed what I think would be a workable solution to achieve the stated goals without unduly eroding the status quo. Do you have any response to it?

reply
ndriscoll
9 minutes ago
[-]
Self categorization has been the status quo since the 90s and has been proven to be insufficient. More generally, assuming people agree that something is a social problem/should be restricted, I don't think "have a third party come up with a solution that people can buy to filter us" makes sense.

We don't give kids special debit cards that detect and block purchases of cigarettes and alcohol and say "make sure your kids don't get cash". We make it a crime to sell those things to a child.

Why is online ID verification a problem for e.g. porn and gambling but it's fine for alcohol? Why should it be fully anonymous? Should we also allow anonymous porn and cigarette vending machines in person? Why is online special?

This whole idea of anonymous access can't even work in a world where you actually pay for things, which makes the whole proposition even more dubious. If you're an adult and spending money online, you already told them who you are (modulo darknet markets with crypto). Or you could buy a porn gift card in person with an ID flash like other restricted physical items. So what's the problem exactly? Ad supported porn specifically somehow is important enough to be special?

reply
sillysaurusx
5 hours ago
[-]
I’m not sure it’s possible to have different priorities without being stupid or ignorant of history. Once you concede a certain right, such as a right to privacy, you rarely if ever get it back. Most people seem not to care about this, despite ample evidence that it’s something worth caring about. Stupid is the obvious term for it, though obtuse could work as well.

Of course, I don’t blame them. They haven’t lived in a context where they need to care. All of the reasons they’ve heard to care have come from stories of people who lived before them. But ignoring warnings for no good reason is still dumb.

A better thing to engage with is whether we can meaningfully change the situation. It might still be possible, but it requires an effective immune response from everybody on this particular topic. I’m not sure we can, but it’s worth trying to.

reply
Kim_Bruning
5 hours ago
[-]
> They haven’t lived in a context where they need to care.

You might believe you don't need opsec, and then new laws are passed, or your national supreme court overturns the case that gave you your rights, or someone invades; and now suddenly you're wanted for anything from overstaying a visa, outright murder, or simply existing.

USA, right now, peoples lives are being destroyed because the wrong people got their data. Lethal consequences exist in Russia, Ukraine, Israel, Palestine, Lebanon, Iran.

Certain professions per definition: Journalists, Lawyers, Intelligence, Military.

Certain Ethnicities. (Jewish, Somali) ; Faiths...

It doesn't need to be quite this dramatic though. But you might accidentally have broken some laws and don't even know about it yet. Caught a fish? Released a fish? Give the wrong child a bowl of soup [1]. Open the door, refuse to open the door. Signed a register; didn't sign a register. The list of actual examples is endless. The less people know about you, the less they can prosecute.

[1] A flaw in the Dutch Asylum Emergency Measures Act (2025) that would have criminalized offering even a bowl of soup to an undocumented person. The Council of State confirmed this reading. A follow-up bill was needed to fix it.

reply
closeparen
4 hours ago
[-]
There is no world where a totalitarian government’s law enforcement ambitions on some object-level question are thwarted by the same government’s enforcement of privacy law. Countries with GDPR that are thinking of rounding up and kicking out the refugees know perfectly well who and where the refugees are.
reply
gzread
2 hours ago
[-]
The law is irrelevant in that case but the actual situation is not. If people have never put their personal information online, the bad government can't get it from online. A new phone coming out during the time of the bad government, that says the government requires you to enter your name and address, will not be received as well as if it comes out during good government times.
reply
nandomrumber
1 hour ago
[-]
> will not be received as well as if it comes out during good government times.

What bearing does that have on anything.

reply
fc417fc802
1 hour ago
[-]
Making the point that people tend to engage in short term thinking. The reception of the same law, product, or practice will be colored by the current government as opposed to potential future ones.
reply
Kim_Bruning
3 hours ago
[-]
You're not entirely wrong; ultimately if they put enough resources towards it they can probably catch quite a number of people. But governments have limited resources and really don't track everyone all the time. Not even in 2026 are they able to do that yet. It helps if you maintain some level of opsec. If they really want to get you, they can get close, but see eg Ed Snowden; who managed to stay ahead of the US government just long enough to reach relative safety (FSVO).
reply
nandomrumber
1 hour ago
[-]
Snowden’s experience doesn’t generalise to, well, anyone really.
reply
Kim_Bruning
27 minutes ago
[-]
Well, I wouldn't personally recommend single-handedly taking on the most powerful nation on earth, myself.

But turns out that if your opsec is decent, and even using mostly publicly available tools like Snowden did, you might survive even that.

In the nuanced case, normal people applying more normal opsec can handle more normal things, would seem to follow.

reply
closeparen
4 hours ago
[-]
I have the right to my own senses, my own observations, my own memories. I have the right to photograph what I can see with my eyes, and to write down what I can remember. Unless enjoined by a specific duty of care (doctor/patient, attorney/client, security clearance, etc) I have the right to discuss my memories with others. This obtains even when using electronic tools and even when working in association with others.

I don’t intend to give up or accept limitations on these rights because you consider yourself to have “privacy rights” or ownership interests in my records, my memories, my perceptions, or the reality in front of me. I find the notion of the government or another person interfering in this process, the perception and recollection of reality, to be creepy and totalitarian by itself.

In 1984, it is not only that the government is aware of Winston, but that it routinely tampers with or destroys evidence of the past & demands to control the perception of the present. I do not think we should let a government do that, even for a good reason like “protect your privacy” any more than we should let it destroy general purpose computing “for the children.”

reply
fc417fc802
1 hour ago
[-]
It can remain legal to operate a security camera while being illegal to upload unencrypted footage to any third party. I'm not worried about individuals, only about big business and the government.

> This obtains even when using electronic tools and even when working in association with others.

I think it is reasonable to place limits on public "speech" (ex uploading videos of people) without interfering with private (in the case of electronics E2EE) communications.

reply
Kim_Bruning
3 hours ago
[-]
I'm actually fine with that; so long as that is restricted to your own senses, observations, and memories; and doesn't somehow spill over and somehow pertain to mine. Basically the typical freedom to swing your fists ends at the tip of my nose argument. This is probably a solvable problem between reasonable people; give or take.
reply
gzread
2 hours ago
[-]
There are many people rights people don't have and they're okay with that and even support not having the right to stab people, not having the right to steal from a store, not having the right to take nude pictures of children... What if this one is like that?
reply
micromacrofoot
4 hours ago
[-]
they are saddled with more problems that they can reasonably care about and broader issues like privacy drop off of their radars because they've never had it
reply
taurath
6 hours ago
[-]
Too many people making too much money - to be honest, people really should blame tech for it, all it takes is RSUs to look the other way. Morally most of the US is running far away from tech and the surveillance state but here it’s still okay to work for monsters and self justify building population control systems and ad networks (often one and the same)
reply
dmix
6 hours ago
[-]
The solution is always to constrain every level of government with more aggressive privacy laws. As long as they are allowed to do it then some private contractors will take the money to help make it ... or government will make their own in house tech teams. Relying on the morals of the general public to limit state surveillance is not a good strategy, but it is of course good when companies take a stand and the tech community creates tools to push back.
reply
taurath
5 hours ago
[-]
Companies create the environment - the government is supposed to be “small” - and it must remain small so the US “consumer” can be leeched from
reply
dmix
2 hours ago
[-]
The US government is very far from small. That said, I'd be open to rules on the data broker industry though considering it's scale and how the foreign governments can buy/hack them bypassing all of Tiktok-esque national security handwaving.
reply
fc417fc802
57 minutes ago
[-]
> Tiktok-esque national security handwaving.

Algorithmic feeds are propaganda tools. A foreign government being able to propagandize your citizens is a legitimate threat, not handwaving.

reply
throwaway173738
5 hours ago
[-]
It should be prohibited outright. If you allow a loophole for corporations then they will just sell it as a service and we will never be free of it.
reply
arcanemachiner
6 hours ago
[-]
By RSU, I'm assuming you mean this:

> Restricted Stock Units (RSUs) are a form of equity compensation where employers promise company shares, typically vesting over time, offering a way to align employee interests with company performance

reply
taurath
5 hours ago
[-]
Yes - you buy the house in the bay, and companies will lock you in with the vesting schedule. Just another 3, 4 years and you’ll be rich enough to afford a second one, or retire early. Some people can self justify what they do, or pretend because they work in a “nicer” part of a company than the core revenue part that it’s all okay that what pays their checks is mass behavior manipulation. I don’t like ads or social coercion, at all.
reply
aucisson_masque
4 hours ago
[-]
Has there even been a time when we really had privacy online ?

It didn't take long for the CIA to sniff everything on everyone, early 2000's.

Maybe you're referring to the 90's but at that time the internet wasn't really that popular, it was a niche thing.

reply
catlifeonmars
5 hours ago
[-]
With respect, this take is a good example of all or nothing thinking. It’s not too late.
reply
NeutralCrane
4 hours ago
[-]
I live in an area that has been declared among the safest in America. Two months ago a 17 year old girl from our city disappeared. Turns out she had been being groomed for a year over Discord and in Roblox by a 39 year old the next state over. He eventually convinced her to let him pick her up, after which he filmed himself having sex with her, killed her, and then dismembered her body. He apparently was grooming other underaged girls in a similar way as well.

The digital age presents with it novel forms of danger for children, and for adults for that matter, and there is absolutely no way to effectively address these risks without some amount of reduction in privacy. And before someone inevitably says “where were the parents?” and wash their hands of the situation, a healthy society should care for and protect all children, especially those whose parents do not.

It’s one thing to hold the opinion “I am willing to sacrifice some number of lives, in order to preserve privacy”. That is an honest and potentially justifiable opinion someone may hold. But declaring the situation to simply be a facade to harvest people’s data seems to me like a reflexive response to avoid uncomfortable truths regarding the situation.

reply
fc417fc802
38 minutes ago
[-]
There will always be weird tail risks. The law should only get involved where there are widespread systemic problems.

People are occasionally hospitalized due to self, family, or friends handling food improperly. That doesn't warrant a legal intervention whereas dining establishments do.

> before someone inevitably says “where were the parents?” and wash their hands of the situation

Nope, that's exactly what I say. The law cannot reasonably replace responsible parenting if society is to remain a pleasant place to live.

reply
chinabot
1 hour ago
[-]
If the government knew every single user on the internet's name, address, phone number and what they had for breakfast, it would not stop monsters like this, or even slow them down.
reply
AJ007
2 hours ago
[-]
Discord & Roblox - no encryption, privacy, or anonymity on either of those platforms, by the way.
reply
mindslight
1 hour ago
[-]
Still none of that necessitates the type of mandatory partial-ID verification being pushed by these laws.

Roblox can straightforwardly require ID verification on their own, of both the parent responsible for the account, as well as the children directly (request documentation from their school, birth certificate, etc. Yes, high touch to verify these documents. But we're talking protecting children here, right?)

If anything this type of legislation is about absolving them of the responsibility of doing so!. Imagine a company making their offering "for adults only", with de facto kid usage as parents relent and just let their kid use an older age on the computer.

reply
SilverElfin
5 hours ago
[-]
For the government it may be surveillance. For the people funding these new laws, it is about advertising profits. See what I said at https://news.ycombinator.com/item?id=47471747
reply
mattmanser
5 hours ago
[-]
Go watch the newest Louis Theroux, into the manosphere.

At points Louis and whatever absolute scumbag he's with walk around the streets while the guy is filming his own content.

There are kids, literally 11/12 year olds, walking up to these predatory, evil, scammers on the street going "oh my god it's MC" or whatever their name is. Multiple times.

And he hardly gets to spend any time with these men because they clock pretty quickly they're not going to come off well.

In the space of like 3 days, Louis caught on camera at least 10/20 young kids recognizing these toxic people from videos they had watched. Even the ones who'd been banned from most platforms, because their videos get reshared under different accounts and insta/tiktok/facebook aren't bothering to catch these reshares.

It really is about the kids.

And it all comes down to these people convincing young men to spend money on scam courses or invest in scam brokerages by getting them to join telegram group chats. And suddenly it's really clear to me why telegram's under scrutiny.

reply
zingerlio
3 hours ago
[-]
I share your observations and concerns. But I don't think the current erosion of digital privacy and the censorship creep were made to address those. There are better ways (even though they are not fully fleshed out yet) to minimize toxic/populist influence, but a blank cheque to sacrifice our rights isn't one.
reply
gzread
2 hours ago
[-]
Some of them clearly are - the one in California.
reply
Keeeeeeeks
5 hours ago
[-]
A theory that’s floating around is that since frontier models are so good at sounding like humans, companies paying for ads are arguing that Dead Internet Theory -> ad costs should go down.

Therefore, the push to ID everyone using the internet (even down to the hardware) is a way to prove that ads are being served to real humans in their target demographic.

reply
phendrenad2
4 hours ago
[-]
It makes a lot of sense, too. Previously, governments wanted everyone to have to swipe their driver's license before accessing the internet. But now, businesses want it too. And that makes all the difference in a world built on capitalism.
reply
jmcgough
6 hours ago
[-]
What's sad is how effective this is. Religious groups figured out a few years ago that anti-porn groups accomplish nothing, but if you start an anti-trafficking group you can restrict porn access.
reply
tangotaylor
5 hours ago
[-]
Their real goals are even worse than that. Some of these groups have admitted they're also about suppressing LGBT+ content.

As the Heritage Foundation admitted:

> Keeping trans content away from children is protecting kids. No child should be conditioned to think that permanently damaging their healthy bodies to try to become something they can never be is even remotely a good idea.

https://arstechnica.com/tech-policy/2024/07/kids-online-safe...

reply
gruez
3 hours ago
[-]
>they're also about suppressing LGBT+ content

>> Keeping trans content away [...]

Isn't it a stretch to round off "trans content" to "LGBT+ content"? I mean, from a pure logical point of view the statement is correct, because "trans content" is a subset of "LGBT+ content", and therefore "suppressing LGBT+ content" is technically correct, but it's at least misleading. The left's version of this would be something like "twitter is suppressing anti-immigration content!", and the actual example is some alt-right commenter saying that immigrants should be lynched. Immigrants being lynched is certainly an subset of "anti-immigration", but it's still misleading.

reply
_moof
2 hours ago
[-]
Hi, I've been openly queer for over 20 years. Using trans people as a wedge to pry apart the entire LGBTQ community is a tale as old as time. This isn't theoretical or a slippery slope argument; it's recent history. It's effective because it sounds "reasonable" on its face, but it's a ploy.

Just one of the many, many, many reasons that trans rights are human rights.

reply
gruez
35 minutes ago
[-]
>This isn't theoretical or a slippery slope argument; it's recent history. It's effective because it sounds "reasonable" on its face, but it's a ploy.

It quite literally is the slippery slope argument. You just don't want to call it that because the term is almost always used in the context of a fallacy, and you think you're right. It's like "freedom fighters" vs "terrorists". Nobody calls themselves terrorists, even terrorists.

Moreover the "It's effective because it sounds "reasonable" on its face, but it's a ploy" argument works equally well for any side, eg. it's not hard to imagine someone on the right saying "today it's Jan 6th protesters and that might seem reasonable, but tomorrow it's anyone at unite the right protests, and when president AOC's in power it's anyone who's protesting against trans surgery for minors".

reply
gzread
2 hours ago
[-]
Trans content is first, it'll be gay content after that.
reply
gib444
4 hours ago
[-]
Heritage's tweet in the screenshot in your link makes no reference to "L", "G", "B" nor "+". Just "T"
reply
abcde666777
4 hours ago
[-]
I mean, I don't disagree with the sentiment of keeping trans ideology away from kids, in exactly the way I'd want to protect them from any kind of religious indoctrination.
reply
simonask
3 hours ago
[-]
The idea that even a single person in the world is trans because of "trans ideology" is what's absolutely insane here.

If you think you could be convinced by anyone that you're not living out your true gender identity, I have news for you... Most people, children too, are not having those thoughts unless there's actually a journey waiting for them.

reply
abcde666777
1 hour ago
[-]
The idea that none of them are trans for that reason is what sound insane to me. Are you not aware of how much humans absorb from the messaging in their environment?

And FYI, I've seen it happen with one of my own family members - someone who so far as I can tell isn't 'a man in a woman's body', but rather just someone who never fit in and was always a bit of a social outcast.

Their struggle was never their bloody gender, it was their struggle to find a way to fit into the world.

And that's what a lot of transitioning actually is. Because human psychology works such that when we're not fitting in, when we feel insecure and out of place, a subconscious pressure emerges to reinvent ourselves due to the current formula not working for us.

It's offensive to me that you'd make such claims whilst clearly so naive about it.

reply
signalfour
3 hours ago
[-]
Is it really insane? In the detrans community, it's not uncommon for people to see their transition as ideologically led, see e.g. https://lacroicsz.substack.com/p/by-any-other-name.
reply
simonask
2 hours ago
[-]
OK, sure, there’s going to be confused and vulnerable people making mistakes. As I understand it, the detrans community is a tiny fraction of an already tiny fraction who ever in their life identified as trans.

This isn’t an actual risk to anybody, and I can’t believe I have to say that.

reply
AuryGlenz
1 hour ago
[-]
It’s weird to say in the same comment “it happens but as far as I know it’s rare” and “this isn’t an actual risk to anybody.”

I do wonder how many would detransition if it wasn’t too embarrassing for them or because they’re effectively stuck that way if they did bottom surgery. Certainly I’m sure there are many more that quietly do it and not be a part of a community around it - I know of one person who did.

I’m sure there are many reasons people transition. For instance, there used to be a subreddit that collected (many) trans people commenting on how much them transitioning or being the other gender turned them on (Reddit being Reddit, it was banned). Some might do it because it feels like a way to get a new start. Some might do it because it’ll get them attention. Some, I’m sure, do it because they genuinely feel like they have the wrong body.

All of those could (of course) be affected by social interactions, with only the last one being positive. Unfortunately it would be really hard for us to ever know the true statistics, as that sort of thing is hard for even the person experiencing it to suss out.

reply
chinabot
1 hour ago
[-]
Respectfully, not everyone shares this view.
reply
Dig1t
2 hours ago
[-]
Do you believe that children are more impressionable than adults? There is a community of detrans people who talk openly about how they became trans because they were influenced by peers and authority figures in their lives.

Go read some threads on the detrans subreddit.

reply
gzread
2 hours ago
[-]
Maybe we shouldn't hide the information then, so they can make their own decisions. Imagine blocking all the information about "am I actually trans or just peer pressured?" but not blocking the peer pressure.
reply
WarmWash
3 hours ago
[-]
Firey take there, but I know a few people who are trans and neck deep in the kool aid. They will tell you that 25-30% of population is trans, and just haven't been liberated/are in denial.

Look, it's cool to be trans, no problem. These women I know are good people and net contributors to society. But they are off the ideological deep-end, and would happily spend 3 hours at the family BBQ lecturing an impressionable 13 year old about how those weird body feelings are very likely gender dismorphia. They're just as drunk on their flavor of delusional social media as any other religious nut is crazy about God.

reply
poly2it
3 hours ago
[-]
Gender dysphoria can be medically studied, is not an ideology and is not a disorder. Hope this clears things up.
reply
abcde666777
1 hour ago
[-]
What we call gender dysphoria is really just a cluster of symptoms around people's sense of their identity.

But identity as a whole is a very murky thing - if you ask me it's largely an adaptive abstraction that our minds invent.

The purpose of said adaption is to adopt a role which functions within the tribe/society for purposes of survival.

I think we way over-simplify the whole thing by making it about gender and gender roles.

And it's that over-simplification that I would label as the ideology. Because that's what ideologies do: they take the complex ambiguities of the world and try to cram them into a simplistic box.

reply
signalfour
3 hours ago
[-]
It stopped being centered around gender dysphoria quite a while ago. Gender identity is where it's at now, and the idea that one does not need to be dysphoric to be trans is currently the most mainstream one.
reply
poly2it
2 hours ago
[-]
Yes, you do not need to be dysphoric to be transgender, however. It is actually quite difficult to compare rates of gendery dysphoria to transgender identity, as transgender identity is inherently self-reported, but studies on gender dysphoria focus on diagnosed cases, not undiagnosed estimates. Therefore it is also not possible to assert that non-dysphoria is dominant among current transitioning people as you do.
reply
phyzix5761
5 hours ago
[-]
Which religious groups specifically are pushing for this and where? I want to know so I can call them out when I see it.
reply
chaostheory
5 hours ago
[-]
It’s meta this time.
reply
mc32
5 hours ago
[-]
Traffickers now use refugee programs as conduits for human trafficking.
reply
txrx0000
3 hours ago
[-]
We have to separate child protection from Internet control so that the "protect the kids" narrative loses its potency. So here's a counter-narrative: we can implement digital child protection without Internet-wide access control, and it requires just 3 simple features that can be implemented in less than a week. There's no need to introduce new laws at all. This could just be done tomorrow if there is genuine will to protect the kids.

1) If you're a platform like Discord or Gmail, give users the option to create an extra password lock for modifying their profile information (which includes age). This could also be implemented at the app level rather than at the account level. Parents can take their child's phone, set the age, and set these passwords for each of their child's apps/accounts.

2) If you're an OS developer, add a password-protected toggle in the OS settings that gates app installation/updates, like sudo on Linux. Parents can take their child's phone and set this password, so they can control what software runs on their child's phone. If we have this, then 1) isn't even strictly needed because parents can simply choose to only install apps that are suitable for their child.

3) If you're a device manufacturer, you should open-source your drivers and firmware and give device owners the ability to lock/unlock the bootloader at will with a custom password. Parents should be able to develop and install an open-source child-friendly OS. Companies like Apple and Samsung have worked against this for years by introducing all kinds of artificial roadblocks to developing an alternative OS for their hardware.

reply
tzs
2 hours ago
[-]
(This is a reply to the dead comment, which was not dead when I start writing this)

I don't know how long their specific proposal would take, but on a Unix or Unix-like system the California bill could be done in a week.

0. Make a directory somewhere, say /etc/age_check, and in that directory create four files: 0-13, 13-16, 16-18, 18+, owned by some system account with permissions 000.

1. This would be the hardest part. Modify whatever is used to interactively create new user accounts to ask for the user age if the account is a child's account, and than add an ACL entry for the appropriate /etc/age_check file that allows the child's account to read that file.

The California bill says you have to ask for and age or birthdate but the API you provide for apps to ask for age information just requires giving an age bracket, so I'm taking that as meaning I am not required to actually store the age. I only have to make the API work.

2. The API for checking age is to try to open the files in /etc/age_check. Whichever open succeeds gives you the user's age bracket.

reply
tzs
3 hours ago
[-]
So basically parents set the child's age and apps rely on that if they need to know if the user is old enough?

That's pretty similar to the California bill. Parents set an age when creating a child's account. The OS provides an API to get the user's age bracket from that, which apps that need to know the age bracket of the user can call.

reply
txrx0000
3 hours ago
[-]
The California bill gets it backwards. Rather than Internet services taking the user's age and deciding what content to serve, the Internet service or app should broadcast the age rating of its content to the OS (if convenience is desired), like how movie ratings work. The responsibility to decide what content is suitable for a child should rest in the hands of that child's parent, not the state or the corporation.

edit: on second thought, realistically, the API solution is too brittle regardless of which way it goes. Because the API requires every service to implement it and that's not happening, whereas an app installation lock only requires one child-friendly OS to implement it, then parents can choose that OS.

reply
cvhc
19 minutes ago
[-]
That's not my understanding. This is what the bill says: Provide a developer who has requested a signal with respect to a particular user with a digital signal via a reasonably consistent real-time application programming interface that identifies [the age group].

So the app requests a signal (like, calling an API), and the OS returns the signal (returning the age group).

Regarding API vs installation lock, TBH I don't think the law concerns that level of details. An OS or app-store installation lock that checks app ratings can be considered as a valid implementation.

reply
gzread
2 hours ago
[-]
Would the content rating be per HTML element and the browser would delete the elements with bad ratings from the DOM, or how would it work?
reply
txrx0000
2 hours ago
[-]
I'd imagine it works like movie ratings. You don't filter movies from scene to scene. There's just one rating for an entire site or app.

But yeah I get the point, API based solutions are complicated and brittle because they require all services to implement it properly. In contrast a user-set app installation password in the OS settings is more effective and easier to implement.

reply
gzread
2 hours ago
[-]
If a chronological social media feed contains both R and G rated elements how would you implement that?
reply
mindslight
1 hour ago
[-]
> the API requires every service to implement it and that's not happening

No it doesn't. A browser/appinstaller with parental/age controls enabled would fail as unavailable if there was no age rating on the website/app. This is exactly the solution we should be aiming for, as it keeps the incentives lined up instead of turning them upside down.

One big problem with the laws currently being pushed is that it leaves the decision for what sites are "appropriate" for kids completely in the lands of corporate attorneys. For example, Facebook will happily make an "under 18" site that uses LLMs to censor posts, but still contains all of the same dopamine drip mechanics. Whereas keeping the decision process of appropriate under the control of the end-device means parents could straightforwardly go beyond what corporate attorneys decide, and block Facebook regardless of the age rating.

I'm responding to another comment of yours here since HN loves the rate limit. In that comment you were talking about locked down bootloaders. But bootloaders are already thoroughly locked down, and most devices are still essentially usable. The current looming threat is remote attestation, which makes it so that websites (and other services) are able to prevent you from running software of your choice when interacting with them! The backwards legislation being currently pushed is all but guaranteed to end up in more demands for remote attestation, whereas the correct direction of information flow (sites/apps publish headers saying they're suitable for <18 etc) would not necessitate remote attestation.

reply
panzi
3 hours ago
[-]
1) Could be simpler for a start if 2) ensures that no web sites that send a special "over 18" server header are displayed. The header could be more detailed and the parent could select what things are allowed, but for a start make it simple.
reply
txrx0000
2 hours ago
[-]
Yes, that's even better. Make apps and websites provide an API that broadcasts the age rating of its content, then let the OS attest the apps and websites, not the other way around.

edit: on second thought, there is a trap here. If hardware manufacturers lock down the bootloader, then we're basically still handing over parental authority to governments and companies in the long run. So I think for a start, we just implement a app-install password lock like sudo. It will be easier to implement than the API. The convenience API can come later when hardware manufacturers are banned from locking bootloaders.

reply
gzread
2 hours ago
[-]
How would you make a website that can be over 18 or not, such as a social media feed? Would it become over 18 as soon as your following list contains a porn star (who may not have been one at the time you followed them), and then if you're under 18 you can't unfollow them because you can't load the page?
reply
sfRattan
4 hours ago
[-]
It's irksome that these laws and bills in multiple countries are trying to put limits on the general purpose computer. It's the wrong solution and arguably put forward in bad faith.

If you want access control, the appropriate point for regulation is with ISPs and cellular providers, and the appropriate mode of regulation is requiring these companies to provide choice and education for families, and awareness of liability.

Require ISPs and cellular network providers to offer a standard set of controls to their customers informing the common person (in common language) who is using those connections and what they are doing with them. For ISPs, this looks like an option for a router with robust access controls, designating some devices (based on MAC address) as belonging to children and filtering those devices' network requests at the network gateway, or filtering one hop up onto the provider's infrastructure (e.g. the ONT for fiber connections). For cellular providers, it looks like an app available to parents' devices and similar filtering for devices designated as belonging to children (based on IMEI).

When a family signs up for Internet service, either at-home access or cellular data, the provider must give both parents a presentation about these tools, and about the liability the parents face for allowing their children unsupervised, latchkey access to adult content, no different than allowing children to drink alcohol.

It may even make sense to require ISPs and cellular providers to track MAC addresses and IMEIs of devices their own customers designate as "for children" and make those providers liable for not filtering Internet for those devices, and also liable for allowing targeted advertising against those devices.

I don't think achieving that setup is likely, but it's fundamentally the right way to solve this problem, and parents are pushing for a solution one way or another. I don't love it, but if it's coming almost inevitably we should at least push to do it right. It's a dead-end, losing strategy to blanket oppose one solution to legislators and provide no alternative. I write all of that as someone who values privacy and liberty, both in meatspace and cyberspace.

reply
gzread
2 hours ago
[-]
> filtering those devices' network requests at the network gateway, or filtering one hop up onto the provider's infrastructure

These things are not possible with any reliability, we spent two decades encrypting everything.

reply
sfRattan
53 minutes ago
[-]
I'm not imagining filtering based on the path. Even with https, hostname is visible before the handshake. And even when Encrypted Client Hello is widely implemented, it's also easy enough for network providers to drop any ECH packets from devices flagged as "for children" and signal to those devices that their handshake must reveal the hostname, at least to the router doing the filtering.
reply
novok
2 hours ago
[-]
IMO instead of age gating everything, it should've been the other way around, which is making unrestricted smartphones or similar an 18 or 16+ device, much like cars.
reply
cs02rm0
6 hours ago
[-]
It's always been internet access control, there is no child protection.
reply
jchook
1 hour ago
[-]
Like the evergreen comic, "How would you like this wrapped?" by John Janik

For decades policymakers have been trying to sell us the same surveillance state they accuse their adversaries of having, wrapped as either security or protecting children.

https://i.redd.it/ifb8agngc7dy.jpg

reply
a-dub
4 hours ago
[-]
how about if i do nothing the internet assumes i'm a child and therefore does not track me, show me ads or permit doom scroll feeds. then if i want i can jump through some hoops and pay some money or something to get a digital id that lets me attach a zkp to all my http requests that then unlock the magic of ads, tracking and doom scroll feeds.

seems like a good plan to me.

reply
pembrook
3 hours ago
[-]
That would be a solution if the people pushing this actually cared about "protecting kids."

But let's be honest, governments want a dragnet they can use to monitor/control all internet communication. The people running western democracies are equally as power hungry and zealously authoritarian (my ideas will bring utopia!) as the people running the CCP.

The only difference is, the CCP has permissionless authority, so they ended internet freedom in China decades ago. They didn't have to ask.

Western authoritarians on the other hand, have to fight a slow battle to cleverly grind you down over time, so that you get tricked into allowing them to gatekeep the internet. It hasn't worked so far. The next step (this one) is "okay, so you don't want to have to ask us permission before you visit a website...but won't anybody think of the poor beautiful innocent children???"

Emotions activated. Rational thought deactivated.

They'll get what they want because they always get what they want. And you'll be convinced it's good for you over time, because most people just follow whatever the mainstream "vibes" are, and the elite sets the vibes. It's amazing a free internet existed this long. Great while it lasted.

reply
a-dub
3 hours ago
[-]
i'm only half joking. adding zkps to http requests is probably the correct privacy preserving technical solution that could be built into something sensible.

the bigger issue is that lawmakers are thinking in terms of smartphones, tablets and commercial pcs as shrink wrapped media consumption devices with a setup step... not protocol level support that preserves parts of computing and the internet they don't even really know exists. seems like the ietf should have lobbyists or something.

reply
gzread
2 hours ago
[-]
ZKPs don't buy anything, since an online service can sell them by the thousand and you're just trusting the client that it belongs to the actual user. You might as well just do "User-Age-Category: 18plus" then and save a headache.
reply
HardwareLust
6 hours ago
[-]
The entire purpose of this exercise is control. "Child protection" is just a ruse to get the stupids onboard.
reply
vsgherzi
5 hours ago
[-]
Y E S. I’m tired of hearing about child proofing the internet. We need a solution that’s not enforcing age or id verification on the os or internet itself like meta is pushing. We need better solutions and we should fight draconian enforcement with extreme prejudice
reply
jameskilton
6 hours ago
[-]
That's the trick, it's always been about control. No-one in such positions actually cares about the children.
reply
mindslight
6 hours ago
[-]
I think the truth is closer to them being tightly bound to one another over their shared "love" of children. Epstein bouncing around the academic community was the tip of an iceberg. Imagine the reputation laundering that goes on with all of these "for the children" NGOs.
reply
plasticeagle
5 hours ago
[-]
AI;DR

It's too late in any case, the Internet as we know it will eat itself. It will be destroyed by AI, and AI agents from without. And it will be destroyed from within by stupid laws such as the ones under "discussion" in this AI-edited and AI-illustrated nothingpiece.

By which I not mean the infrastructure. I mean the current crop of social media websites. The infrastructure will remain, and perhaps something better will come along to use that infrastructure.

reply
dlcarrier
5 hours ago
[-]
For the US, the worst of it started in 2019, when the held YouTube liable for all content that a child might access. (https://en.wikipedia.org/wiki/YouTube_and_privacy#COPPA_sett...) That's what pushed all of the content networks to lobby for the liability to go somewhere else.
reply
cluckindan
6 hours ago
[-]
It’s not even a debate if these controls are problematic. The litmus test is to mentally substitute the age field for an ancestry field and place the system in 1930’s Germany.

Coincidently, that system was provided by IBM.

reply
bluegatty
3 hours ago
[-]
'Preventing children from buying guns is Nazism!'

Actually, this sentiment is a 'litmus test' for common sense.

We use age discrimination universally in all affairs, across the globe, across all cultures.

Of course the same thing is going to apply to 'content', it's just a lot harder and creates ugly externalizations.

It's a real problem, with no real solutions, at least not yet.

reply
jjk166
5 hours ago
[-]
The people pushing for "child protection" went to the island. It's not even about control, it's about shifting liability away from platforms so they can further gut moderation, reducing their expenses and getting away with doing nothing to stop the actual bad actors.
reply
wakawaka28
4 hours ago
[-]
It's not about gutting moderation. They want you to dox yourself to get online. It's a pro-censorship authoritarian-friendly move. I don't believe the narrative that Meta is behind it all either. If they are, they are probably serving someone else.
reply
echelon
3 hours ago
[-]
Here are just some of the things you can do with tracking:

- Dox, coerce, blackmail, and ruin political candidates, powerful CEOs, and wealthy people. If they watch a category of porn that is embarrassing or have an affair, suddenly you have leverage against them. You can parlay that to accomplish lots of things.

- Make it impossible to talk about certain things and eventually eliminate those things. Porn today, abortion tomorrow. LGBT, women's rights ... it's a tool to start enforcing an ideology. Eventually these things can be disappeared entirely, not just the discourse. You just cordon off and begin washing it away bit by bit, year by year. Once the control mechanisms are in place, it cannot be stopped.

- Kill anonymous communication. This can pin identities to online comments. You can then punish people of the ideology you don't like by denying them jobs, auditing them, etc. This has a chilling effect on political opposition. This also makes it much harder to leak or report information safely and harms the ability to whistle blow.

- In general, this also pushes society into more religious, more conservative views. With it comes a lack of skepticism and a greater appreciation for authority.

- Ultimately, this is a step into 1984. If we go down that route, we will eventually be owned in whole by the authoritarian powers at top. This entire conversation will be memory holed.

Once a right is lost, we will not get it back. Then it's just one step after another into hell.

We must fight this.

Our lives, our freedom, our future - depend on it.

reply
scott_paul
3 hours ago
[-]
I disagree with almost all of your political opinions, and some of your positions I very much hate. But we should be free to have the argument, without the thread of handcuffs or the threat of starvation. Although I use my real name here, sometimes I prefer not to, and that should be allowed.

The right to actual real privacy is the same thing as the right to actual real freedom of speech, and we should harm anyone who is trying to take that most basic foundation of all rights away.

I agree with Alexander Solzhenitsyn.

reply
echelon
3 hours ago
[-]
Regardless of how we (mis?)align on social and economic issues, we should align on dislike of authoritarianism and surveillance. It is our common enemy.

----

Edit: I can't respond to comments anymore (HN rate limits on downvotes and commenting within a single thread), but I also wanted to respond to a sibling comment:

> "your team"

Just because I believe in personal freedom of people from the government does not mean I'm left-wing. I agree with some democratic party policies, and I disagree with some others.

I'm not strictly a libertarian either, because I believe government regulation is necessary to prevent monopolies. But over-regulation is also stifling to progress.

But it shouldn't matter what my politics are. Social and economic issues are orthogonal, and frankly, not as potentially dangerous as this one issue.

Democrats and Republicans alike should be aligned on their disdain of surveillance and authoritarianism. Either party in power (or any power) can use it against the "other side" (or the entire population outside of the oligopoly).

These tools are nothing but evil and designed to control. Once they start sinking their teeth in, they only sink in deeper. Every free person should hate them.

reply
pembrook
3 hours ago
[-]
You've accurately described what could happen with right-wing authoritarians in power. You've not described what could happen with left-wing authoritarians in power.

Don't be fooled that your team doesn't have people with the same impulses. Privacy and civil liberties exist to protect us from abuse of authority on all sides.

- "Oh I see John is connected to this account. I really don't like this HN comment and opinion he posted, I find it deeply offensive. Put him on the bank KYC fail list."

- "We'd love to give you this mortgage backed by the US government, but why didn't you post the right flag in support of the new hip thing?"

- "Before you login to your retirement account, how much wealth are you secretly harboring there from this job we think you unfairly got due to your privilege?"

- "If you just let us monitor your activity and the ideas you see, we'll stop you from wrong-think and will create a utopia"

reply
rdevilla
3 hours ago
[-]
Good luck, man. Nobody cared in 2012, and even less people care now. The west is lost. 1984 is already here.
reply
echelon
3 hours ago
[-]
Don't give up!

If you think the heat has started, you're mistaken. We're not even in the fire yet. It can and will get waaaay worse.

We've been able to push back against these efforts time and time again. Don't stop. Call your legislators. Talk with your friends and get them to do the same. Vote against politicians that support it.

It does work.

reply
rdevilla
3 hours ago
[-]
Whatever you think the scale of surveillance is, I assure you it is 100x worse.

North America is rooted. There is no recovery plan.

reply
timschmidt
3 hours ago
[-]
My understanding is that Abraham Lincoln literally had all the nation's telegraph lines routed through DC during the civil war, and AT&T has been an honorary branch of the US government ever since.
reply
rdevilla
3 hours ago
[-]
That tradition was carried into the modern era.

https://en.wikipedia.org/wiki/Room_641A

reply
timschmidt
3 hours ago
[-]
I'm aware. And the GP you're responding to's username is echelon. I think they're aware as well.
reply
scotty79
3 hours ago
[-]
> ruin political candidates, powerful CEOs, and wealthy people

This is mostly fantasy propagated by works of fiction. In the real world release of any evidence of sins has practically zero impact on the wealthy people and when it very occasionally does have an impact it just happens in cases of people who weren't wealthy enough for the circumstances.

reply
rudhdb773b
3 hours ago
[-]
The government can do a whole lot more than embarrass CEOs and powerful people they don't like. Look at how China controls its tech CEOs by making them disappear until their views align.
reply
echelon
3 hours ago
[-]
The Epstein Island isn't just a fantasy playground for sickos.

Every single one of those people has a noose around their neck and is being told what to do. They have a gun to their head now.

The intelligence apparatus has been exploiting dynamics like this for a long time.

reply
rdevilla
3 hours ago
[-]
The west runs on blackmail. If they can't find any dirt on you, you're not getting into power, and that's a fact.
reply
mpalmer
4 hours ago
[-]
It's mostly Meta lobbying for this, in every state. Sensationalizing and exaggerating does not help.
reply
hunterpayne
4 hours ago
[-]
But why is Meta lobbying for this? The bills they push move compliance onto the app stores. And Meta doesn't run an app store. I think the execs think its some sort of 4D chess move to put liabilities onto their competitors. I'm not sure it will work out that way. Seems like FB has a lot more to lose than they think.
reply
gruez
4 hours ago
[-]
> I'm not sure it will work out that way.

why? If age restriction get legislated into the OS, it puts a damper on further attempts on adding restrictions to sites, because they can point to the existing legislation and claim it's enough.

reply
Scandiravian
4 hours ago
[-]
They're pushing for an API at the system level, where they can query the age

Such an API can then be extended to provide location data to "help the police find bad guys", track purchase histories to "prevent fraud"; all the stuff that Apple and Google blocked fb from sniffing from user devices

It's circumvention of these privacy protections with added vengeance since now Google and Apple will be sitting with the cost of implementation and the liability

reply
gruez
3 hours ago
[-]
>Such an API can then be extended to provide location data to "help the police find bad guys", track purchase histories to "prevent fraud"; all the stuff that Apple and Google blocked fb from sniffing from user devices

/s?

In case this is serious, why do they need an age API to ask for a location backdoor API?

reply
Spooky23
3 hours ago
[-]
They would be no longer responsible for doing it.

My kid had classmates as young as 8 using it. Facebook knows this.

reply
peyton
3 hours ago
[-]
Think of how many quadrillions of hours under-18 spend online. Ads for verified 18+ are more profitable.
reply
cyanydeez
5 hours ago
[-]
I mean sure; but look at it from their POV, controlling the medium is the message right from 1984. Like LLMs, you can't learn about doing evil things without seeing how they benefit yourself.
reply
gruez
5 hours ago
[-]
>The people pushing for "child protection" went to the island.

What does this even mean aside from a thinly veiled accusation that such efforts are being pushed by a shadowy cabal of pedophiles elites? I'm sure you can find some overlap between people who want to push age verification laws and people who went to the island, but what about everyone else pushing for the law but who didn't go?

reply
catapart
4 hours ago
[-]
Like who? Name some names of people pushing for this, and we can dissect their motivation.
reply
gruez
4 hours ago
[-]
How about the first country to ban social media for kids, Australia[1]? So far as I can tell the PM/party leader was not in the files. Of course, if you make your inclusion criteria absurdly wide (eg. anyone who voted or advocated for age based restrictions in any shape or form), you'll probably find some pedophiles or even epstien island visitors from sheer luck alone.

[1] https://en.wikipedia.org/wiki/Online_Safety_Amendment_(Socia...

reply
bigfatkitten
4 hours ago
[-]
This has been on Labor’s agenda, in various forms for many years.

https://www.abc.net.au/news/2012-11-09/government-abandons-p...

reply
gruez
4 hours ago
[-]
That doesn't change the conclusion, unless you're trying to imply the entire party is full of pedophiles.
reply
bigfatkitten
46 minutes ago
[-]
My point is that this has been consistent with their policy for the last couple of decades, and that the recent round of scandals have nothing to do with it.
reply
scotty79
3 hours ago
[-]
At this point the burden of proof is on the party. Benefit of the doubt has ran out.
reply
gruez
3 hours ago
[-]
>Benefit of the doubt has ran out

...because they're pushing age verification legislation? Did I miss some massive Labor pedophile scandal? If not, this just feels like a tautology. Labor is only pushing age verification because they're pedophiles, and they're pedophiles because they're pushing age verification.

Moreover even if we ignore that, what does that mean for the rest of their platform items? If Labor is pro net-zero, is it fair to characterize the situation as "the people pushing for net-zero are pedophiles"?

reply
smallmancontrov
4 hours ago
[-]
These laws were passed almost exclusively by the party of self-proclaimed free speech warriors led by Epstein's best friend.

    State             | Effective Date | Legislature Control
    ------------------+----------------+----------------------
    Alabama           | Oct 1, 2024    | Republican
    Arizona           | Sep 26, 2025   | Republican
    Arkansas          | Jul 31, 2023   | Republican
    California        | Jan 1, 2027    | Democratic
    Florida           | Jan 1, 2025    | Republican
    Georgia           | Jul 1, 2025    | Republican
    Idaho             | Jul 1, 2024    | Republican
    Indiana           | Aug 16, 2024   | Republican
    Kansas            | Jul 1, 2024    | Republican
    Kentucky          | Jul 15, 2024   | Republican
    Louisiana         | Jan 1, 2023    | Republican
    Mississippi       | Jul 1, 2023    | Republican
    Missouri          | Nov 30, 2025   | Republican
    Montana           | Jan 1, 2024    | Republican
    Nebraska          | Jul 18, 2024   | Nonpartisan (unicameral)
    North Carolina    | Jan 1, 2024    | Republican
    North Dakota      | Aug 1, 2025    | Republican
    Ohio              | Sep 30, 2025   | Republican
    Oklahoma          | Nov 1, 2024    | Republican
    South Carolina    | Jan 1, 2025    | Republican
    South Dakota      | Jul 1, 2025    | Republican
    Tennessee         | Jan 13, 2025   | Republican
    Texas             | Sep 19, 2023   | Republican
    Utah              | May 3, 2023    | Republican
    Virginia          | Jul 1, 2023    | Divided
    Wyoming           | Jul 1, 2025    | Republican
reply
tredre3
4 hours ago
[-]
It's curious that you've omitted California (Democrats) and Colorado (Democrats) from your list.
reply
smallmancontrov
3 hours ago
[-]
I thought the Colorado bill died.

https://leg.colorado.gov/bills/sb25-201

Looks like the CA bill went through though.

https://legiscan.com/CA/text/AB1043/id/3269704

I updated the list. Still looks rather tilted to me!

reply
gruez
3 hours ago
[-]
This table seems suspect. I spot checked Texas, and while the party affiliation is correct, the dates are not. You put Sept 19, 2023 as the date for Texas, but Wikipedia[1] says it "Enacted September 1, 2024" and "Enacted June 13, 2023". Looking at the other dates, I'm not sure how you got Sept 19, 2023, even through a typo.

[1] https://en.wikipedia.org/wiki/SCOPE_Act

reply
wakawaka28
4 hours ago
[-]
Can you cherry-pick harder? Geez...
reply
bdangubic
4 hours ago
[-]
25 states isn't cherry-picking :) geeeeeeeeeeez!
reply
rkomorn
4 hours ago
[-]
I think/hope they were being sarcastic.
reply
wakawaka28
4 hours ago
[-]
No, it's bipartisan and even fucking international. I think there is a very obvious conspiracy to get this done, but maybe it's a big coincidence that governments and politicians everywhere suck now.
reply
wakawaka28
4 hours ago
[-]
I was talking about the party. This shit is and always has been pushed from both parties. Even democrat states like California and Colorado are on board. See also, the OS age verification legislation.
reply
cvhc
3 hours ago
[-]
TBH California one doesn't require age verification (while many other states do). It only requires the OS to provide a mechanism for the user to indicate their age group and apps should use the information (instead of asking for PII themselves). It's a fake one, but somehow drew most attention.
reply
wa7dj229de6
4 hours ago
[-]
reply
maweaver
3 hours ago
[-]
It means that especially those who went to the island but also most of the others don't care about protecting children. They merely see a way to consolidate power and are jumping on it.
reply
girvo
5 hours ago
[-]
> shadowy cabal of pedophiles elites

Its a shame that this used to just be a conspiracy theory one could mostly ignore, but we simply can't pretend that there isn't rampant CSA by those in power, because we've had proof of it despite their best efforts. Without wanting to get into politics, the leader of the United States right now was friends with the supposed ring-leader...

> but what about everyone else pushing for the law but who didn't go?

Useful idiots, perhaps? Wanting to protect their own power and gain more?

It's certainly not actually about protecting children. Never has been.

reply
Tarq0n
4 hours ago
[-]
I don't like the "those in power" framing because it implies that they all participated and that such a homogenous group even exists.
reply
foltik
4 hours ago
[-]
In the USA it literally is two homogenous groups though? One of which is majorly complicit in covering up the files, against their constituents’ wishes.
reply
scotty79
3 hours ago
[-]
I wouldn't even call them two groups. It's just one group ostensibly and publicly split in half, but it's still one group that intermingles behind the courtains.
reply
DaSHacka
3 hours ago
[-]
I would say both parties are complicit at this point.

Keep in mind Epstein died in 2017. We had two GOP terms and one Democrat term from then to now.

With what we know from the files that have been released thus far (and how obviously the worst if it has either been shredded or will never see the light of day), the fact they refused to release/prosecute those implicated tells you all you need to know.

reply
smallmancontrov
4 hours ago
[-]
Yes, and many people have an extreme incentive to retreat to that framing because

* In 2024, they had a choice between pedophiles and not pedophiles and chose the pedophiles.

* In 2020, they had a choice between pedophiles and not pedophiles and chose the pedophiles.

* In 2016, they had a choice between pedophiles and not pedophiles and chose the pedophiles.

There was plenty of evidence of this association in 2016 (bragging about creeping into Ms Teen USA dressing rooms, bragging about being Epstein's best friend in the same sentence as acknowledging he's a pedo, victim testimony under oath that he diddled kids, etc etc), so "I didn't know" isn't an excuse if they cared one iota about the children at any step of the way.

It should be good news that the powerful pedophiles are largely (but not exclusively) concentrated in one party, but those who put them in power will do anything to avoid admitting culpability.

reply
pipes
4 hours ago
[-]
I might be misreading you, but are you saying that the whole Qanon thing isn't a baseless conspiracy theory?
reply
girvo
3 hours ago
[-]
Qanon is absolutely a baseless conspiracy theory.

The overall idea that far too many of those in power politically and economically are involved in CSA isn't though, it seems.

reply
Dylan16807
56 minutes ago
[-]
The threshold for "far too many" is like, a single digit number. It's an extremely weak claim. Even if those in power were half as likely to be involved as the average adult, that would still be far too many.
reply
gruez
4 hours ago
[-]
>we simply can't pretend that there isn't rampant CSA by those in power, because we've had proof of it despite their best efforts

What's "rampant"? The news coverage provides no shortage of people, but ringing off 100 (or whatever) people that are in the files doesn't say much, even if we make the questionable assumption that inclusion in files implies guilt. I'm sure that everyone would prefer the amount of pedophiles that are in power to 0, but if it's the same rate as the general population that can hardly be considered "rampant", or a "conspiracy". Given some neutral inclusion criteria (eg. members of legislative bodies), is there any evidence they have disproportionate amount of pedophiles?

>the leader of the United States right now was friends with the supposed ring-leader...

You conveniently omit the fact that they broke up 5 years before he was first convicted. From wikipedia:

"Trump had a falling out with Epstein around 2004 and ceased contact. After Epstein was said to have sexually harassed a teenage daughter of another Mar-a-Lago member in 2007, Trump banned him from the club. "

>Useful idiots, perhaps?

So basically https://en.wikipedia.org/wiki/False_consciousness?

> Wanting to protect their own power and gain more?

How does adding age verification help in that? Are they blackmailed by the shadowy cabal? Are they just doing what the voters/lobbyists want? If so, what makes invocation of this reasoning more suitable than for any other political issue? Is everything from tax policy to noise ordinances just something pushed by pedophile elites, helped by useful idiots and people who want to "protect their own power and gain more"?

reply
foltik
4 hours ago
[-]
You sure are giving them quite the benefit of the doubt. Why?

https://www.pbs.org/newshour/amp/politics/see-the-alleged-tr...

reply
gruez
4 hours ago
[-]
>https://www.pbs.org/newshour/amp/politics/see-the-alleged-tr...

1. "alleged"

2. I'm not sure what you're trying to refute. I specifically quoted a passage saying that they broke up in 2004, which implies they were together prior to that.

3. For the specific claim that Trump's a pedophile, a "drawing of a curvaceous woman" is hardly proof. At best it's a proof that he's a womanizer, but we hardly need proof of that given the "grab her by the pussy" quote.

reply
foltik
1 hour ago
[-]
Oh nothing, just wanted you to see how you’d attempt to rationalize away what you see with your own eyes.
reply
gruez
54 minutes ago
[-]
Sounds like you don't have any arguments, if you have to resort to thinly veiled personal attacks.
reply
foltik
1 minute ago
[-]
Sounds like you have no argument besides “I don’t believe the evidence.”
reply
gosub100
4 hours ago
[-]
The resistance to the release of the files including redactions and outright refusal of Congressional order is enough to reveal the magnitude of what's going on. I would even dare say this Iran war is in part due to blackmail gained on DJT.
reply
gruez
3 hours ago
[-]
>The resistance to the release of the files including redactions and outright refusal of Congressional order is enough to reveal the magnitude of what's going on.

I agree this makes him look suspect, but it's hardly conclusive. Moreover Democrats did a similar U-turn a few years before. The only difference is that they weren't bombastically pushing the conspiracy theory during the election campaign, which made it easier for them to backtrack later.

>When Maxwell was charged in 2020, Democrats continued to push for transparency. [...] After Biden took office in 2021, Democrats appeared to dial back their public calls for Epstein records’ release.

https://www.aljazeera.com/news/2025/8/6/fact-check-did-democ...

reply
gosub100
3 hours ago
[-]
I dont disagree with anything you have written in the above reply. But why does democrats' reversal somehow annul or invalidate the claim about trump? Do you honestly believe it could all be an overly-embellished fable? If you do, then do you think the hundreds, some-say-thousands, of women who claim they were raped are lying?

Another explanation could be the democrats' AIPAC handlers told them to back off because it wasn't the precise time to leverage the material yet.

reply
gruez
1 hour ago
[-]
>But why does democrats' reversal somehow annul or invalidate the claim about trump? Do you honestly believe it could all be an overly-embellished fable? If you do, then do you think the hundreds, some-say-thousands, of women who claim they were raped are lying?

What claim about Trump? That's he's a pedophile? Based on the rest of your comment it seems like the goalposts are subtly getting moved from "Trump raped kids" to "Trump committed sexual crimes".

reply
theshackleford
4 hours ago
[-]
> You conveniently omit the fact that they broke up 5 years before he was first convicted.

And? It doesn’t change the reality of the original statement.

The president of the United States was friends with the alleged ring leader of a large pedophile network.

reply
GeorgeWBasic
1 hour ago
[-]
And more importantly, resisted releasing the files as hard as he could.
reply
gruez
3 hours ago
[-]
>The president of the United States was friends with the alleged ring leader of a large pedophile network.

You're making some leaps logic here here. If someone's outed as a pedophile, everyone who's friends with him should be assumed to be a pedophile? Surely not, given that pedophilia is considered taboo, we'd expect them to hide it, and therefore at least some friends might not be in the know. That's not to say there's no conspirators, but "he was friends with a pedophile therefore he's a pedophile too" is just guilt by association. What you need to prove is that he knew, or ought to have known that his friend was a pedophile. A conviction works decently for this, because it's presumably public knowledge, although even that's questionable because most people don't do a background check on people they met. In the case of Epstien he also hired reputation management firms to suppress his conviction from showing up in the results, which weakens the case even more.

reply
theshackleford
2 hours ago
[-]
> You're making some leaps logic here here.

No, you’re just shifting the goalposts.

The original claim was “The president of the United States was friends with the supposed leader of a pedophile ring.”

Your response to that was to imply that over time, they had a falling out. To which my point was, so what? It doesn’t materially change the original claim you challenged.

A falling out in NO way changes that the original statement was correct, the current president of the United States, Donald J Trump, was good friends with the alleged leader of a large scale pedophile network.

> If someone's outed as a pedophile, everyone who's friends with him should be assumed to be a pedophile?

If a given friend had their own history of acting like a creepy sex pest when it comes to young women, had a known and close relationship with the alleged leader of a pedophile network AND knew about “the girls”, would I assume them also to be a pedophile? At a minimum, I may in fact conclude that the odds they are also a pedophile are significantly higher than that of the average individual. Birds of a feather and all…

It’s not to say they are of course and it may in fact be as simple as they are nothing more than a creepy sex pest with a bad taste in friends, but NOT a pedophile. I gotta be honest but, me personally, I’d rather be neither.

reply
gruez
55 minutes ago
[-]
>To which my point was, so what? It doesn’t materially change the original claim you challenged. [...]

It changes the claim in the same way that "he ran over a kid" isn't "materially changed" by the addition of the detail that the kid jumped in front of the car and he had no time to stop. The original statement is still technically true, but it's a massive omission to leave the latter part out. That's doubly true if you're invoking that fact in the context of trying to imply the person did other crimes.

reply
micromacrofoot
4 hours ago
[-]
you mean the guys who are working alongside a bunch of pedophiles and doing little about it?
reply
afh1
5 hours ago
[-]
Those are just stupid.
reply
aga98mtl
4 hours ago
[-]
> but what about everyone else pushing for the law but who didn't go?

Who exactly is influential & organized enough across many western countries to push legislation that no one is asking for? Notice that epstein said he worked for [withheld] in some of his emails.

reply
gruez
4 hours ago
[-]
>Who exactly is influential & organized enough across many western countries to push legislation that no one is asking for?

The anti-social media sentiment has been brewing for a while now, not least due to books like The Anxious Generation (2024). It's also reflected in opinion polls and media coverage. Unless you want to imply there's some massive conspiracy by The Elites™ (ie. not just a few lobbyists Meta hired, but those in academia and media as well), it's probably organic.

reply
smallmancontrov
5 hours ago
[-]
I don't know the precise combination of stupidity vs evil that compelled the "think of the children" crowd to choose the single most publicly implicated man in the Epstein scandal as their champion and elect him over someone who wasn't and hasn't been implicated at all in the slightest, but they did. Either way, they receive the culpability for doing so and we should expect their future decision making to be equally compromised.
reply
Velocifyer
4 hours ago
[-]
Which island?
reply
jadamson
4 hours ago
[-]
Little St. James
reply
nssnsjsjsjs
3 hours ago
[-]
Epstein's most likely.
reply
cdrnsf
3 hours ago
[-]
The people pushing these bills are the same that are looking to ban library books. They’re either bad or ineffective parents (or both). Instead of having a healthy relationship and discussion with their kids they’d rather impose their own regressive ideas by way of legislation on everyone.
reply
rustyhancock
4 hours ago
[-]
Quite mind boggling to me that a nanny state can exhurt such a large amount of global control.

It's darkly comedic that the single most toxic experience since the pop up ad - the cookie consent popup was similarly imposed.

The solution is simple. Websites and services (including ISPs) become governed by the country in which they operate not the whims of foreign entities.

reply
bluegatty
3 hours ago
[-]
Where do all these people come from?

The 'nanny state' prevents people from driving cars without a license?

That prevents you from buying myriad substances without a note from the doctor?

That makes it illegal for you to buy a gun?

" become governed by the country in which they operate not the whims of foreign entities"

... is not going to work, at face value, because 'operation' involves the consumer and the producer, each of whom may be in different jurisdictions, and even if they were in the 'same nation' ... this is still a hard problem.

No easy answers, and there are legit concerns.

reply
AvAn12
2 hours ago
[-]
You can’t determine age from a face scan. And it’s trivial to hold up a photo of an older person. Seriously if a website wants an image of your government ID or facial image, maybe ask yourself if you really need to access that site.

There WILL be breaches and those drivers license scans will get loose in the world sooner or later. Fully agree that this is all about access control. No thank you.

reply
funnybookbinder
10 minutes ago
[-]
ban porn altogether
reply
nirui
46 minutes ago
[-]
You all saw the Epstein scandal, right? If you saw one cockroach this randomly, then you know there are thousand hiding. Maybe that's why Epstein is un-lived.

So I found it very ionic that, to quote on quote "protect" child from online harms, they asks you to upload the photo ID of you and your child to, guess what, real potential pedophiles.

Of course they're going to claim your information is totally safe... just like Bill Gates told his wife it's safe to have sex with him after his STD infestation.

Sure, I don't really know how the companies will actually handle your personal photos, but there's a history where a tech CEO made an attractiveness comparison website using photo obtained from their user uploads without user agreeing. So go figure.

The best way to protect your child is to tech them how to use Internet for their own benefit, and only allow them to create accounts after they've learned how to use Internet correctly. The companies and governments will NEVER do that for you, they'll only steal and steal even more.

reply
wewewedxfgdf
6 hours ago
[-]
You must be crazy, who could possibly object to governments "protecting the children"?
reply
skybrian
3 hours ago
[-]
Devices with child locks turned on really shouldn't have access to everything on the Internet. A simple protocol could let cooperating websites know when child locks are on, so they don't show inappropriate content. Whitelisting or blacklisting could handle the rest.

This doesn't mean every device needs to implement child locks. It also shouldn't affect anyone using unlocked devices at all.

reply
AJ007
2 hours ago
[-]
How does that even begin to make sense?

I want to protect my child from X type of content -- one of many jobs of a parent, but I will trust all content to self report to be child inappropriate? "Inappropriate" is entirely subjective and can not be defined as some sort universal bool -- and that's before you get to the point of actively malicious actors like Meta and Tiktok actively exploiting children for their content farms generation and ad impression factories.

If the user owns and controls their computers -- as they should -- then that subjective content filtering layer belongs there, in the owners control. If its a child's, then the parent owns the device, not the child.

reply
skybrian
2 hours ago
[-]
The idea is that society should have some common standards for what's inappropriate for children. For example, parents don't want their kids to buy cigarettes, but also, stores don't want to sell them cigarettes. When there's consensus on this, cooperation is possible. Parents have an easier time when they get cooperation from the rest of society.

But there isn't going to be consensus on everything, so content filters are still needed.

reply
gzread
2 hours ago
[-]
> a simple protocol could let cooperating websites know when child locks are on, so they don't show inappropriate content.

Isn't that literally the California law?

reply
skybrian
2 hours ago
[-]
Not which law you mean, but I think there's a distinction between "disallows children under 16 from creating an account" (which apparently requires age verification) and "disallows creating or logging into a social media account from a device with a child lock on." (Which doesn't.)
reply
gzread
2 hours ago
[-]
The California law is the one where parents select how old their child is and apps must respect that as a child lock.
reply
jacquesm
2 hours ago
[-]
For almost three decades authorities have been wondering how to put this 'free communications' genie back into the bottle without taking the GFW approach. It looks like this time they just might get it.

If you really believe that this is about child protection then you are much too gullible, that was never the main reason. If the authorities really wanted to do something about child protection online they'd spend a fraction of what they are going to spend on this on building out the departments in the various countries that actually work on that problem exclusively. As it is they have more work they can handle, which leaves a lot of cases lying and far more of these perps active than what would otherwise be the case.

So as long as you don't see that you know for a fact that this child protection is not the real reason.

reply
SilverElfin
2 hours ago
[-]
It is all about advertising profits. See what I said at https://news.ycombinator.com/item?id=47471747

But what do we do about it? Look at social media comments on this topic. There’s huge support for these age verification laws. Parents chime in about how their kids were affected by social media and how badly this is needed, instead of taking responsibility for raising their kids properly. That article by the Pinterest CEO calling for these laws is naively seen as some sort of sacrifice of profits for the good of everyone’s kids. And no one talks about privacy or the effects on speech.

And all these well funded nonprofits pushing these dishonest bills onto legislators have time and resources. Feels like the privacy friendly people are losing the battle.

reply
mrcwinn
16 minutes ago
[-]
This happening and it can’t be stopped. There is bipartisan consensus, so hard to come by otherwise, because both parties share the same corporate interests. Deal with it.
reply
1vuio0pswjnm7
3 hours ago
[-]
Controlling access to certain websites, i.e., so-called "social media", is not "internet access control". The web is not the internet. Nor are these laws limiting access to _all_ websites. Third, not all operating systems are controlled by corporations like Apple, Google, etc. and used to protect and promote corporate interests
reply
braiamp
3 hours ago
[-]
I fail to see why the "protections" that child data deserves, isn't also the same kind of protection that everyone deserve. In what way are children special, in a digital world, that adults shouldn't be protected the same way?
reply
gzread
2 hours ago
[-]
In this world where we are committed to not making laws affecting the freedom of an adult in any way, even to be forced to consent to things.
reply
kepeko
5 hours ago
[-]
Maybe the positive is that access control might break the illusion of privacy.

Okay it's quite private in the sense that we don't know our friends browsing history but we know somebody, somewhere is collecting data and selling it to their 100 partners.

Do you think there might ever be a moment when someone decides, legally or not, dump enormous amount of info, in a way that allows people to see what google searches other people did or browsing history etc? A moment when people's embarrassing secrets come into light.

reply
andai
5 hours ago
[-]
Saw a mini documentary once, which was filmed in China, that showed how easy it was to buy this data. Many apps spy on location and sells it to brokers. In the documentary, they showed a common practice: people buying their romantic partner's location history to make sure they haven't been doing anything naughty.
reply
squarefoot
6 hours ago
[-]
Access control and pervasive surveillance has been the plan since day one; child protection is the leverage. Also, I don't expect people who repeatedly hide the contents of certain files to care about children.
reply
1970-01-01
3 hours ago
[-]
I'd be ok with this if both ends of the spectrum were covered. Sorry, you're too old to access this computer. Go ask a younger adult if you want to read the news or see photos of your grandkids.
reply
Diffusion3166
5 hours ago
[-]
Given that it seems Meta is commissioning these laws, I wonder if a viral open source license that explicitly fails to grant Meta a license to use or modify the software would effectively deter future lobbying for regulations which are especially difficult for the open source community to comply with.
reply
tzs
4 hours ago
[-]
> Given that it seems Meta is commissioning these laws

That's not given. Someone found some good evidence that Meta was supporting (and even supplying language) for some of the earlier laws. Those were the laws doing age checks on websites and typically requiring uploading ID documents or face scans to those websites.

I've not seen anyone provide evidence that Meta has anything to do with the laws that are like the California one, which do not require providing any documentation or proof whatsoever of age. They just required that the parent of a child who uses a device be asked to provide a birthdate or age when setting up the child's account, and that the OS providing an API that apps on that device can use to get the age bracket of the child.

reply
baal80spam
6 hours ago
[-]
It was never about children...
reply
abcde666777
4 hours ago
[-]
The more people that use something the more it inevitably trends toward average mediocrity.

A lot of these trajectories aren't really for us - the techy folk.

reply
Beestie
4 hours ago
[-]
Well age verification works so well to keep alcohol, tobacco and weed beyond the reach of minors so....
reply
superkuh
2 hours ago
[-]
Here we see the danger of the lay perception that multi-media screens are the same as chemical drugs. They are not. Not even close.
reply
cat-turner
5 hours ago
[-]
parents need to do their job and raise their children, and moderate their content.
reply
mamami
4 hours ago
[-]
Because of course it's so easy. You obviously have never visited a site your parents would have disapproved of
reply
Ylpertnodi
5 hours ago
[-]
Whose great-grandparents are you going to blame?
reply
k33n
3 hours ago
[-]
We had a good run when the internet was a disruptive force. But mass adoption of anything always leads to where we are. The internet is an established institution. The wild west days are over. If you're looking for that vibe, p2p technology in small corners will be where you can find it.
reply
ginko
4 hours ago
[-]
Just ban children from using the internet.
reply
einpoklum
5 hours ago
[-]
But the whole point of bringing up child protection was to restrict Internet access, to police Internet content and to legitimize mass surveillance.

Or do we really believe that states which condone support, fund and sometimes engage in the mass killings children are motivated by genuine moral concern for the young?

-----

Still, there is somewhat of a silver lining: Perhaps this will encourage young people, and people who value their privacy, to avoid those "social networks" in favor of places where there is no age verification, 2FA with a physical phone number, etc. etc.

reply
varispeed
5 hours ago
[-]
The people who want to control internet access use children to achieve their means. Why these creeps get to power? Normally people thinking too much about children would be casted out of society at best.
reply
dzogchen
4 hours ago
[-]
Am I the only one that simply disregards everything that follows an AI slop image?
reply
tzs
3 hours ago
[-]
Is there something wrong in particular with that image? The composition fits in well with the content of the article, and the art seems pretty well down. I'm only seeing one error that would make me think "a human probably didn't draw this" and took a while to notice that.
reply
mamami
4 hours ago
[-]
You don't understand, the children need to be exposed to Nick Fuentes, Andrew Tate, and algorithmically generated suicidal ideation from Facebook. It's crucial for their development, actually
reply
kgwxd
4 hours ago
[-]
The only people on the planet that care about this, and understand it enough to maybe do something about it, are reading this thread right now. I got nothing. Anyone else got any ideas?
reply
bfivyvysj
4 hours ago
[-]
Too late

- Australia

reply
TomGarden
4 hours ago
[-]
For many it's not about the children. For many it is.

I haven't made my mind up on this topic, but Jesus, the comments here strawmanning everyone who supports this kind of thing as disingenuous or worse... Wow.

I'm not sure how we make any corner of the internet usable within the next few years without verification given all the misinfo, bots & AI slop anyway.

reply
windowliker
5 hours ago
[-]
Arguments about erosion of privacy miss the point: that is exactly what they want.
reply
borissk
6 hours ago
[-]
The big tech is going to be one of the big winners from Internet Access Control. This will give them a more reliable way to link a user account to an actual human being - a link that can be monetized in a variety of ways. All kind of political regimes can use such regulations to enhance their control of the population. And the loosers are going to be the Internet users and small companies.

The unfortunate true is IAC is coming to most countries in the world, no matter how much the Hacker News audience hates it...

reply
pstuart
4 hours ago
[-]
The moment "think of the children!" enters the chat is when suspicions should be heightened.
reply
psyclobe
4 hours ago
[-]
If you think u can control a kids imagination to circumvent these controls then you are part of the problem
reply
SilverElfin
5 hours ago
[-]
I read in some other discussions that this is about social media companies being able to increase their profits and nothing else. But the social media companies lobbying for these laws are shamelessly making it look like some kind of protect the children thing. It is all pushing more ads annd getting more users.

The way it works: today, social media companies cannot advertise to children under 13 under COPPA. So these companies have to do their best to guess the user’s age, and if it is possibly a child, they can’t advertise and have to lose those profits even though MAYBE the user is an adult. Now they can shift the legal compliance costs and liability to the operating system provider or phone manufacturer and not be responsible for the user’s identity. And then they can advertise much more at that point, without being conservative. This also lets them have a different experience for minors that doesn’t advertise to them, but targets them carefully to keep them as users until they are older, so they start to become a source of advertising profits later.

It’s well known that Meta is behind a lot of funding for nonprofits pushing these laws under a “protect the children” thing. But now even Pinterest’s CEO is shamelessly saying parents don’t have a responsibility to manage their own kids, and is supporting all of this. See https://www.gadgetreview.com/reddit-user-uncovers-who-is-beh... and https://time.com/article/2026/03/19/pinterest-ceo-government...

Evangelist/theocratic conservatives welcome these laws because they view it as enabling and validating age-based restrictions for other things. For example, Project 2025 called for a ban on porn. And separately, the Heritage Foundation pushed age-verification for porn websites, and has openly admitted it is a defacto porn ban. That should have been ruled unconstitutional on free speech grounds, but the current SCOTUS upheld it unfortunately. They’ll next use age-based verification for all sorts of content - maybe for LGBTQ stuff, maybe for something else.

In the end, everyone else will lose. If you have to prove your identity to anyone, there is a high chance this information can be accessed and surveilled by the government. There is a high chance at some point, no matter what they claim, your identity data will be hacked and sold. And of course if you can be identified online, then anything you say or do can be traced back to you, and that can be used against you by the government. Suddenly, being a protester in these chaotic times will become a lot more risky.

reply
holyhnhell
5 hours ago
[-]
I’m okay with internet access control if it means less AI slop like this shit. Bring it on. I’ll be there when it happens.
reply
amarant
5 hours ago
[-]
Why would IAC lead to less slop? What's the mechanism here?
reply
charcircuit
1 hour ago
[-]
Imagine if it was possible to ban people who spam instead of only being able to ban the IP of a spammer.
reply
amarant
42 minutes ago
[-]
So IAC is magically removing slop AND spam from the internet? It's still not clear how? Are we always broadcasting our government issued identity documents to every website we visit? I don't think that's gonna work out the way you think it's gonna work out
reply
kogasa240p
5 hours ago
[-]
Lol no
reply