Online age verification is the hill to die on

(x.com)

344 points | by Cider9986 2 hours ago

41 comments

  • Bender 2 hours ago
    The one and only method I will participate in is server operators setting a RTA header [1] for URL's that may contain adult or user-generated or user-contributed content and the clients having the option to detect that header and trigger parental controls if they are enabled by the device owner. That should suffice to protect most small children. Teens will always get around anything anyone implements as they are already doing. RTA headers are not perfect, nothing is nor ever will be but there is absolutely no tracking or leaking data involved. Governments could easily hire contractors to scan sites for the lack of that header and fine sites not participating into oblivion.

    I a small server operator and a client of the internet will not participate in any other methods period, full-stop. Make simple logical and rational laws around RTA headers and I will participate. Many sites already voluntarily add this header. It is trivial to implement. Many questions and a lengthy discussion occurred here [1]. I doubt my little private and semi-private sites would be noticed but one day it may come to that at which point it's back into semi-private Tinc open source VPN meshes for my friends and I.

    [1] - https://news.ycombinator.com/item?id=46152074

    • hooverlabs 12 minutes ago
      Servers can then infer user’s ages by whether or not the client renders pages given those headers or not no? See if secondary page requests (e.g images, scripts) are made or not from a client? A bad actor could use this to glean age information from the client and see whether the person viewing the page is a small child. That should be scary
      • Bender 3 minutes ago
        I disagree. The ability to render a page could simply mean that parental controls were not enabled on the device. Some parents trust their children to be psychologically ready for adult situations. The client could be literally any age.

        Today devices do not default to accounts being child accounts. Some day this may change and may require an initial administrator password or something to that affect but this can evolve over time.

    • kevin_thibedeau 9 minutes ago
      > fine sites not participating into oblivion.

      That would also amount to compelled speech.

      • Ekaros 2 minutes ago
        Clients could refuse to show content that does not have headers set.

        On other hand servers might choose to lie. After all that is their free speech right.

        So maybe you need some third party vetting list. Ofc, that one should be fully liable for any damages misclassification can cause... But someone would step up.

      • Bender 6 minutes ago
        I disagree. The legal requirement to apply a warning label is a well known, understood and accepted process that is applied to a myriad of hazards to children and adults.
    • big85 1 hour ago
      Back in the late 90s or so, there was a proposal to have sites voluntarily set an age header, so parents/employers/etc could use to block the site if they wish. People said it would never work, because adult sites had a financial incentive not to opt in to reduce their own traffic.
      • masfuerte 1 hour ago
        The porn companies already set the RTA header. It was designed by an organisation funded by the porn companies.

        https://en.wikipedia.org/wiki/Association_of_Sites_Advocatin...

        • motbus3 45 minutes ago
          It seems there is a GitHub repo somewhere mapping Meta money to lobbyists inside other companies Which is at least interesting
      • thesuitonym 1 hour ago
        What, in the same way movie studios wouldn't comply with the Hayes Code, or comic book publishers wouldn't comply with the CCA, or games publishers wouldn't comply with the ESRB? The financial incentive is to police yourself, because government policing is much, much worse.
        • nine_k 1 hour ago
          There's a great relevant quip: "If you think that the cost of compliance is high, try noncompliance".
        • breezybottom 15 minutes ago
          Sure but the government doesn't police corporations in the US anymore. The Hayes code was before neoliberalism.
          • shevy-java 13 minutes ago
            Quite true. The US corporations act like a giant global rabid dog. Fake legislation appears in the USA - lo and behold, it is copy/pasted into the EU. At the least lobbyists are getting rich right now.
            • htek 4 minutes ago
              At least the EU has GDPR. In the US, our personal data is collected by every app and website and company and packaged, sold and sifted through by a vast collection of private data brokers which the government already ingests.
      • btilly 8 minutes ago
        People were wrong.

        We pay money online mostly through credit cards. Credit card transactions can be reversed. If children spend money on porn, those payments are likely to be reversed. This is really bad for the ability of the porn sites to continue receiving credit card payments, and continue making money.

        An age header is a trivial step that can reduce the odds of the adult site receiving payments that later get reversed. Win, win.

        But if someone is willing and able to pay, then the adult industry wants the choice of whether to access content to be up to them. If government tries to regulate them, they'll engage in malicious compliance - do the minimum to not be sued, in a way that they can still reach customers.

        For example Utah tried to institute age verification. The porn industry blocked all IP addresses from Utah. Business boomed for VPN companies in Utah. Everyone, including porn companies, knows that a lot of that is for porn. But if you show up with a Nevada IP address, the porn's position is, "You're in Nevada. Utah law doesn't apply." Even if the credit card has a Utah zip code.

        If you live in Utah, and you're able to purchase a VPN, the porn companies want your money.

      • iamnothere 1 hour ago
        You’d think that one could simply block sites that don’t have the age header set on child computers. This may block kids from hobbyist sites that don’t bother to set their headers as kid-friendly, but commercial sites would surely set their headers properly. Over time sending proper rating headers would become more normalized if they were in common use.

        This still isn’t perfect, as it creates an incentive for legislators to criminalize improper age header settings and legislate what is considered kid-appropriate. But it’s still better than this age verification crap.

        • Scaled 15 minutes ago
          Yes, that's how parental filters already work. They use a combination of rta tags and external data to block pages. Even works with Google safe search, firewall devices, etc. The rta ecosystem is already built out and viable.
      • Bender 1 hour ago
        What I am suggesting could address most of that. If they do not participate they get fined. The government loves to fine companies. This assumes they put enough "teeth" into a law that prevents companies from accepting fines as the cost of doing business. This would also require legislation that could block sites that operate from countries that do not cooperate with US laws. Mandatory subscriptions to BGP AS path filters, CDN block-lists which already exist, etc... People could still bypass such restrictions with a VPN but that would not apply to most small children. Sanctions and embargoes are always an option.
        • Barbing 1 hour ago
          >fined

          Exactly. If you’re hurting kids to make more money selling porn videos, straight to jail.

          I’m glad there are solutions that won’t ruin the Internet. Now the uphill battle to convince our legislators (see: encryption & fundamentally technically ignorant calls for backdoors).

          I’m here to die on this hill!

      • Lammy 1 hour ago
        > Back in the late 90s or so, there was a proposal

        This one: https://www.w3.org/PICS/

        • Bender 1 hour ago
          PICS was very complicated and attenpted to cover all possible "categories" of adult content. It was confusing, incomplete and only a handful of sites voluntarily labelled their sites with it. RTA is one simple static header that any site operator could add in seconds unless they get more complicated with it by dynamically adding it to individual videos say, on Youtube which means in that case the server application would need to send that header for any video tagged as adult.

          I added PICS to my forums but it was missing many categories of adult content. I ended up just selecting everything as I could not predict what people may upload which made for a very long header.

          • dylan604 34 minutes ago
            > unless they get more complicated with it by dynamically adding it to individual videos say, on Youtube

            YT already does this. I never watch YT signed in, and I often see videos that require you to be logged in as the video is age restricted.

            • Bender 19 minutes ago
              Agreed though in my example the point would be to set the header in the case the child is logged in but for whatever reason the site does not know their age. Instead of a third party site, a header is sent with the video tagged as adult that triggers parental controls.
    • _ink_ 1 hour ago
      How are they supposed to fine sites out of their jurisdiction?
      • Bender 1 hour ago
        One possible method [1] though I am sure the network and security engineers here on HN could come up with simpler methods. Just blocking domains on the popular CDN's would kill access for most people as by default most browsers are using them for DoH DNS.

        [1] - https://news.ycombinator.com/item?id=47950843

        • filoleg 25 minutes ago
          The question was about fining entities outside of the original jurisdiction, so I am not sure what you have in mind that could be done by network/security engineers here.
          • Bender 18 minutes ago
            In terms of fines if they do not pay the fine their country is at risk of sanctions or embargoes which is probably a bit heavy handed but may incentivize their government to also enforce the rules, collect fines keeping some for themselves and passing the original fine back to the countries implementing child safety controls.
    • duped 58 minutes ago
      This doesn't address the wider array of age-verification related problems that people want to solve, like social media where age verification is needed to police interactions between users.
      • jdasdf 16 minutes ago
        Such censorship shouldn't exist in the first place.
      • Bender 44 minutes ago
        I could be misunderstanding the context but to me that sounds like a moderation issue assuming we even want small children on social media in the first place. There should probably be a dedicated child-safe social media site that limits what communication can take place for small children and has severe punishments for adults pretending to be children.
      • svachalek 44 minutes ago
        This is assuming children should be on social media at all, which I for one would debate.
  • ronsor 1 hour ago
    There's an angle everyone misses.

    Mandatory age surveillance everywhere is only going to result in massive, normalized ID fraud. You thought fake and stolen IDs were a problem before? You haven't seen anything yet.

    And half of it will be from adults trying to avoid privacy invasion.

    • dylan604 29 minutes ago
      Not so sure about that. Handing an ID to a bouncer at a bar or similar is not logging anything. Mainly it's some big man that you can see gears turning to see if the date is correct and a cursory glance to see if the photo matches. Sophisticated places might have a scanner that does what ever validation it does, but again, it's just another cursory check of the photo. Most of these people really don't care.

      A tech company doing scans for validation could actually connect to a state database to verify the ID is legit and is not already being used for a different account. It would then be saved. I don't think real world vs tech world usage of fake IDs are the same at all.

      • schnitzelstoat 7 minutes ago
        The tech companies care even less than the bouncers do.

        They just want a plausible defence should it ever end up in court.

      • chimeracoder 8 minutes ago
        > Not so sure about that. Handing an ID to a bouncer at a bar or similar is not logging anything.

        > Sophisticated places might have a scanner that does what ever validation it does, but again, it's just another cursory check of the photo.

        Many/most bars do scan IDs now. Ostensibly it's to verify that it's real, but they do use those systems to keep a log of everyone who enters.

  • mzmzmzm 3 minutes ago
    If you don't use X/Twitter anymore, XCancel makes it possible to read threads when not logged in: https://xcancel.com/GlennMeder/status/2049088498163216560
  • goda90 1 hour ago
    Age verification can be achieved without destroying anonymity and privacy online using anonymous credential systems, but it has to be designed that way from the ground up, and no one pushing age verification is interested in preserving privacy.
    • Aurornis 1 hour ago
      This comes up in every thread, but the purpose of the laws is not to verify that someone can access an anonymous token. If we had a true anonymous token system then everyone would just share tokens around.

      The real world analog would be if you could buy beer at the store with anyone's ID because they didn't make any effort to reasonably check that the ID was yours or discourage people from sharing or copying IDs.

      The systems enforce identity checking because that's the only way age verification can be done without having some reason to discourage or detect credential sharing.

      The retort that follows is always "Well it's not perfect. Nothing is perfect." The trap is convincing ourselves that a severely imperfect system would be accepted. What would really happen is that it would be the trojan horse to get everyone on board with age verification, then the laws would be changed to make them more strict.

      • miloignis 47 minutes ago
        Matthew Green talks about this in his blog on the subject: https://blog.cryptographyengineering.com/2026/03/02/anonymou...

        The two methods that seem feasible are making it hard to copy (putting it in the secure element in your phone, for example, which I don't love) or doing tokens that can only be used a limited number of times per day, like in : https://eprint.iacr.org/2006/454

      • goda90 1 hour ago
        Make it a duplication resistant hardware token that you can get for free then. The stakes just aren't high enough to worry about these kinds of edge cases.
        • dpark 1 hour ago
          Yeah, right. So the government is going to spend billions on “porn tokens”. That’s going to get through the legislature.

          I’m sure there wouldn’t be a brisk illicit trade in these tokens either. Certainly no one would be incentivized to sell these tokens to teenagers for easy profit.

          • snackbroken 30 minutes ago
            Further, "porn tokens" are the pointy end of the wedge, because it's easy to misconstrue any opposition as advocating for "kids should have access to porn, actually". The broad end that is being hammered towards is "kids aren't allowed on social media because it's harmful to them" AKA "free speech tokens".
        • akersten 1 hour ago
          The stakes just aren't high enough for us to implement any of this crap for the Internet in the first place. Let alone an entire government-administered hardware supply chain.
    • dpark 1 hour ago
      No it really can’t. Age verification requires identification.

      Even if you could anonymously verify age to issue a “confirmed adult” credential, the whole chain of trust breaks down if one bad actor shares their anonymous credential and suddenly everyone is verifiably an adult.

      The solution to that attack is naturally to have some kind of system for sites to report obviously-shared credentials. Which means tracking.

      • armchairhacker 6 minutes ago
        But likewise, someone can share (or have stolen) their ID

        https://news.ycombinator.com/item?id=47951372

      • goda90 43 minutes ago
        There's already authorities that know your age, so verifying age with them to get the credential isn't the part that needs to be anonymous. The issue is them knowing what you do with your credential, which anonymous credentials solves by making it impossible to track tokens back to the credential holder. As far as sharing, there are some possible mitigations.
        • dpark 35 minutes ago
          Right. And the possible sharing mitigations generally amount to tracking.

          This isn’t even getting to the issue that mandating government-issued credentials is the “foot in the door”. If you mandate the use of government creds for accessing websites, it’s an obvious step to turn around and demand that sites report credential use to “fight credential fraud”.

    • wesselbindt 1 hour ago
      The destruction of privacy is the whole point.
    • everdrive 1 hour ago
      This is something that's technologically feasible, but will never happen in practice.
    • Forgeties79 1 hour ago
      And they continue to act like opposition just wants a wild west/don't care about kids, which is the oldest trick in the book. We just don't want "protect the kids" leveraged to tear up our rights.

      It's addressing a real problem in a bad way.

      • jMyles 1 hour ago
        I mean, it's more than that. I _want_ to protect kids' right to be part of the human connectome. The "protect the kids" (by disallowing them their freedom of thought on the internet) is just naked ageism.
        • dpark 1 hour ago
          So do you want 5 year olds driving on the highway and 8 year olds doing shots of tequila or are you ageist?

          Or perhaps protecting kids isn’t really ageism at all.

          • Forgeties79 50 minutes ago
            It depends on what you're depriving them of too. Those are very extreme examples with little to no upside.
            • dpark 3 minutes ago
              Disagree. We can discuss what restrictions are appropriate or reasonable without calling it ageism.

              Calling it ageism is an emotional appeal, not a principled stance.

              • Forgeties79 2 minutes ago
                Ageism is a legally defined form of discrimination as well as the subject of ethical discussions. It's a real, defined thing. Just because we disagree on what qualifies as ageism doesn't mean you get to call foul and say it's irrational/emotional.
          • IAmBroom 35 minutes ago
            Quit arguing as though the topic is binary. It's not.
            • dpark 8 minutes ago
              I’m not saying anything is binary. I’m saying it’s not ageism to restrict child access. It could be a bad idea but that doesn’t make it ageism.
        • Forgeties79 1 hour ago
          Fair point
    • bigbugbag 1 hour ago
      the EU is. but their verification age process shows the design flaw that preserving privacy means the system can be easily circumvented with a mitm allowing to circumvent the age verification process.
      • kro 23 minutes ago
        Young people setting up a MITM and getting deeper into tech rather than consuming short-form-content is something I'd appreciate as a nice bonus effect.

        Of course the EU solution isn't perfect and there are bypasses (there will always be and have always been), but let's appreciate it that way rather than too many PII, if it must come. I'd prefer the Age/RTA header and parental responsibility too.

    • devmor 1 hour ago
      They are interested - interested specifically in opposing it. These groups don't care about age verification - it is a trojan horse for censorship.
    • intended 1 hour ago
      AFAIK there are designs in the EU that respect privacy. There is a range of options being pushed around the world, and theres definitely a few of them which are more technically defensible than others.
  • tim333 1 hour ago
    >age verification requires identity verification. Identity verification requires digital IDs. Digital IDs require everyone — not just children — to prove who they are before they can speak...

    Not if it's done in a half arsed way. I'm in the UK and so far my age verification has involved doing a selfie with the webcam for Reddit. That's it. No one needing my name, ID number etc. (Apart from banks of course).

    Really this is just the modern equivalent of putting the porn mags on the top shelf at the newsagent to stop the kids getting them. We don't need more.

    • afh1 9 minutes ago
      A photo identifies you. This is the digital equivalent of having a photo taken of you upon entering the mag store, stored digitally forever, shared with government, and tied to every magazine you read and purchase.
    • cmiles74 47 minutes ago
      In the US, the plan is to require adults to take a picture of their state ID and upload it to a third party that provides age verification. It's not explicitly part of the proposed law but there are only a handful of companies who meet the qualifications to provide this service (id.me, Persona) and this is how they do it.

      I believe if you are a "minor" then you can go the post-a-selfy route.

      • Buttons840 22 minutes ago
        If someone wanted to be a martyr and just uploaded all their personal documents so they could be accessed by everyone, I wonder if an interesting court case might follow.

        I could imagine it ending with a court ruling that people are responsible to protect their own personal documents which... yeah, that would muddy the waters in a world where every website expects to see your ID.

      • chimeracoder 2 minutes ago
        > In the US, the plan is to require adults to take a picture of their state ID and upload it to a third party that provides age verification.

        That's not just the plan - that's what's already legally required in many US states.

        These laws were introduced by the explicitly religious right-wing groups like Exodus Cry and Morality in Media, as ways to de facto outlaw pornography (in their own words). They've since been laundered into the mainstream so the general public is unaware of the root cause.

      • motbus3 42 minutes ago
        Imagine so if that was a pltr right Or like someone who uses pltr What could possibly go wrong? People are being paranoid for no reason!
    • mohamedkoubaa 14 minutes ago
      Whether it can be done this way is besides the point. It is about how regimes like ours in the US that have demonstrated an interest in spying on their subjects choose to regulate this over time.
    • conradfr 45 minutes ago
      Reddit is one thing but would you do the same for a porn site?
  • gslepak 4 minutes ago
    Good: some commenters here realize it's an attack on privacy

    Bad: some still entertain the idea that we should do age verification using some sort of crypto primitives

    There is no reason for age verification at all.

    I am from the goatse generation. Rotten.com. steakandcheese. Horrific stuff tbh, I mostly stayed away from it, and I didn't need a helicopter government to protect me from it.

    The moment you accept the narrative that kids need to be protected from the Internet you have already lost.

    You've already condemned those kids to a life of slavery. So much for protecting them.

    What we need is not online verification, but a competent government that does its existing job well.

    Who's been arrested over the Epstein files? Who is protecting those kids?

    No one.

    That same government wants to "protect" your kids by KYCing everyone.

    Give me a break.

  • wxw 2 hours ago
    How are folks recommended to get involved? Contact your local Congress member? I feel this thread has a lot of passion but is missing concrete, actionable steps.
    • Barbing 1 hour ago
      Heroes @ EFF have our guide (USA residents):

      https://www.eff.org/pages/help-us-fight-back#main-content

      • ethagnawl 1 hour ago
        Of course Chuck Schumer won't let me contact him using this helpful tool.

        Perhaps we NYers should organize a rally outside his office in Manhattan like we did for PIPA/SOPA?

        • Barbing 1 hour ago
          Dumb- BUT immediate links to sites of the right legislators!

            Adam B. Schiff
          
            Sorry, this legislator cannot be contacted with our tool. To message them, visit their website instead.
          
            Alex Padilla
          
            Sorry, this legislator cannot be contacted with our tool. To message them, visit their website instead.
        • Finnucane 1 hour ago
          Use every means necessary. If that can be organized, do it.
      • trueno 1 hour ago
        man the EFF owns
    • chainingsolid 1 hour ago
      I've contacted my congressmen and I would also advocate for telling/explaining this to non technical people you know. They either won't have heard of this or won't know whats bad about it.
      • Barbing 1 hour ago
        Any tips for writing the letter, maybe even a starting point?
  • cooper_ganglia 11 minutes ago
    THe government shouldn't be raising anyone's children, that's what parents are for. If you're a bad parent, your kids will get access to bad things and could become an adult failure.

    The future of your family and your legacy is up to you, not the government. We don't need age verification to restrict the social darwinism of raising children.

  • retired 1 hour ago
    Age verification on Australian social media has loopholes. Underage influencers use an agency to manage their social media for them. So anyone with enough followers or money can continue using social media under the age of 16.

    If you are going to implement age controls, you should implement a ban on underage influencers as well.

    • Barbing 1 hour ago
      How could one protect the, call it one in 1 million… the speech of the (young) Greta Thunbergs, for example?

      I bet there is a 15 year-old much smarter than me making political videos and I wouldn’t necessarily want them to be forced to stop. What if they’re on my “team”! ;) (I kid)

      Recalling how we had lots of political debates in high school: if some of those kids made videos and got really popular, and the law made them stop, they would have been incentivized to vote $responsibleParty out.

      (Socials bad for kids though maybe they could selfhost their monologues instead)

      • mystraline 1 hour ago
        I believe every government disenfranchises young people because they are young.

        Its not about intelligence. Else a whole lot of over-age-of-majority wouldn't pass either.

        Theres also no old-age cutoff, when their mental faculties significantly decline.

        Yeah, the voting majority keeps 'under age' from voting. But at least in the USA, we have children as young as 11 being tried as adults but with none of the benefits.

        • Barbing 56 minutes ago
          You’re right that it shouldn’t be about intelligence! Overall definitely unfair.

          After posting, I questioned whether political speech is special. Like should fifteen-year-olds who love film be able to make videos about them and get lots of followers… but I couldn’t be thought police. So maybe-

          The platform just has to be designed non-addictively.

          Is this accurate?: In reality, Facebook was so powerful the regulators could never make them stop at any turn. Now that they finally got sued big time, we finally educated ourselves enough as constituents to raise enough of a stink to trigger straight up bans. (educated ourselves, or politicians legislate based how bad headlines are, or it was so egregious it genuinely ticked them off… …)

    • ilovecake1984 30 minutes ago
      That’s not really a loophole though. We have child actors in Harry Potter.
    • everdrive 1 hour ago
      >Underage influencers

      Anyone who has hone so far as to become an influencer is already a lost cause. No law could save them.

    • logicchains 1 hour ago
      >If you are going to implement age controls, you should implement a ban on underage influencers as well.

      That just makes it even worse, why deprive the younger generation of one of the few remaining methods they have to make a decent income? We should be encouraging youth entrepreneurship, not making them spend even longer in classrooms learning things that LLMs will do better than them.

      • jrajav 1 hour ago
        This is almost verbatim the same argument that people make in support of allowing child labor in factories.

        Children do not need, nor are they entitled to, any kind of "freedom" to work for a living.

      • retired 1 hour ago
        People under the age of 16 shouldn't be worried about "making a decent income". They should focus on school.

        In the weekends they can stock shelves, deliver pizza, deliver newspapers, wash dishes, babysitting, feed animals or other typical jobs for children in the age range of 12 to 16.

        • hackinthebochs 28 minutes ago
          >They should focus on school.

          Why? Presumably so they can go to college and get a high paying job that may not exist in 10 years? The direction we give kids coming up always seems to lag behind reality by 10 or 20 years. Perhaps we shouldn't stand in the way of the new generation figuring things out for themselves in this brave new world. The old playbooks to a solid middle class life are increasingly outdated.

      • connoronthejob 1 hour ago
        Since when did being an influencer become 'one of the few remaining methods' to make a decent income?
        • bluefirebrand 9 minutes ago
          I don't think it truly is, but I do think that the younger generations think it is.

          My nieces and nephews really don't know what they are going to do in their futures because so much is uncertain right now.

          If it feels like a longshot to expect normal 9-5 office jobs to be around in 5 years, and it's also a longshot being an influencer, then why not go for the influencer thing?

  • ericmay 49 minutes ago
    Just requiring it for social media companies is probably enough of a win to not have to pursue any further. We require age verification for sports betting and things like that, I'm not sure why we wouldn't do the same or some variation of that for other massively addicting products that we know as a matter of scientific study have a very bad impact on some number of kids.
  • motbus3 44 minutes ago
    It is not like a digital control for id verification could be used anyway to control a narrative in war times right?
  • dirtikiti 1 hour ago
    And the piece nobody is even considering...

    Responsible parents don't have separate OS accounts for their children.

  • ilovecake1984 32 minutes ago
    I’d wager most people want more censorship of the internet.
  • midtake 34 minutes ago
    I agree, doxxing yourself to some shady gray-market adjacent data broker is not acceptable as age verification, and age verification was safer using the honor system as before. But for some communities, especially social media communities, some kind of verification is better than none, otherwise what's to stop them from being overwhelmed with alt accounts that are used simply for harassment or other targeted objectives?

    People should not be able to misrepresent themselves on the internet, it may have been safe in low volumes but it is scary now and will be outright dangerous as a modality in the hands of AI agents. If you think teen mental health is bad now, wait until social media campaign capabilities previously only available to nation states fall into the hands of ordinary school bullies.

    Maybe age verification isn't the way to mitigate this obvious risk, but there has to be something that can be done to stop rampant sockpuppeting.

  • aalaee 19 minutes ago
    For a forum that supposedly consists of hackers and tech-savvy people, this number of comments supporting age verification is concerning.

    The author has said a lot about what kind of future awaits with mass surveillance and AI, but I believe it’s not enough. Technofascism Is not that far away.

  • jrexilius 1 hour ago
    I can't agree with this enough and yet I think the long term danger is masked by the current problems for the majority of voters. I'm not hopeful.
  • stared 1 hour ago
    Online age verification is an example of the Motte-and-bailey fallacy (https://en.wikipedia.org/wiki/Motte-and-bailey_fallacy, https://slatestarcodex.com/2014/11/03/all-in-all-another-bri...).

    It is easy to defend on the motte hill (protection of children, protection against abuse and heinous crimes), and easy to expand and farm on the bailey (universal surveillance, mass data collection, and the erosion of privacy).

  • kaboomshebang 43 minutes ago
    Kids will always find ways around regulation. Look at cigarettes, vapes, alcohol, weed; they will just get it from their dealers. Pornography? I expect something like: download a Torrent, get it from a classmate, share HardDrives in school, get it through an older brother.
    • ilovecake1984 27 minutes ago
      It’s just defence in depth and wholly appropriate for it to be imperfect
    • IAmBroom 38 minutes ago
      And bootleggers will always bootleg, and smugglers will always smuggle. For that matter, murderers will find ways to murder.

      Shall we just abolish all laws? None of them have any effect whatsoever, if they are even slightly imperfect... by your rule.

  • giantg2 1 hour ago
    So many pieces of law are flawed today, and the reason why should be concerning to all.

    I find it disgusting that most laws today are based on creating a perfect world instead of addressing harms in the least intrusive way. There is no balancing of interests, even when they state that there are. Every side complains about the others and potential future abuses, except when it is their plan. Nobody tries to design the law with a devil's advocate perspective to make as effective as reasonably possible (not perfect!) while limiting overreach.

    The real problem is the pursuit of perfection. A perfect world does not exist, nor will it ever (laws of nature, physics, etc). One person's view of perfect is not the same as another's. We've lost the capacity for legislative empathy through are impatience and self importance. It's no longer about restricting government and providing people with rights. It's about how we can use government to shove the desires of a majority or plurality onto the total population.

    There are ways to do age verification with reasonable anonymity, but they aren't perfect and can create underground markets (see gaming in China). At a certain point, we need to step back and put the responsibilities where they belong - with parents, instead of causing massive negative externalities on everyone else.

    Yeah, yeah, but the children...

  • seydor 50 minutes ago
    Usually Fear is the realm of governments. Modern republics are basically legitimized around the fears of something terrible happening, it can be communism, narcotics, the ozone hole, corona virus, terrorists, immigration, globalization, unrecycled waste or greenhouse effect.

    Private entities being frontrunners in AI Fear either means that these companies have too much unchecked power or that they have are covert instruments of governments.

  • Havoc 57 minutes ago
    I have a fair bit of fatalism on this one.

    Saw it with the UK laws. It just gets rammed through. Whether it’s ignorance, malice, hidden force, a desire for surveillance state, genuine concern for children - doesn’t matter, the forces in favour are substantially more and seemingly motivated to try over and over until it sticks.

    Much like brexit or for that matter trump reelection I just don’t have much faith in wisdom of the democratic collective consensus anymore and I don’t think it’ll get any better in an AI misinformation echo chamber world. Onwards into dystopia

    Exceeding gloomy take I know

    • baggachipz 18 minutes ago
      Contacting my representatives is about as effective as making a silent wish. Whenever I've done it, I'll either get no response, or a boilerplate reply which basically says "I'm doing this, go fuck yourself". Then I'll be added to their spam list. The truth is that my reps don't represent me and they're going to do what they want regardless. After all, I'm not the one backing the truck of money up to their front door.
  • Ritewut 1 hour ago
    Just a reminder that the YC funds many of the companies pushing these laws and building the surveillance state.
  • yawniek 1 hour ago
    ironically i think we need more social and stronger local social networks that have high identity validation and are "safe" spaces for the plebs. so that the perceived "threat level" from the free internet gets lower. basically hide the real internet a bit behind a small rock. its a slippery slope but it might be the better strategy unless some democratic societies achieve to put more modern "freedom guarantees" into their consitution.
  • baxtr 1 hour ago
    Ok, maybe that’s a silly thought, but… couldn’t this be provided by Apple/Google anonymously?

    When you set up kids devices in your family they ask you to provide the birthday anyway.

    I’m keen to see the arguments against this.

    • add-sub-mul-div 57 minutes ago
      Further empowering and depending on either of those companies as a middleman in our lives should make us nauseous.
  • semiquaver 2 hours ago
    Why is it always “think of the children” used to abrogate the rights of adults?
    • subscribed 1 hour ago
      Because it's very easy for the creeps already thinking of your children to paint these rejecting this type of the laws as those who want to see children hurt.

      Regardless how stupid this argument is, rags will always pounce on it.

      This is just a dirty trick of the creeps to make the resistance harder.

    • scythmic_waves 1 hour ago
      I think it's because, without further context, it's so hard to argue against. Pretty much every person in every culture cares deeply about their children. So if you can successfully hitch your position to that idea, it too becomes hard to argue against.

      It's the same with tough on crime. "What, you want criminals to keep getting away with it?!"

    • LaGrange 1 hour ago
      Because adults remain children. As in, their parent’s kids and therefore property. [edit: I should mention also property of the state beyond that] It’s less explicit in US I guess but in some places that’s very blunt - if you don’t support your parents enough you can be sued for abuse. And there are situations where an adult in us has been declared too irresponsible and forced into conversion camps by parents in the US. It’s insane, yes, and if you’re lucky enough this might be entirely invisible to you. But if you’re gray or trans or autistic and get a but unlucky this can become a very harsh reality.

      Protect the children refers to a type of property, not a type of human.

  • cft 2 hours ago
    There is a sudden concerted international push for online age verification, and we do not know where this push originates from. That is the scariest thing about it.
    • jrajav 1 hour ago
      It's not _completely_ shrouded in mystery - it started after Facebook got slapped by the EU for irresponsible handling of underage users, and since began a heavily funded lobbying push to drag competitors down with them. https://github.com/upper-up/meta-lobbying-and-other-findings...

      Of course, it's probably also been coopted by the neverending stream of nanny-state political power grabs in both the US and EU.

      • cft 49 minutes ago
        [flagged]
  • anonym29 2 hours ago
    It's worth pointing out that full digital identity verification ("doxxing" yourself to an untrustworthy, unauditable, legally unconstrained private company) is NOT the only way to verify adulthood. We have had a system in place which enables adulthood validation without enabling digital surveillance infrastructure, with a degree of false negative risk that society has deemed acceptable for nearly 100 years now. This idea is not my own, but I'm happy to share a reasonable proposal for it.

    The Cashier Standard – Age Verification Without Surveillance

    https://news.ycombinator.com/item?id=47809795

    https://claude.ai/public/artifacts/7fe74381-a683-4f49-9c2b-1...

    • jaykru 2 hours ago
      The "cashier standard" you advocate for has already crept toward centralized state tracking in places like Utah. When you go to a restaurant and order a drink, the staff are required to take it to the back and scan it for verification. The scanned data is also compared with a state database of DUI offenders. It's not clear whether the database is stored on site, or if that data goes out on the wire for the check; presumably the latter. Scanned data is also stored for up to 7 days by the restaurant, and it's easy to imagine further creep upping that storage bound.
      • anonym29 2 hours ago
        This is not the case in most of the country. Utah is largely influenced by a Mormon / LDS culture that expresses heavy opposition to drinking. I am clearly not proposing that the cards be scanned Utah style, I am proposing that they be glanced at by a cashier, everywhere else style.
        • rationalist 1 hour ago
          More and more places I go in other states besides Utah, try to scan IDs when purchasing alcohol.
          • anonym29 1 hour ago
            Again, the proposal isn't for a system which requires scanning of IDs, it's for a system where the cashier glances at the ID. You're arguing against a strawman. You may argue that the system proposed could evolve into the system you're describing, but still, you're arguing against a hypothetical future fiction. If we're going to be arguing about what the proposal might evolve into in the future, we might as well be arguing about what we should be doing when aliens arrive, since they might arrive in the future, too.
            • Supermancho 1 hour ago
              > we might as well be arguing about what we should be doing when aliens arrive, since they might arrive in the future, too.

              Did aliens land in multiple states already? Strawman deflections aside, scanning is the natural evolution and has already happened across multiple kinds of exchange (money markers, various ids, various phone apps, etc). Government issue has a benefit of an independent verification system. It's super expensive for various government agencies to integrate into businesses. Constituents and businesses don't want that, leading to a much more comfortable adversarial relationship, imo.

    • _ink_ 2 hours ago
      How does this prevent a second market for one time codes? I as an adult can just get a code and sell it someone else.
      • HWR_14 18 minutes ago
        Stings that catch adults reselling codes.

        It doesn't have to be perfect.

      • anonym29 1 hour ago
        It doesn't prevent it, it just disincentivizes it. As an adult, you can also go buy a beer and sell it to a minor. That said, mandatory age verification with photo ID upload and facial scans doesn't prevent workarounds either - kids use their parents' photo ID and pass facial scans with a variety of techniques, too.

        Nobody who understands how adversarial systems like this work is seriously expecting a 100% flawless performance of blocking every single minor and accepting every single adult, the question is how much risk is acceptable, and the risks posed by this system are acceptable for alcohol, cigarettes, and other adult items that can arguably pose much more acute risk of serious injury or bodily harm to kids.

    • hypeatei 2 hours ago
      This type of system is a horrible idea for the following reasons:

      1) the cards can just be re-sold which creates a black market and defeats the "cashier physically saw the person buying the card" angle

      2) nickle and dimes people for simply browsing the internet (verification can dystopia anyone?)

      3) related to #2, it creates winners in the private sector since presumably you need central authorities handing out these codes

      I abhor the idea of digital ID verification, but if we're going to do it, let's not create a web of new problems while we're at it.

      • arowthway 1 hour ago
        Is it even theoretically possible to have bearer anonymity and no reselling option at the same time?
        • terangaway 1 hour ago
          With digital tokens being generated by a user (the seller) on demand, you could have a bond system where the seller places something costly on the line, that the buyer can choose to destroy or obtain. For instance, if Alice gives her age token to Bob, Bob can (if he is a troll) invalidate the token in a way that requires Alice to go to a physical location to reset her ID.

          I imagine this could be done with appropriate zero-knowledge measures so that the combination of Alice's age token and Bob's private key creates a capability to exercise the option, but without the service (e.g. a social media site) knowing that the token belongs to Alice, and without the ID provider (e.g. the state) knowing that Bob was the one who exercised it.

          While honest customers have no reason to make use of this option, if Alice blindly sells her tokens to anybody willing to pay, there's bound to be some trolls out there who will do it just for the laughs.

          This is far from a perfect system since a dishonest site could also make use of the option. But it theoretically works without revealing anybody's identity (unless the option is used, and then only if the service and the ID provider collude).

      • anonym29 1 hour ago
        First - Alcohol and cigarettes can just be resold too. The black market for them is effectively zero because the consequences for giving them to kids are severe and the room for meaningful profit is close to zero, same applies here.

        Second - The codes would be priced on the order of magnitude of pennies per verification - think 10 cents or less, accessible even to low / fixed income folks without really making a dent in their budget.

        Third - the proposal explicitly mentions a nonprofit running it as an option, and the idea would be that law codifies the method to be approved, not a specific vendor, so competitive markets could emerge, too. Would you argue that restrictions on the sale of alcohol are creating artificial winners in the private sector of alcohol manufacturing?

        • arowthway 1 hour ago
          'consequences for giving them to kids are severe and the room for meaningful profit is close to zero, same applies here.'

          I don't think it applies, the difference is that codes are digital and can be sold over the internet, anonymously, in a scallable manner.

          I still like this solution because all the solutions I've seen have flaws, this one being so easy to explain makes it great to campaign for.

        • hypeatei 1 hour ago
          You're doing a huge logical jump in your first point. Alcohol and cigarettes are physical goods, digital ID is not, but you're proposing a system that turns it into a physical problem. I'm merely pointing out that's what you're doing and the issues with it.

          Second, it doesn't matter what it costs, it's inconvenient and I already spent time (possibly money too) obtaining a government ID... on top of a theoretical mandate that says I need to show the ID on a bunch of websites.

          Third, I'm not sure I follow your point on alcohol restrictions creating winners? The non-profit idea could potentially be good, but I'm not hopeful that real world legislation would be crafted that way.

          EDIT: also more on #1 and "severe consequences" for re-selling... yes that's exactly what we want to avoid: creating more reasons to put people in prison and a bigger burden on law enforcement and the court system.

  • SirMaster 1 hour ago
    "But age verification requires identity verification. Identity verification requires digital IDs."

    Um, no? iOS is doing age verification just by your credit card. I never saw people all that upset about giving their credit card info to their phone wallet app or even to a bunch of websites.

    • Finnucane 1 hour ago
      Are you going to give your cc number to every website in the world? Also, is that really an ID?
  • callamdelaney 1 hour ago
    Agree
  • KaiserPro 1 hour ago
    If it was the hill to die on, then we should have done a better job of stopping pervasive fraud, abuse and harm to everyone so that we wouldn't have been a need to bring in age verification.

    The reason we are up shit creek is because large companies didn't want to spend 2-5% of profits on decent editorial controls to stop bad actors making money from bending societal red lines (ie pile ons, snuff videos, the spectrum of grift, culture of abusing the "other side")

    They also didn't want to stop the "viral" factor that allows their networks to grow so fucking fast.

    This isn't really about freedom of speech, its about large media companies not wanting to take responsibility for their own shit.

    meta desperately want kids to sign up. There are no penalties for them pushing shit on them. If an FCC registered corp had done half the shit facebook did, they'd have been kicked off air and restructured.

    So frankly its too fucking late. Meta, google and tiktok will still find ways to push low quality rage bate to all of us, and divide us all for advertising revenue.

  • shevy-java 14 minutes ago
    I agree. I don't call it "age verification" though - it is age sniffing. And it has nothing to do with children - that is the lie.

    What is fascinating is to see how governments ALL fall for it. There is zero resistance. This is fascinating to me. It shows how little real effort is necessary once you have the lobbyists in place. Kind of scary to witness too.

    It is an apartheid system. All apartheid slavery systems will eventually die, so age sniffing will die too. But it will most likely be a long fight as more and more money will be invested by crazy corporations such as Palantir and others.

    The whole "debate" is already not logical by the way. Let's for a moment assume the "but but but the kids!" is a real argument rather than a strawman argument, which it is. Ok so ... I am a "concerned parent", for the sake of discussion. I have three young kids. I am not a tech nerd. The kids see "unfitting content" on the antisocial media such as facebook and what not. So, what do I do? Well ... they have a smartphone? Aha, so ... I am not so concerned? Having no smartphone is no option? Ok so ... I say they can have a smartphone, but they may not use antisocial media. Ok. First - in any free society, is it acceptable that this kind of censorship is done on ALL kids? What if I, as a parent, do not agree with this? Well, tough luck - the laws force you into the age sniffing routine suddenly. But, even those parents who want the state to act as totalitarian: why would I want to hand over control to ANY politician for that matter? That makes no sense to me. I am aware that some parents may think differently, but do all parents think like that, even IF they buy into the "we protect the children" lie? I don't want ANY information from ANY of my computers to go into private hands here. So the whole argument already makes zero sense from the get go.

    Of course those who know how things work, they know that this is the build up towards identifying everyone on the world wide web at all times AND to make access to information conditional, e. g. if the state does not know you, you can not access information. Aka a passport system for the www. Built right into the operating system too. Windows already complied. MacOSX too. The battle for Linux will be interesting; it may be some hybrid situation, like systemd. And the systemd distributions will all succumb to age sniffing, courtesy of Poettering "this is really harmless if we store your age in the database, just trust me".

  • fithisux 1 hour ago
    We now know all the arguments. No more need to persuade anyone.

    People will show what they are made of.

  • selectively 2 hours ago
    An attestation-like system to detect humanity at time of post is absolutely for useful online spaces in the era of AI slop.

    The writing style of the author is very annoying.

    • HWR_14 16 minutes ago
      Until people hit "attest" and then copy the text from ChatGPT.
    • MiddleEndian 1 hour ago
      And people should be free to pick and choose whether they want to use sites that do that or not. Whatever hacker news does seems to be fine for me, and I did not need to verify my ID in any way (even though it's very easy to figure out who I am from this profile)
    • goda90 1 hour ago
      It could be done with anonymous credentials though. No tracing to who the human is.
      • selectively 1 hour ago
        Anonymous in terms of it not being possible to derive the real world identity of the human from the value, sure. Anonymous in terms of providing no durable way to ban that human from the platform? No.
  • inquirerGeneral 1 hour ago
    [dead]
  • zzzeek 2 hours ago
    [flagged]
    • rationalist 1 hour ago
      Please don't play identity politics, it's a fallacy and further perpetuates the problems that we are all facing.
      • zzzeek 1 hour ago
        "prepper" is not an identity. I looked up this guy, he has crazy takes he's free to change at any time (so not an "identity" in any way) and id rather not read him.
        • rationalist 1 hour ago
          "right wing" is considered politics that people identify others in.

          > free to change at any time (so not an "identity" in any way)

          I'm free to change my name at any time, is my name not considered a part of my identity?

          • zzzeek 1 hour ago
            Right wing is not an identity either. Don't mess around with terms you don't understand. Do some research.

            > I'm free to change my name at any time, is my name not considered a part of my identity?

            You can change your name to conceal elements of your identity just like you could wear clothes that cover your skin color. So sure, you can hide elements of your identity but changing your name won't change your family history or ethnicity.

            Growing up in a Republican family might be an identity but producing lots of blogs, videos, and commercial products within the right wing prepper lifestyle is not an identity. Just do a little research thanks

            • rationalist 1 hour ago
              > Do some research

              I have... (???)

              > Just do a little research thanks

              Just have an open mind thanks

  • kelseyfrog 2 hours ago
    [flagged]
    • bee_rider 1 hour ago
      Basically every article on this site has a comment complaining that the article is AI. Who knows. Maybe “complaining about AI” is the new AI way of fitting in.
      • acheron 1 hour ago
        I just flag all the AI complaints. Perfect example of the guideline about “don’t complain about tangential issues” or whatever the wording is.
        • mplanchard 1 hour ago
          This feels different to me than complaining about the font or whatever. I don’t want to read or comment on anything not written by a human. I also agree with GP here that using AI instead of your own words has bearing on the content itself, insofar as it’s a signal that the author doesn’t care enough to write it themself.

          As a corollary, I also want to know if a project posted here is predominantly vibe-coded, since that to me is a signal that it may be of lower quality, have fewer edge cases worked out, and is more likely to be abandoned in the near future.

        • arduanika 52 minutes ago
          Caring enough to put in the effort of thinking and writing is not a "tangential issue". Laziness is a substantive defect, and sadly, I think that kelseyfrog has clocked this one correctly. There are borderline cases, but the cadence of this tweet thread is unmistakeable.

          We don't have to live like this. We don't have to accept it. We don't have to upvote it even if we agree (as I do) with the explicit point. The medium is the message, and the message that this poster is putting out here is that online age verification isn't actually worth getting that worked up about.

        • bakugo 53 minutes ago
          AI-generated content being passed off as human-written is not a tangential issue. HN staff agree, because posting AI generated comments is explicitly forbidden. I suspect the only reason this isn't extended to submissions is because pretty much all articles about AI are also written by AI, and effectively forbidding positive discussion of AI is obviously against the interests of a VC firm.

          HN's guidelines were written under the assumption that submitted articles about [thing] would be written by people who care about [thing] and made a good faith effort to write something interesting about [thing], so it's only fair that any comments would be expected to respect the author's effort and discuss the article in equally good faith.

          This assumption completely falls apart when you add AI generated submissions into the mix. If the "author" didn't care and thus couldn't be bothered to write about [thing] themselves, choosing to instead outsource that work to an LLM while they supposedly did something they deemed more valuable with the time they would've spent writing, then why should commenters be expected to dedicate more effort into their discussion of the article than the author dedicated to writing it? It's a bit unfair towards the commenters, don't you think?

      • bakugo 1 hour ago
        > Basically every article on this site has a comment complaining that the article is AI

        Just a hunch, but it may have something to with the fact that basically every article on this site is AI.

      • noident 1 hour ago
        No, it's because authentic writing on HN has been drowned out in an ocean of slop, in such quantities that calling it out is becoming an exercise in futility
        • mwigdahl 28 minutes ago
          There was an article yesterday where people were complaining about "AI smell", the author showed up in comments to state clearly and unequivocally that he didn't use AI to write it, and people wouldn't believe him.

          AI didn't get its tropes and tics ex nihilo. Some people just write in a way that "smells" like AI to others.

      • devmor 1 hour ago
        If everyone is complaining about the smell of shit, maybe it's because there's shit everywhere.
        • stronglikedan 1 hour ago
          It's more likely that they're just virtue signaling about {{current-controversial-thing}}, as evidenced by the fact that they often accuse content of being AI generated when it would only appear that way to the most naive readers.
          • kelseyfrog 1 hour ago
            It doesn't feel like virtue signaling. It feels like pointing out a contradiction in the text: I care deeply about this topic; I don't care enough to write it myself.

            > Virtue signalling is a pejorative neologism for the expression of a moral viewpoint with the intent of communicating good character, frequently used to suggest hypocrisy.

            What virtue am I signalling and what hypocrisy am I trying to hide?

    • WarmWash 1 hour ago
      You should add an AI clause to that license agreement in your profile.
      • kelseyfrog 59 minutes ago
        How would it read?
        • arduanika 30 minutes ago
          Holy crap, I only just saw your license agreement. Oh no. We've argued on here before, although this time we're in agreement. Please don't use this hidden license to dox me!

          (It's an unenforceable joke, right? There's no way I'm bound by anything here other than maybe the site ToS.)

          • kelseyfrog 12 minutes ago
            Lol, who knows. I don't give legal advice. Wanna find out? ;)
            • arduanika 7 minutes ago
              Absolutely not!

              > How would it read?

              Sounds like you don't really care for this idea, so maybe just have Claude write it for you.

  • streetfighter64 1 hour ago
    Seriously, who cares this much about the internet? I for one will be happy if my kids spend less time online than me. Similar to what a smoker would feel seeing cigarettes finally be banned, I suppose.

    It's also ironic that this guy is so adamant about protecting the children on xitter. It's like preaching against racism on 4chan.

    • rationalist 1 hour ago
      > who cares this much about the internet?

      The Internet pretty much runs our lives now, so: I do.

      Lots of things require having Internet access, an email address, being able to visit a website, coordinate with others on a Facebook page for a local group, etc.

      No one requires me to buy a pack of cigarettes to register for classes, pay bills, submit something to the government, etc.

  • stackedinserter 1 hour ago
    Very unpopular opinion here on HN: one can't stop it without direct physical action against those who push it.
  • eykanal 2 hours ago
    Alternative take: The fact that twitter / facebook / whatever allow arbitrary, unverified posting enables large-scale misinformation that led to, among other things, Russia's manipulation the US electorate and ultimate impacting the presidential election.

    This one-sided view has some good points, but for goodness sake, don't pretend that the alternative has no downsides.

    • CamperBob2 2 hours ago
      You'll need to explain how age verification fixes that.
    • nradov 2 hours ago
      Really? How many Electoral College votes did Russia's clumsy attempt at manipulation actually change? Please quantify that for us based on hard evidence.
    • Larrikin 2 hours ago
      Playing devil's advocate outside of debate club only serves to promote the devil's point of view.

      State your well reasoned opinion where you have considered the facts. Or just say you are in support of this openly.

      • bit-anarchist 1 hour ago
        Disagreed. I'm against invasive age verification methods, but to allow innacurate expectations to proliferate often becomes a bubble that pops, causing many to rebound to the other side, even if it's objectively worse. I much prefer to keep the tradeoffs clear, as it prevent betrayed expectations while still showcasing the unnacceptible downsides.
        • Larrikin 1 hour ago
          I'm firmly against the idea of Internet arguments presenting an opposing position under the guise of it not being their actual opinion so they can run away from debate. Devil's advocate is a technique that should be used in school to learn how to make stronger arguments.

          All it does is covertly promote the idea by presenting it as reasonable and on an equal level to the other idea. While at the same time being able to shut down debate, by pretending they don't actually think that.

          Anybody can say something like "but what about the good side of the African slave trade" but they will be debated and the argument shut down if they present it as their actual argument and engage in good faith with the comments. Using the devil's advocate technique is an extremely useful way to argue in bad faith, anonymously on the Internet.

          Critique of the author's style is fine. An opposing view should honestly be presented as such.

  • speak_plainly 2 hours ago
    The argument being made seems plausible but it’s complete fear mongering. The surveillance mechanisms already exist and are in play and people can be identified in endless ways.

    States have broad power to do what is being feared in the thread and haven’t already and to think that they’re waiting for this final piece of the puzzle to enact some insane regime is laughable. They could do that right now without the internet at all.

    Social media is probably not healthy and kids should probably not be on social media. Age verification and age limits for social media will be a good thing for kids.

    Instead of fear mongering, finding a middle ground, like governments adding some rules and protections on how this information or system is used is probably a better response.

    I might be in the minority, but I think incorporating an identity layer into the internet itself should happen with the right protections for users and should have happened at the beginning of the net and is probably a result of lack of foresight by the creators of ARPANET.

    • rationalist 2 hours ago
      What I'm hearing you say:

      > Our freedom is already being eroded, saying that it is being eroded more is just fear mongering.

      > They want to hurt you, instead of fear mongering, find a middle ground where they're hurting you differently.

    • selectively 2 hours ago
      Social Media is not a thing at all. Social media is a website. Websites are not health or unhealthy. Food is healthy or unhealthy. Websites are light and potentially sound, not something with health effects.
      • Kbelicius 40 minutes ago
        Go look directly at the sun without any protection or go listen to sounds of 120dB if you want to test your hypothesis that light and sound can't be unhealthy.

        Or maybe you aren't being litteral and are just saying that what children see and hear has no influence on their developmemt. Either way, total bullshit.

      • crdrost 1 hour ago
        This is simply false -- the literature is full of discussion about the health effects of social media.

        More generally you're committing I believe two separate fallacies of ambiguity? Like one in going from the institution of social media to its reification in the form of specific websites, and then a second fallacy when you go from the specific websites to all websites in general? Like if you said "Gun ownership is not a thing at all. Gun ownership is a piece of metal. Pieces of metal cannot be healthy or unhealthy." OK but, you owning a gun is known in the scientific literature to significantly correlated with a bunch of very adverse health effects for you, such as you dying by suicide or you dying from spousal violence or your protracted grief and wasting away because your child accidentally killed themselves. Like to say that it's impossible for the institution to have adverse health effects because we can situate the objects of that institution into a broader category which doesn't sound so harmful, is frankly messed up.

        [1]: Bernadette & Headley-Johnson, "The Impact of Social Media on Health Behaviors, a Systematic Review" (2025) https://pmc.ncbi.nlm.nih.gov/articles/PMC12608964/ - the content you consume can promote healthy or unhealthy behaviors

        [2]: Lledo & Alvarez-Galvez, "Prevalence of Health Misinformation on Social Media: Systematic Review" (2021) https://www.jmir.org/2021/1/E17187/ is notable not just for its content but also like a thousand papers that cite it getting into all of the weeds of health influencers sharing misinformation to make a buck

        [3]: Sun & Chao, "Exploring the influence of excessive social media use on academic performance through media multitasking and attention problems" (2024) https://link.springer.com/article/10.1007/s10639-024-12811-y was a study of a reasonably large cohort showing correlations between social media usage and particular forms of multitasking that inhibit academic performance -- more generally there's broad anecdata that the current "endless scrolling constant dopamine hits" model that social media gravitates to, produces kids that are "out of control" with aggressive and attentional difficulties -- see Kazmi et al. "Effects of Excessive Social Media Use on Neurotransmitter Levels and Mental Health" (2025) (PDF warning - https://www.researchgate.net/profile/Sharique-Ahmad-2/public...) for more on the actual literature that has probed those questions

        [4]: The APA has a whole "Health advisory on social media use in adolesence" https://www.apa.org/topics/social-media-internet/health-advi... which is pretty even-handed about "these parts of social media are acceptable, those parts can maybe even be downright good -- but here are the papers that say that for adolescents, it can mess with their sleep, it can expose them to cyberhate content that measurably promotes anxiety and depression, it has been measured to promote disordered eating if they use it for social comparison..."

        • selectively 1 hour ago
          You posted a giant, AI generated block of junk science.
  • cvoss 2 hours ago
    > If you love your family, you must stop online age verification.

    > If you want the best for your children, you must stop online age verification.

    > Your children are being targeted. The infrastructure being built under the cover of child safety is designed to enslave them for the rest of their lives.

    Jumped the shark on that one, and really off-color. I'm less inclined to listen to guy, not because of his actual points, but because of how unreasonable he sounds when articulating them. A great lesson in how not to do rhetoric.

    • emptybits 2 hours ago
      When I read those seemingly outrageous claims, I didn't immediately dismiss the author. I allowed him to substantiate the claims and kept reading. I found myself agreeing with his argument and his train of thought of how, once digital IDs are accepted as a norm, they won't be unwound, and all online activity will likely require them and then, as he says,

      "Your children will never know what it was like to think freely online. They will never explore ideas anonymously. They will never question authority without it being logged in their permanent profile. They will never speak freely without fear that every word will be used against them.

      They will grow up in a digital cage. And you will have to tell them you saw it being built and did not stop it when you had the chance."

      So I'm with the author on this one. Under the cover of child safety, digital IDs will cage us (or at least children entering the verification age), and it will probably never be rolled back.

      • paisawalla 2 hours ago
        That's the role of rhetoric as a skill: all the true and sufficient syllogisms in the world will be ignored by most readers, if the argument leads with priors-triggering hyperbole and bombast.
      • Ifkaluva 2 hours ago
        The best way to not be in a digital cage is to opt out of the current digital products.

        Would that be such a bad thing? Frankly I would welcome a world in which kids are not using Instagram or TikTok. They don’t have to live in a cage if we don’t let them in the cage.

        Personally, my plan is that when age verification laws get passed, every service that requires ID is a service I stop using. And I expect my life to be better for it!

        • noah_buddy 1 hour ago
          What if all services require ID?

          Let’s take a basic example: Wikipedia, which hosts pornography, easily could be a target of such legislation. Now there is infrastructure in place to know when you read about “Criticisms of policy X” and maybe it’s handled safely or maybe it’s handed directly to the government.

          What about news? It’s a hop skip and leap from “age verify pornography with ID” to “age verify content about sexual abuse or violence.” Now the infrastructure is in place to see the alt-news criticisms you read.

          Twitch or YouTube wouldn’t even wait to comply, ID verification is something that these corporations are already perfectly fine with. Now, you watching a history of your government’s crimes is a potentially tracked red flag that you’re a dissident to be watched.

          Do you think if this sort of legislation is enacted, it will stop at large websites? It will be an excuse used by the government and supported by big tech firms to shut down any small websites which don’t comply. After all, Google, MS, et al, they would rather that your entire concept of the internet start and end in a service they control.

        • matheusmoreira 1 hour ago
          > The best way to not be in a digital cage is to opt out of the current digital products.

          But will your friends and family opt out? Their phones are always listening. They can just as easily listen to you, even if you go to great pains not to expose yourself to technology. They'll make a shadow profile of any avoidant user whether they want it or not.

        • pessimizer 46 minutes ago
          > The best way to not be in a digital cage is to opt out of the current digital products.

          Bullshit. These are all-encompassing monopolies and government services. More likely, they'll ban you and you'll end up having to go to court out of desperation to demand that they service you.

          This is very limited thinking. If you lacked this sort of imagination 20 years ago, you wouldn't have been able to predict today.

          > Frankly I would welcome a world in which kids are not using Instagram or TikTok.

          This is the sort of passive reactionary nonsense that causes the danger that we're in. Everything isn't something to give up lightly, even if you think that it will force your neighbor to turn his music down, or get rid of bad reality television. I don't like kids on social media either. I don't like adults on it. I think kids are suffering more from surveillance than from TikTok.

      • acheron 1 hour ago
        Nah that’s silly, because Google has been doing all that already for the past quarter century. This “age verification” shit isn’t going to move the needle on the Google-created dystopia we already have.

        The time to worry about not having a digital cage was quite awhile ago. Instead tech people pushed Chrome and Android and Gmail and ads onto us.

        • mcdow 1 hour ago
          Chrome, Android, and Gmail are optional to use.
          • vrganj 1 hour ago
            So is social media.
            • afpx 1 hour ago
              It's framed as being only for social media. But, really, it's about network access. Without network access, it's difficult to thrive in the modern world.

              Are you not alarmed at the possibility that a person's network access could be cut arbitrarily and at-will?

        • Barbing 1 hour ago
          Is Google tracking which teenagers make which posts on 4chan?

          Curious about via Google Chrome versus not

    • simplyluke 51 minutes ago
      Is it? Digital ID is the point being made here by the x thread. It's being brought in under age verification. It's arguing that the laws being passed, and infrastructure to enforce them, to protect your kids today will be used to abuse them in the decades to come.

      It's trivially easy for me to imagine covid playing out +10 years from now instead of -5, and the ~3 established identity/age verification players in that market at that point responding to exactly the same pressure that was applied to the handful of social media companies to censor people disagreeing with the administration's approach. Real people were fired from real jobs for exactly this in working memory, they did it under their real identities on social media sites. The future being driven towards now would ensure that there's no anonymous forums to avoid that risk on controversial subjects in the future.

      You disagreed with the logic of mask mandates, or suggested that a lab leak was a plausible theory, that was tied to your identity, and now it's not a ban on instagram for a while -- you can't use AI tools you need for your job, your email, your online banking. You are functionally unemployable and excommunicated from all internet tools that rely on those handful of services. The smartest legal minds argue that this wasn't actually a violation of your first amendment rights, after all, those three private businesses just decided that they didn't want to do business with someone engaging in dangerous rhetoric, and the fact that the administration sent them an email letting them know isn't material to that (Murthy V Missouri, 2024).

      We're already living in the early innings of political controversies coming from the fact that the youngest politicians had social media accounts where they said dumb things as teenagers/young adults. Is the future where the dumb ideas 16 year olds post on reddit are tied to their government IDs for eternity a good one?

      How is this rhetoric jumping the shark?

    • jasonjayr 52 minutes ago
      A lot of people dismissed RMS's "Right to Read"[1] essay long ago. All the things it was warning about have come to pass, in spades.

      1: https://www.gnu.org/philosophy/right-to-read.html

      • matheusmoreira 43 minutes ago
        It's mind boggling how far Stallman saw into the future. Saddest part is we're losing this war. They're going to destroy freedom of computation, freedom of information, and it turns out that... Nobody cares. Nobody but a bunch of nerds.
    • awkward 1 hour ago
      Responding to tone but not to content is what a dog does.
      • therobots927 1 hour ago
        looks like you ruffled some feathers with this one
        • Barbing 1 hour ago
          Tone was off
        • streetfighter64 1 hour ago
          Yeah, calling people "dogs" for pointing out that TFA is a hyperbolic (AI-written) screed without substance would ruffle some feathers.

          Edit: yes it is hyperbolic and ridiculous to suggest people will be "enslaved" because they don't have access to the internet. Do you realize that makes everybody who grew up in the 90s or earlier a "slave"?

          • matheusmoreira 1 hour ago
            Nothing "hyperbolic" about the points made. If anything it's not nearly extreme enough. People have no idea how bad things really are.
    • nandomrumber 2 hours ago
      Do you actually have an argument to make?

      He’s 100% correct.

      For a start, child are parents responsibility, and the state should stay out of that as much as reasonably possible.

      Nothing more would need to me said on the matter if that’s as far as it went, but it isn’t.

      There can be no free speech if the state can imprison you for what you say, and they know everything you say.

      I dropped the word ‘online’ from the above paragraph, because on is the real world. Touch grass, but there’s no way online isn’t real. Are these words not real simple because I telegraphed them to you?

      That’s not a world I want to live in.

      • raverbashing 1 hour ago
        > For a start, child are parents responsibility, and the state should stay out of that as much as reasonably possible.

        Yes

        That's why stores let kids buy alcohol and tobacco, of course, because no responsible parent would let them buy that, right?

        That's why any kid can go watch any movie in the cinema right?

        Yes it's the parents responsibilities. Do you think a middle class single mother has the resources to keep their kids entertained and out of social media for the whole day?

        The problem with age verification is 100% the lack of anonymity in its implementation (which I do agree has ulterior motives) - but honestly not the age check in itself

        • rationalist 1 hour ago
          > That's why any kid can go watch any movie in the cinema right?

          Yes. At least in the U.S., the federal government does not regulate that, it is voluntary by the MPA (formerly MPAA) and theaters. A kid can buy a ticket for a PG movie and walk into an R-rated movie.

          > Do you think a middle class single mother has the resources to keep their kids entertained and out of social media for the whole day?

          Mine did. While not everyone has a backyard, things like pencils, papers, books, used toys, etc can be found inexpensively or for free.

          • olelele 1 hour ago
            Did social media exist when you grew up?
            • rationalist 1 hour ago
              Xanga and MySpace are what my friends had; yes
        • hackable_sand 19 minutes ago
          It's weird that none of your arguments or proposals hold accountable the responsible parties.

          You want to force us to compromise when we were minding our own goddamn business.

    • bondarchuk 2 hours ago
      >They are counting on you caring more about sounding reasonable than protecting your kids from a system designed to control them forever.
    • peyton 1 hour ago
      The kids are our future adults. It should be pretty obvious that getting them used to the state yanking access is a future problem. I don’t see anything off-color or unreasonable.
    • jrm4 1 hour ago
      Maybe you're not the target, then.

      I haven't heard too many people say these extreme-sounding, yet at least arguably true points out loud.

      Someone should be saying them, and the fact that it's not your particular cup of tea may not be the biggest issue here.

    • pessimizer 54 minutes ago
      > how unreasonable he sounds

      It's important to remember that they're targeting your children. You grew up with freedom from surveillance and constant identification. You were able to communicate anonymously and without the content of your speech being sold to Walmart and the cops. They are putting in effort to make sure that your children will never have that reality as a reference point. The idea of the government and a dozen corporations not knowing everything that they are doing at all times, and not using and selling that information freely, will sound like the ramblings of a delusional old fool.

      It's important that you engage with that. Denial is not something to brag about.

    • therobots927 1 hour ago
      I’ve been noticing a trend among a lot of HN members where instead of contending with the arguments made in an article, they focus on the “off putting rhetoric” used by the author.

      Make no mistake you are engaging in your own form of rhetoric when you respond like this. You are in effect moving the discussion away from the subject at hand, and towards the perceived faults in the author’s communication style. This is a rhetorical slight of hand and it’s highly disingenuous.

      • jcheng 1 hour ago
        "Disingenuous?" Just because someone finds the style irksome, and chooses to share that here, they're deceptively, calculatingly trying to derail the conversation? That's an extremely cynical and uncharitable take.

        If I were the author of the post, I'd value the feedback.

        • JohnMakin 1 hour ago
          Except that is not what this place is for, at all, and flirts with several explicit posting guidelines. It doesn't make for good discussion, doesn't address the topic at hand, etc.
    • babypuncher 2 hours ago
      5 years ago I would have agreed, but seeing how the GOP has been fighting tooth and nail to protect actual child sex traffickers, I don't think so anymore. There's just no possible way that the safety of children is an actual concern to any of them. To these people, kids are little more than sex toys for billionaires.
    • streetfighter64 1 hour ago
      Ironic that he's relying on the same ridiculous "think of the children" rhetoric that's being used to promote age verification. Really says a thing or two about online discourse in our day and age.