• lmmarsano@group.lt
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    4 hours ago

    Cool: agreed. Your objection was ambiguous.

    If we had to choose, though, I’d consider the professor’s suggestion preferable to age verification. While I disagree with mandating it, it’d pretty much do nothing, because it’s already reality: most mainstream OSs include parental controls. The “criteria” would establish standards for parental controls, which isn’t altogether a bad idea. A better idea would be to promote a standard & replace mandates with public services to provide parental control technologies free & to educate parents.

    In the late 90s, when US Congress attempted to regulate access of adult content to minors, those laws commissioned studies that drew similar conclusions even then. The studies & federal courts concluded that to meet the government’s compelling interest in “protecting minors from harmful content”, there were more narrowly tailored alternatives to criminalization & age verification that are less restrictive to fundamental rights & are at least as effective:

    • client-side filters to block content from the receiving end
    • government programs to train parents & provide them resources to “protect” their children from “harmful content”
    • public education campaigns.

    They pointed out while client-side filters may have false positives & negatives

    • they can be monitored & corrected
    • they’re a more complete solution that can restrict all internet protocols (not just web) from any geographic source (not only in legal jurisdiction) with content of any type (including dynamic such as live chat)
    • they allow restriction of other kinds of content (eg, violence, hate speech)
    • they can vary restrictions per child (eg, age-appropriateness)
    • they let parents disable them
    • they don’t obstruct access by adults.

    Criminalizing access to adult content at the source obstructs everyone’s access & burdens them with loss of privacy & with security risk.

    Despite their age, those studies’ findings remain relevant.

    • COPA Commission

      In October 1998 Congress enacted the Child Online Protection Act and established the Commission on Online Child Protection to study methods to help reduce access by minors to certain sexually explicit material, defined in the statute as harmful to minors. Congress directed the Commission to evaluate the accessibility, cost, and effectiveness of protective technologies and methods, as well as their possible effects on privacy, First Amendment values and law enforcement. This report responds to the Congressional request.

    • National Research Council

      In November 1998, the U.S. Congress mandated a study by the National Research Council (NRC) to address pornography on the Internet (Box P.1).

    COPA Commission summary

    The COPA Commission found Age Verification ID to have the highest adverse impact on cost, privacy, fundamental rights, and law enforcement and to score poorly on effectiveness and accessibility. They found other technologies & methods to be more effective & accessible with much lower adverse impact including

    • client-side filtering
    • family education programs
    • acceptable use policies
    • top-level domains for materials “not harmful” to minors
    • “greenspaces” containing only child-appropriate materials.

    Some recommendations to highlight

    Public Education:

    • Government and the private sector should undertake a major education campaign to promote public awareness of technologies and methods available to protect children online.
    • Government and industry should effectively promote acceptable use policies.

    Consumer Empowerment Efforts:

    • Resources should be allocated for the independent evaluation of child protection technologies and to provide reports to the public about the capabilities of these technologies.
    • Industry should take steps to improve child protection mechanisms, and make them more accessible online.
    • A broad, national, private sector conversation should be encouraged on the development of next-generation systems for labeling, rating, and identifying content reflecting the convergence of old and new media.
    • Government should encourage the use of technology in efforts to make children’s experience of the Internet safe and useful.

    Industry Action:

    • The ISP industry should voluntarily undertake “best practices” to protect minors.
    • The online commercial adult industry should voluntarily take steps to restrict minors’ ready access to adult content.
    NRC summary

    The NRC found “no single or simple answer”, agreed on the capabilities of filters in preventing inadvertent or unhighly-motivated exposure, but also stressed social & educational strategies in addressing motivation, coping, & responsible behavior.

    Social and educational strategies are intended to teach children how to make wise choices about how they behave on the Internet and to take control of their online experiences: where they go; what they see; what they do; who they talk to. Such strategies must be age-appropriate if they are to be effective. Further, such an approach entails teaching children to be critical, skeptical, and self-reflective of the material that they are seeing.

    An analogy is the relationship between swimming pools and children. Swimming pools can be dangerous for children. To protect them, one can install locks, put up fences, and deploy pool alarms. All of these measures are helpful, but by far the most important thing that one can do for one’s children is to teach them to swim.

    Perhaps the most important social and educational strategy is responsible adult involvement and supervision.

    Internet safety education is analogous to safety education in the physical world, and may include teaching children how sexual predators and hate group recruiters typically approach young people, how to recognize impending access to inappropriate sexually explicit material, and when it is risky to provide personal information online. Information and media literacy provide children with skills in recognizing when information is needed and how to locate, evaluate, and use it effectively, irrespective of the media in which it appears, and in critically evaluating the content inherent in media messages. A child with these skills is less likely to stumble across inappropriate material and more likely to be better able to put it into context if and when he or she does.

    Education, supervision, & parental controls/filters seem a more compelling solution. However, bring that up in regard to legislation to age-restrict social media & the tune at lemmy dramatically changes: seems inconsistent.