• 0 Posts
  • 18 Comments
Joined 1 year ago
cake
Cake day: June 18th, 2023

help-circle
  • In 1962 Phillip K Dick put out a book called “Man in the High Castle.” In it there was a scene that stuck out to me, and seems more and more relevant as this AI wave continues.

    In it a man has two identical lighters. Each made in the same year by the same manufacturer. But one was priceless and one was worthless.

    The priceless one was owned by Abraham Lincoln and was in his pocket on the night he was assassinated. He had a letter of certification as such, and could trace the ownership all the way back to that night.

    And he takes them both and mixes them up and asks which is the one with value. If you can no longer discern the one with “historicity,” then where is it’s value?

    And every time I see an article like this I can’t help but think about that. If I tell you about the life and hardship of an artist, and then present you two poems, one that he wrote and one that was spit out by an LLM, and you cannot determine which has the true hardship and emotion tied to it, then which has value? What if I killed the artist before he could reveal which one was the “true” poem? How do you know which is a powerful expression of the artist’s oppression, and which is worthless, randomly generated swill?




  • testfactor@lemmy.worldtoTechnology@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    3
    ·
    1 month ago

    Sure, maybe, but I’d also say you shouldn’t let the perfect be the enemy of the good.

    Yes, we should absolutely have better mental healthcare safety nets. Yes, false positives are probably a pretty common prank.

    But this isn’t a zero sum game. This can work on tandem with a therapist/counsellor to try and identify someone before they shoot up a school and get them help. This might let the staff know a kid is struggling with suicidal ideation before they find the kid OD’d on moms sleeping pills.

    In an ideal world would this be unnecessary? Absolutely. But we don’t live in that ideal world.


  • testfactor@lemmy.worldtoTechnology@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    2
    ·
    1 month ago

    That argument could be expanded to any tool though.

    People run people over with cars or drive drunk. Ban cars?

    People use computers to distribute CP. Ban computers?

    People use baseball bats to bludgeon people to death. Ban baseball?

    The question of if a tool should be banned is driven by if its utility is outweighed by the negative externalities of use by bad actors.

    The answer is wildly more nuanced than “if it can hurt someone it must be banned.”


  • testfactor@lemmy.worldtoTechnology@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    14
    ·
    1 month ago

    You say “the last time this happened” as if this wasn’t a generalized trend across all schooling for the past decade or so.

    Out of the tens of thousands of schools implementing systems like this, I’m not surprised that one had some letch who was spying on kids via webcam.

    And I’m all for having increased forms of oversight and protection to prevent that kind of abuse.

    But this argument is just as much of a “won’t someone think of the children” as the opposite. Just cause one school out of thousands did a bad thing, doesn’t mean the tech is worthless or bad.


  • testfactor@lemmy.worldtoTechnology@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    36
    arrow-down
    20
    ·
    1 month ago

    This article feels pretty disingenuous to me.

    It glosses over the fact that this is surveillance on computers that the school owns. This isn’t them spying on kids personal laptops or phones. This is them exercising reasonable and appropriate oversight of school equipment.

    This is the same as complaining that my job puts a filter on my work computer that lets them know if I’m googling porn at work. You can cry big brother all you want, but I think most people are fine with the idea that the corporation I work for has a reasonable case for putting monitoring software on the computer they gave me.

    The article also makes the point that, while the companies claim they’ve stopped many school shootings before they’ve happened, you can’t prove they would have happened without intervention.

    And sure. That’s technically true. But the article then goes on to treat that assertion as if it’s proof that the product is worthless and has never prevented a school shooting, and that’s just bad logic.

    It’s like saying that your alarm clock has woken you up 100 days in a row, and then being like, “well, there’s no proof that you wouldn’t have woken up on time anyway, even if the alarm wasn’t there.” Yeah, sure. You can’t prove a negative. Maybe I would usually wake up without it. I’ve got a pretty good sleep schedule after all. But the idea that all 100 are false positives seems a little asinine, no? We don’t think it was effective even once?




  • I think the issue is that, while a country is certainly allowed to write it’s own laws, the idea that it is deeply fundamentally immoral for the government to prevent someone from saying something (or compel them to say something) is very deeply baked into the American zeitgeist (of which I am a part.)

    So in the same way that a country is perfectly within its sovereign rights to pass a law that women are property or minorities don’t have the right to vote, I can still say that it feels wrong of them to do so.

    And I would also decry a country that kicks out a company that chooses to employ women or minorities in violation of such a law, even if that is technically their sovereign right to do so.