cross-posted from: https://lemmy.world/post/43768262

Some may have believed they were against AI being used for war… they just don’t want it to make the final kill decision.

The argument for them is that AI in the military was inevitable, so their position is a reasonable one.

  • crunchy@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    49
    ·
    8 hours ago

    Well yeah, Anthropic’s statement said as much. And they already had a contract with the DoD. This isn’t a gotcha.

      • Serinus@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        3
        ·
        4 hours ago

        I mean, yeah. They’re the only ones trying to hold out some reasonable constraints.

        • XLE@piefed.social
          link
          fedilink
          English
          arrow-up
          5
          ·
          4 hours ago

          Anthropic has been willing to throw away so many reasonable constraints, and they always have. The companies are disturbingly similar, and Anthropic is worse in some ways.

          Ignoring the similarities to imply Anthropic as a hero is just bad journalism.

  • Iconoclast@feddit.uk
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    1
    ·
    8 hours ago

    My understanding is that it’s not military use broadly that they object to but the use of their systems for the development of fully autonomous drones.

    • XLE@piefed.social
      link
      fedilink
      English
      arrow-up
      2
      ·
      3 hours ago

      This is incorrect.

      Anthropic contradicts this, showing an aggressive willingness to work with the Trump department of “War.”

      Even fully autonomous weapons (those that take humans out of the loop entirely and automate selecting and engaging targets) may prove critical for our national defense… We have offered to work directly with the Department of War on R&D to improve the reliability of these systems

      • Iconoclast@feddit.uk
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        1
        ·
        edit-2
        7 hours ago

        Here’s the full quote including the parts you conveniently left out.

        Even fully autonomous weapons (those that take humans out of the loop entirely and automate selecting and engaging targets) may prove critical for our national defense. But today, frontier AI systems are simply not reliable enough to power fully autonomous weapons. We will not knowingly provide a product that puts America’s warfighters and civilians at risk. We have offered to work directly with the Department of War on R&D to improve the reliability of these systems, but they have not accepted this offer.

        Our strong preference is to continue to serve the Department and our warfighters—with our two requested safeguards in place.

        Source

        • XLE@piefed.social
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          7
          ·
          edit-2
          7 hours ago

          You said Anthropic didn’t want to develop autonomous weapons. Anthropic contradicts you. They do want to develop them.

          Can you acknowledge this fact?

          I love how Anthropic only draws the line at autonomously killing American “warfighters” too. I guess some lives are worth more than others.

            • Serinus@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              4 hours ago

              It’s not a very solid point. They said they may become necessary at some point, but right now they’re irresponsible.

              They’re not ruling it out in the future, but their focus is on today’s problem.

              • XLE@piefed.social
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                4 hours ago

                Serinus, did you see the part where Anthropic wants to develop them with the US military?

  • criticon@lemmy.ca
    link
    fedilink
    English
    arrow-up
    1
    ·
    5 hours ago

    Is there a “good-ish” AI chatbot? I uninstalled chatgpt and started using claude over the weekend 🤡

    I don’t use it that much, maybe a couple of questions a week and some help understanding Japanese grammar

    • breadguy@kbin.earth
      link
      fedilink
      arrow-up
      1
      ·
      49 minutes ago

      check huggingface, it’s got all the open models. inference is pay as you go, though most models are insanely cheap

    • frongt@lemmy.zip
      link
      fedilink
      English
      arrow-up
      4
      ·
      3 hours ago

      If you can’t validate the answers it gives, I would recommend not using it. It could be giving you complete nonsense in Japanese and you’ll have no way to know until years later someone looks at you funny when you say something and you explain “I learned Japanese from chatgpt”.

    • onnekas@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 hours ago

      I’m quite happy with Mistrals LeChat. I have not done much research on Mistral but from the headlines I reas they don’t seem like bad guys.

    • floquant@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      8
      ·
      5 hours ago

      Depends on your definition, Anthropic has been the somewhat less evil on the scene, doing a lot of research to actually understand what they’re building instead of just making bold claims whenever they launch a new model, but it’s still relative to huge AI companies. The more ethical choice would be local models, again depending on what you see as the ethical issues of AI

    • partial_accumen@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      5 hours ago

      Depends on your definition of “good-ish”. Do you mean:

      • performance/accuracy?
      • ethical origin?
      • ethical ongoing operation?
      • privacy/future data harvesting concerns?

      Running one locally on your own hardware would likely reach “good-ish” with some sacrifices against performance/accuracy (unless you’ve got a lot of expensive hardware to run very large models). As far as ethical origins, there are few small models trained on public domain/nonstolen content, but their functions are far more limited.

      • criticon@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        3 hours ago

        I mean good-ish in the lesser-evil type of thing. I don’t expect any of those to be 100% ethical but there are some that are a lot worse than others

        I don’t really have a computer capable of running a local AI. I have an i3 laptop from around 6 years ago with 12gb of RAM and integrated graphics

        • partial_accumen@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          3 hours ago

          I mean good-ish in the lesser-evil type of thing. I don’t expect any of those to be 100% ethical but there are some that are a lot worse than others

          Ethics are subjective. “Good-ish” to you may mean you’re fine if its trained on copyrighted works as long as it wasn’t done with electricity from diesel generators belching exhaust into the local Memphis atmosphere (I’m looking at you Grok). Llama doesn’t do the diesel generator thing, but its a product of Facebook corporation. So is that “good-ish” to you or not? I don’t know. That’s up to you.

          It may not be fast, but your i3 laptop with 12GB of system RAM can absolutely run a local LLM. This is where that “performance/accuracy” question I raised comes in. It won’t be very fast, and you won’t be able to run the most common large models like GPT-5 etc. However, if your needs are light, light models exist. Give this a read

    • Iconoclast@feddit.uk
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      5 hours ago

      If by “good” you mean one that more reliably answers your questions correctly, then no. That’s not really what these systems are good at. They’re fully capable of giving a solid, accurate answer, but you can simply never trust it to be correct. They’re great for chit-chat and bouncing around ideas if you’re into that, but it’s not an oracle.

      When it comes to translating languages, that’s one of the few things LLMs are actually somewhat decent at, and I don’t think there’s much difference between them in that regard.

      • criticon@lemmy.ca
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 hours ago

        No, I mean good in the sense of less evil. I don’t ask any question to a bot where I need complete accuracy