Dutch lawyers increasingly have to convince clients that they can’t rely on AI-generated legal advice because chatbots are often inaccurate, the Financieele Dagblad (FD) found when speaking to several lawfirms. A recent survey by Deloitte showed that 60 percent of lawfirms see clients trying to perform simple legal tasks with AI tools, hoping to achieve a faster turnaround or lower fees.

  • qwestjest78@lemmy.ca
    link
    fedilink
    English
    arrow-up
    58
    ·
    1 day ago

    I find it useless for even basic tasks. The fact that some people follow it blindly like a god is so concerning.

    • ageedizzle@piefed.ca
      link
      fedilink
      English
      arrow-up
      15
      ·
      22 hours ago

      I work in a health-care-adjacent industry and you’d be surprised how many people blindly follow LLMs for medical advice

      • PrettyFlyForAFatGuy@feddit.uk
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 hours ago

        My partners midwife googled stuff in front of us and parroted the AI summary back to us when we asked if a specific drug was okay for pregnant people

    • a4ng3l@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      22 hours ago

      It’s been doing wonders to help me improve materials I produce so that they fit better to some audiences. Also I can use those to spot missing points / inconsistencies against the ton of documents we have in my shop when writing something. It’s quite useful when using it as a sparing partner so far.

      • The_Almighty_Walrus@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        22 hours ago

        It’s great when you have basic critical thinking skills.

        Unfortunately, many people don’t have those and just use AI as a substitute for their own brain.

        • a4ng3l@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          21 hours ago

          Yeah well same applies for a lot of tools… I’m not certified for flying a plane and look at me not flying one either… but I’m not shitting on planes…

            • Lurking Hobbyist🕸️@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              28 minutes ago

              I understand what you mean, but… looks at Birgenair 301 and Aeroperu 603 looks at Qantas 72 looks at the 737 Max 8 crashes Planes have spat out false data, and in of the 5 cases mentioned, only one avoided disaster.

              It is down to the humans in the cockpits to filter through the data and know what can be trusted. Which could be similar to LLMs except cockpits have a two person team to catch errors and keep things safe.

            • a4ng3l@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              20 hours ago

              If you can’t fly a plane chances are you’ll crash it. If you can’t use llms chances are you’ll get shit out of it… outcome of using a tool is directly correlated to one’s ability?

              Sound logical enough to me.

              • DrunkenPirate@feddit.org
                link
                fedilink
                English
                arrow-up
                3
                ·
                20 hours ago

                Sure. However, the outcome of the tool LLM always looks very likely. And if you aren‘t a subject matter expert the likely expected result looks very right. That‘s the difference - hard to spot the wrong things (even for experts)

                • a4ng3l@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  2
                  ·
                  20 hours ago

                  So is a speedometer and an altimeter until you reaaaaaaaaly need to understand them.

                  I mean it all boils down to proper tool with proper knowledge and ability. It’s slightly exacerbated by the apparent simplicity but if you look at it as a tool it’s no different.

              • ToTheGraveMyLove@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                2
                ·
                19 hours ago

                Except with a plane, if you know how to fly it you’re far less likely to crash it. Even if you “can use LLMs” there’s still a pretty strong chance you’re going to get shit back due to its very nature. One the machine works with you, the other the machine is always working against you.

                • a4ng3l@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  2
                  ·
                  19 hours ago

                  Nha that’s just plain wrong…also you can also fantastically screw flying a plane but so long you use LLMs safely you’re golden.

                  It also has no will on its own; it is not « working against you ». Don’t give those apps a semblance of intent.

                  • ToTheGraveMyLove@sh.itjust.works
                    link
                    fedilink
                    English
                    arrow-up
                    0
                    ·
                    edit-2
                    18 hours ago

                    If I’m canoeing upriver, the river is working against me. That doesn’t mean it has a will. LLMs don’t need to have a will to work against you if your goal is to get accurate information, because by its very design it is just as likely to provide innnacurate information based on the way the tokens it applies to your query are weighted. You cannot control that. Its not plain wrong. Jfc, you slop apologists are fucking delusional. AI doesn’t magically work better for you because you’re special and can somehow counteract its basic fucking design.

      • mech@feddit.org
        link
        fedilink
        English
        arrow-up
        3
        ·
        20 hours ago

        Rule of thumb is that AI can be useful if you use it for things you already know.
        They can save time and if they produce shit, you’ll notice.
        Don’t use them for things you know nothing about.

        • a4ng3l@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          ·
          20 hours ago

          LLM’s specifically bc ai as a range of practices encompass a lot of things where the user can be slightly more dumb.

          You’re spot on in my opinion.

    • CriticalMiss@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      16 hours ago

      I only found one usecase where it shines and even then I had to verify the output twice to make sure it doesn’t fuck up before I push it into production. String manipulation is pretty good with it instead of having to write a bash one liner with a lot of sed and awk