https://futurism.com/the-byte/government-ai-worse-summarizing

The upshot: these AI summaries were so bad that the assessors agreed that using them could require more work down the line, because of the amount of fact-checking they require. If that’s the case, then the purported upsides of using the technology — cost-cutting and time-saving — are seriously called into question.

    • UlyssesT [he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      21
      ·
      3 months ago

      Continuing to look under LLM rocks of varying size and shininess in search of the solve-every-problem robot god of the future yud-rational

        • UlyssesT [he/him]@hexbear.net
          link
          fedilink
          English
          arrow-up
          12
          ·
          edit-2
          3 months ago

          Stating that newer models that perform better than old models somehow implies that the newer models are completely living up to marketing hype, up to and including calling it “artificial intelligence” to begin with.

          And yes, it’s a known and established issue where some people that stan for these treat printers do see them as replacements for people, not tools. There’s already an entire startup industry of “AI companions” selling that belief, so what I said isn’t as absurd as you claim it is. Besides, I said “robot god of the future” there, not “AI” waifus, but there’s certainly a connection that some true believers make between the two concepts.

            • UlyssesT [he/him]@hexbear.net
              link
              fedilink
              English
              arrow-up
              11
              ·
              edit-2
              3 months ago

              I didn’t mention marketing

              That’s too bad, because “AI” as it stands, and what is branded as “AI,” is not what it claims to be on the label. There are certainly scientific efforts underway to make rudimentary versions of that, but large language models and related technology simply isn’t it, and to believe otherwise is marketing, whether you accept it or not.

              Benchmarks designed to test the machine’s abilities to perform reasoning like humans. And they’re being improved on constantly

              Again, you’re believing in the marketing.

              https://bigthink.com/the-future/artificial-general-intelligence-true-ai/

              https://time.com/collection/time100-voices/6980134/ai-llm-not-sentient/

              Sorry if that rubs ya the wrong way.

              You’re not sorry, this isn’t /r/Futurology or /r/Singularity, and the smuglord closer to your post only makes it worse.

                • UlyssesT [he/him]@hexbear.net
                  link
                  fedilink
                  English
                  arrow-up
                  10
                  ·
                  edit-2
                  3 months ago

                  You seem to have a kind of “head in the sand” approach to this

                  Even more smuglord and there’s so much more text to read. Here we go.

                  (I get it, we have to protect our egos)

                  Maybe educate yourself on what some of the research in this field looks like.

                  Maybe stop ignoring entire fields of research that, to this date, are still figuring out what biological brains are doing and how they are doing them instead of just nodding along to what you already want to believe from people that have blinders for anything outside of their field (computers, in this case). It’s a case of someone with a hammer seeing everything as a nail, and you buying into that.

                  Honestly you sound scared about this stuff.

                  More like tired. If you weren’t so religiously defensive about the apparent advent of whatever you’re hoping for, you’d know that I have on many occasions stated that artificial intelligence is possible and may even be achieved within current lifetimes, but reiterating and refining the currently hyped “AI” product simply isn’t it.

                  It’s like if people were trying to develop rocketry to achieve space travel, but you and yours were smugly stating that this particularly sharp knife will cut the heavens open, just you wait.

                  • soupermen [none/use name]@hexbear.net
                    link
                    fedilink
                    English
                    arrow-up
                    9
                    ·
                    edit-2
                    3 months ago

                    Hey there, I’ve got no stakes here and I don’t want to speak for anyone but I think what happened here was QuillCrestFalconer and DPRK_Chopra were simply pointing out that the technology is rapidly evolving, that it’s capabilities even just a couple years ago were way less than now, and it appears that it will continue to develop like this. So their point would be that we need to still prepare and anticipate that it may soon advance to the point where employers will be more willing to try to replace real workers with it. I don’t think they were implying that this would be a good thing, or that it would be a smart or savvy move, just that it’s a possible and maybe even a likely outcome. We’ve already seen various industries attempt to start doing that with the limited abilities of “AI” already so to me it does seem reasonable to expect them to want to do that more as it gets better. Okay, thanks for reading. 👋

                  • impartial_fanboy [he/him]@hexbear.net
                    link
                    fedilink
                    English
                    arrow-up
                    6
                    ·
                    3 months ago

                    Maybe stop ignoring entire fields of research that, to this date, are still figuring out what biological brains are doing and how they are doing them instead of just nodding along to what you already want to believe from people that have blinders for anything outside of their field (computers, in this case).

                    Well first, brains aren’t the only kind of intelligent biological system but they aren’t actually trying to 1 for 1 recreate the human brain, or any other brain for that matter, that’s just marketing. The generative side of LLM’s is what gets the focus in the media but it’s really not the most scientifically interesting or what will actually change that much all things considered.

                    These systems are absolutely fantastic at finding real patterns in chaotic systems. That’s where the potential lies.

                    It’s like if people were trying to develop rocketry to achieve space travel, but you and yours were smugly stating that this particularly sharp knife will cut the heavens open, just you wait.

                    More like trying to go to the moon with a Civil War era rocket, it is early days yet. But progress is insanely quick.