• Mossy Feathers (She/They)@pawb.social
    link
    fedilink
    arrow-up
    19
    ·
    edit-2
    1 month ago

    Everyone’s trying to recapture the dotcom bubble; but they don’t realize tech is gonna need considerably more money than they already have to do something that crazy again. Furthermore, when it comes to AI specifically, if you give them the money they need to actually achieve AGI, then there’s a very real chance your investments will be worthless the moment they succeed.

      • Mossy Feathers (She/They)@pawb.social
        link
        fedilink
        arrow-up
        9
        arrow-down
        1
        ·
        1 month ago

        Why would money become worthless if AGI is invented? Best case scenario is a benevolent AGI which would likely use its power to phase out capitalism, worst case scenario is that the AGI goes apeshit and, for one reason or another, decides that humanity just has to go. Either way, your money is gonna be worthless.

        The only way your money would retain its value is if the AGI is roped into suppressing the masses. However, I think capitalists would struggle to keep a true AGI reigned in; so imo, it’s questionable as to whether or not the middle road would be “true” AGI or just a very competent computer program (the former being capable of coming to its own conclusions from the information it’s given, the latter being nothing more than pre-programmed conclusions).

            • nickwitha_k (he/him)@lemmy.sdf.org
              link
              fedilink
              arrow-up
              3
              ·
              1 month ago

              There’s a vocal group of people who seem to think that LLMs can achieve consciousness despite the fact that it is not possible due to the way that LLMs fundamentally work. They have largely been duped by advanced LLMs’ ability to sound convincing (as well as a certain conman executive officer). These people often also seem to believe that by dedicating more and more resources to running these models, they will achieve actual general intelligence and that an AGI can save the world, releasing them of the responsibility to attempt to fix anything.

              That’s my point. AGI isn’t going to save us and LLMs (by themselves), regardless of how much energy is pumped into them, will not ever achieve actual intelligence.

              • Mossy Feathers (She/They)@pawb.social
                link
                fedilink
                arrow-up
                2
                ·
                1 month ago

                But an AGI isn’t an LLM. That’s what’s confusing me about your statement. If anything I feel like I already covered that, so I’m not sure why you’re telling me this. There’s no reason why you can’t recreate the human brain on silicon, and eventually someone’s gonna do it. Maybe it’s one of our current companies, maybe it’s a future company. Who knows. Except that a true AGI would turn everything upside down and inside out.

                • nickwitha_k (he/him)@lemmy.sdf.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  1 month ago

                  I think, possibly, my tired brain at the time thought that you are implying LLM -> AGI. And I do agree that that’s no reason, beyond time and available technology that a model of a brain cannot be made. I would question whether digital computers are capable of accurately simulating neurons, at least, without requiring more components (more bits of resolution).

                  For full disclosure, I am supportive of increasing the types of sentience in the known universe. Though, not at the expense of biosphere habitability. Whether electronic or biological, sharing the world with more types of sentients would make it a more interesting place.

                  Except that a true AGI would turn everything upside down and inside out.

                  Very likely. Especially if “human rights” aren’t pre-emptively extended to cover non-human sentients. But, the existence of AGI, alone, is not likely to cause either doomsday or save us from it, which seem to be the most popularly envisaged scenarios.

                  • Mossy Feathers (She/They)@pawb.social
                    link
                    fedilink
                    arrow-up
                    1
                    ·
                    1 month ago

                    I think, possibly, my tired brain at the time thought that you are implying LLM -> AGI.

                    Ah, okay. I’ve been there lol. I hope I didn’t come off as confrontational, I was very confused and concerned that I had badly explained myself. My apologies if I did.

        • psud@aussie.zone
          link
          fedilink
          arrow-up
          1
          ·
          1 month ago

          So you make an AGI, what gives it the power to do any damage? We have loads of biological intelligences, even pretty damn clever ones like Ted Kaczynski (the Unabomber)

          They rarely got significant power. Those that did were super charismatic. Do you expect charisma to be easily accessible to an AGI?

          The usually proposed path to paperclip maximiser is that the AGI is put in charge of a factory that can make nano machines and follows orders strictly. We don’t have such factories.

          I can’t imagine anyone handing over nukes to AGI as human leaders like being in charge of them

          What makes the machine brain so much more effective than Ted Kaczynski?