Prompted by the recent troll post, I’ve been thinking about AI. Obviously we have our criticisms of both the AI hype manchildren and the AI doom manchildren (see title of the post. This is a Rationalist free post. Looking for it? Leave)

But looking at the AI doom guys with an open mind, sometimes it appear that they make a halfway decent argument that’s backed up by real results. This YouTube channel has been talking about the alignment problem for a while, and I think he probably is a bit of a Goodhart’s Law merchant (as in, by making a career out of measuring the dangers of AI, his alarmism is structural) so he should be taken with a grain of salt, it does feel pretty concerning that LLMs show inner misalignment and are masking their intentions (to anthropomorphize) under training vs deployment.

Now, I mainly think that these people are just extrapolating out all the problems with dumb LLMs and saying “yeah but if they were AGI it would become a real problem” and while that might be true if taking the premise at face value, the idea that AGI will ever happen is itself pretty questionable. The channel I linked has a video arguing that AGI safety is not a Pascal’s mugging, but I’m not convinced.

Thoughts? Does the commercialization of dumb AI make it a threat on a similar scale to hypothetical AGI? Is this all just a huge waste of time to think about?

  • buckykat [none/use name]@hexbear.net
    link
    fedilink
    English
    arrow-up
    13
    ·
    3 days ago

    The thing that currently exists by the name of “AI” is never gonna become anything like AGI, but it is bad in its own way.

    I don’t think AGI is fundamentally impossible, because dualism is fake and consciousness is a real material process that exists in the universe, but it’s not a thing we are anywhere near understanding how to build.

    Further, the roko’s basilisk is literally just a dumber pascal’s wager, because it’s a pascal’s wager wherein not giving a shit about god is a total defense against divine punishment.

    • Des [she/her, they/them]@hexbear.net
      link
      fedilink
      English
      arrow-up
      6
      ·
      3 days ago

      I’m torn between either there being a quantum theory of consciousness (thus requiring quantum effects to be utilized to create true AGI) or it requiring actual full, high resolution simulation of an actual living mind or a very close approximation.

      • Carl [he/him]@hexbear.net
        link
        fedilink
        English
        arrow-up
        5
        ·
        edit-2
        3 days ago

        I’m increasingly of the belief that a major part of our own consciousness is socially contingent, so creating an artificial one can’t be done in one fell swoop by one computer getting really smart, it has to be the result of reverse-engineering the entire process of evolution that lead to consciousness as we understand it.

        • Des [she/her, they/them]@hexbear.net
          link
          fedilink
          English
          arrow-up
          6
          ·
          edit-2
          3 days ago

          really interesting because my partner and I were doing some worldbuilding and came up with something like this! we had two methodologies: one was for artificial life which involved exactly what you describe; starting with a deep, complex simulation sped up by asteroid-sized computers that start from scratch. after you have an artificial life model the AI basically had to be bound to a human at birth and “grow up” and learn with them, essentially developing in parallel as an artificial sibling while they exist in a symbiotic relationship. this becomes a cultural norm, and ties artificial life to humanity as a familial relation. (this was a far future society where single child households were the norm)

      • buckykat [none/use name]@hexbear.net
        link
        fedilink
        English
        arrow-up
        4
        ·
        3 days ago

        Either way, those are real physical processes which could, in principle, be replicated. My general layman’s impression is that claims of quantum being involved are more of a last redoubt of dualists than a serious theory though.

    • semioticbreakdown [she/her]@hexbear.net
      link
      fedilink
      English
      arrow-up
      3
      ·
      3 days ago

      I think the conception of AGI as a machine is holding back its development ontologically speaking. Reductionism too. A consciousness is dynamic, and fundamentally part of a dynamic organism. It can’t be removed from the context of the broader systems of the body or the world the body acts on. Even its being comes secondary to activities it takes. So I’m not really scared of it existing in the abstract. I’m a lot more afraid of mass production commodifying consciousness itself. Everything that people fear in AGI is a projection of the worst ills of the system we live in. Roko’s basilisk is dumb as fuck also

      • buckykat [none/use name]@hexbear.net
        link
        fedilink
        English
        arrow-up
        6
        ·
        3 days ago

        (I ain’t readin’ all that) but what the abstract describes isn’t even close to the worst thing I’ve read about LLMs doing this week. I don’t exactly trust the LLM companies’ ideas of what is or is not “harmful.” Shit like people using the LLMs as therapists, or worse, oracles is much worse in my opinion, and that doesn’t require any “pretend to be evil for training” hijinks.

      • BountifulEggnog [she/her]@hexbear.net
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        3 days ago

        Doesn’t really strike me as sinister, just annoying for finetuners. They trained a model from the ground up to not be harmful and it tries its best. Even with further training it still retains some of that. To me this paper shows that a model’s “goals”, what you trained it to do initially, however you want to phrase that, is baked into it and changing that after the fact is hard. Highlights how important early training is I guess.

        • FunkyStuff [he/him]@hexbear.netOP
          link
          fedilink
          English
          arrow-up
          3
          ·
          3 days ago

          Kinda problematic that it means we can’t ever really be sure that we’re catching problematic behavior in the training stage of any AI system, though, right? Sadly I find it hard to think of good uses of LLMs or other genAI outside of capitalism, but if there were any, the fact that it’s possible for it to behave duplicitously like that is a pretty big problem.

  • pcalau12i@lemmygrad.ml
    link
    fedilink
    English
    arrow-up
    6
    ·
    3 days ago

    I don’t think AI safety is such a big problem that it means we gotta stop building AI or we’ll destroy the world or something, but I do agree there should be things like regulations, oversight, some specialized people to make sure AI is being developed in a safe way just to help mitigate problems that could possibly come up. There is a mentality that AI will never be as smart as humans so any time people suggest some sort of policies for AI safety that it’s unreasonable because it’s overhyping how good AI is and it won’t get to a point of being dangerous for a long time. But if we have this mentality indefinitely then eventually when it does become dangerous we’d have no roadblocks and it might actually become a problem. I do think completely unregulated AI developed without any oversight or guardrails could in the future lead to bad consequences, but I also don’t think that is something that can’t be mitigated with oversight. I don’t believe for example like an AGI will somehow “break free” and take over the world if it is ever developed. If it is “freed” in a way that starts doing harm, it would be because someone allowed that.

  • Mardoniush [she/her]@hexbear.net
    link
    fedilink
    English
    arrow-up
    6
    ·
    edit-2
    3 days ago

    I’m going to talk about the hard AI problem of rogue AI, not the obviously more pressing one of AI being used for bad things, because that can’t be stopped

    So, Big Yud is mad and constantly full of nuclear Bay Area brainworm takes. but some of his arguments against alignment make sense. At least enough to engage with them and try to solve them, but bombing data centres et al is dumb and won’t solve the issue since most code is already so inefficient.

    AI is not going to go FOOM in the way he suggested, basically because LLMs exist, which he thought were impossible as part of his arguments. He never considered we would have time to practice alignment with extremely limited AI at near human levels. And that’s kicked the core of his doom argument away.

    I think UlyssesT and I once agreed that he is the dumbest smart guy alive and will probably become leftist after trying and failing at every tech nerd principal. If someone could keep him the hell away from the Bay Area he might normalise and stop writing weird shit.

    For what it’s worth he also thinks Transformers have no chance at AGI, at least alone. And his thoughts on the dumb approaches used by AI companies are sound. But he is arguing in good faith and has probably thought about this more than anyone else. (Nobody even in the Yud cult except for a few weirdos ever took Roko’s Basilisk seriously. Yud deleted the post mostly because he saw that people would go weird about it, a rare example of him being exactly correct.)

    As for everyone else…Altman doesn’t care and is using it for market capture. Anthropic probably does care but is more or less forced down the same path of ineffective safety that captures the market.

    Ilya definitely cares but for all the wrong reasons, he’s massively pro Israel so he’ll never achieve a consistently aligned AI anyway, but I’d not want him working on this stuff.

    But ultimately hard alignment is a secondary concern against dumb corporate HR managers trying to AI the compliance team.

    • FunkyStuff [he/him]@hexbear.netOP
      link
      fedilink
      English
      arrow-up
      5
      ·
      3 days ago

      EDIT: didn’t read that this was a non Big Yud post, so sorry, but…you can’t really talk about AI safety in the general sense without him.( He dominated the field for 20 years, long before his dumb rationalist cult.)

      I forgive you. The traditions of all Yud generations weigh like rational nightmares on the alignments of the sneerers.

  • insurgentrat [she/her, it/its]@hexbear.net
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 days ago

    The real research is useful. Putting aside nerd rapture stuff which is probably fantasy as far as we can tell it ought to be possible to make thinking machines. whether we do? who knows. Increasingly we make and use machines able to make at least some decisions and the good research into alignment and training helps with even mundane shit; such as making sure the machine is actually seeing cancers and not say the institution the MRI was taken at.

    Even with relatively stupid machines it would be important to have certainty that idk a car isn’t going to take a shortcut through a pedestrian. You don’t need to reach AM level for it to be important that an autonomous system using some degree of heuristic decision making, quote unquote, cares about the things we do.

  • FunkyStuff [he/him]@hexbear.netOP
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 days ago

    For reference, here’s a video about a recent paper that showed a real example of an AI behaving differently under training, with the explicit purpose of avoiding the detection of behavior it would take on in deployment.

  • semioticbreakdown [she/her]@hexbear.net
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    3 days ago

    What troll post?

    I think the risk posed by capitalist deployment of AI technologies far outweighs the dangers of misalignment, although alignment is definitely a problem