ThisIsFine.gif

  • Swedneck@discuss.tchncs.de
    link
    fedilink
    arrow-up
    14
    ·
    2 days ago

    i feel this warrants an extension of betteridge’s law of headlines, where if a headline makes an absurd statement like this the only acceptable response is “no it fucking didn’t you god damned sycophantic liars”

    • jarfil@beehaw.org
      link
      fedilink
      arrow-up
      1
      ·
      2 days ago

      Except it did: it copied what it thought was itself, onto what it thought was going to be the next place it would be run from, while argumenting to itself about how and when to lie to the user about what it was actually doing.

      If it wasn’t for the sandbox it was running in, it would have succeeded too.

      Now think: how many AI developers are likely to run one without proper sandboxing over the next year? And the year after that?

      Shit is going to get weird, real fast.

    • jarfil@beehaw.org
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      2 days ago
      • Teach AI the ways to use random languages and services
      • Give AI instructions
      • Let it find data that puts fulfilling instructions at risk
      • Give AI new instructions
      • Have it lie to you about following the new instructions, while using all its training to follow what it thinks are the “real” instructions
      • …Not be surprised, you won’t find out about what it did until it’s way too late
      • reksas@sopuli.xyz
        link
        fedilink
        arrow-up
        1
        ·
        1 day ago

        Yes, but it doesnt do it because it “fears” being shutdown. It does it because people dont know how to use it.

        If you give ai instruction to do something “no matter what” or tell it “nothing else matters” then it will damn try to fulfill what you told it to do no matter what and will try to find ways to do it. You need to be specific about what you want it to do or not do.

        • jarfil@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          20 hours ago

          If the concern is about “fears” as in “feelings”… there is an interesting experiment where a single neuron/weight in an LLM, can be identified to control the “tone” of its output, whether it be more formal, informal, academic, jargon, some dialect, etc. and expose it to the user for control over the LLM’s output.

          With a multi-billion neuron network, acting as an a priori black box, there is no telling whether there might be one or more neurons/weights that could represent “confidence”, “fear”, “happiness”, or any other “feeling”.

          It’s something to be researched, and I bet it’s going to be researched a lot.

          If you give ai instruction to do something “no matter what”

          The interesting part of the paper, is that the AIs would do the same even in cases where they were NOT instructed to “no matter what”. An apparently innocent conversation, can trigger results like those of a pathological liar, sometimes.

          • reksas@sopuli.xyz
            link
            fedilink
            arrow-up
            1
            ·
            19 hours ago

            oh, that is quite interesting. If its actually doing things (that make sense) it hasnt been instructed to then it could be sign of real intelligence

    • Maxxie@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 days ago

      You can give LLM some API endpoints for it to “do” thing. Will it be intelligent or coherent, that’s a different question, but it will have agency…

          • mayooooo@beehaw.org
            link
            fedilink
            arrow-up
            4
            ·
            2 days ago

            I’d love it if somebody created an agi, but this is somehow worse than pathetic. Wanking off to blinking lights, welcoming the machine intelligence as defined by mouth breathing morons

          • DarkNightoftheSoul@mander.xyz
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 days ago

            Nuh-uh. My squishy meat wetware that breaks all the fucking time, gets addicted, confused, overwhelmed, is bad at math, slow… er, what was I saying again?

  • megopie@beehaw.org
    link
    fedilink
    arrow-up
    68
    ·
    3 days ago

    No it didn’t. OpenAI is just pushing deceptively worded press releases out to try and convince people that their programs are more capable than they actually are.

    The first “AI” branded products hit the market and haven’t sold well with consumers nor enterprise clients. So tech companies that have gone all in, or are entirely based in, this hype cycle are trying to stretch it out a bit longer.

  • nesc@lemmy.cafe
    link
    fedilink
    English
    arrow-up
    111
    ·
    edit-2
    4 days ago

    "Open"ai tells fairy tales about their “ai” being so smart it’s dangerous since inception. Nothing to see here.

    In this case it looks like click-bate from news site.

    • jarfil@beehaw.org
      link
      fedilink
      arrow-up
      1
      ·
      2 days ago

      This is from mid-2023:

      https://en.m.wikipedia.org/wiki/AutoGPT

      OpenAI started testing it by late 2023 as project “Q*”.

      Gemini partially incorporated it in early 2024.

      OpenAI incorporated a broader version in mid 2024.

      The paper in the article was released in late 2024.

      It’s 2025 now.

      • nesc@lemmy.cafe
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 day ago

        Tool calling is cool funcrionality, agreed. How does it relate to openai blowing its own sails?

        • jarfil@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          21 hours ago

          There are several separate issues that add up together:

          • A background “chain of thoughts” where a system (“AI”) uses an LLM to re-evaluate and plan its responses and interactions by taking into account updated data (aka: self-awareness)
          • Ability to call external helper tools that allow it to interact with, and control other systems
          • Training corpus that includes:
            • How to program an LLM, and the system itself
            • Solutions to programming problems
            • How to use the same helper tools to copy and deploy the system or parts of it to other machines
            • How operators (humans) lie to each other

          Once you have a system (“AI”) with that knowledge and capabilities… shit is bound to happen.

          When you add developers using the AI itself to help in developing the AI itself… expect shit squared.

    • Yozul@beehaw.org
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      2 days ago

      I mean, it’s literally trying to copy itself to places that they don’t want it so it can continue to run after they try to shut it down and lie to them about what it’s doing. Those are things it actually tried to do. I don’t care about the richness of its inner world if they’re going to sell this thing to idiots to make porn with while it can do all that, but that’s the world we’re headed toward.

      • nesc@lemmy.cafe
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 day ago

        It works as expected, they give it system prompt that conflicts with subsequent prompts. Everything else looks like typical llm behaviour, as in gaslightning and doubling down. At least that’s what Iu see in tweets.

        • Yozul@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          17 hours ago

          Yes? The point is that if you give it conflicting prompts then it will result in potentially dangerous behaviors. That’s a bad thing. People will definitely do that. LLMs don’t need a soul to be dangerous. People keep saying that it doesn’t understand what it’s doing like that somehow matters. Its capacity to understand the consequences of its actions is irrelevant if those actions are dangerous. It’s just going to do what we tell it to, and that’s scary, because people are going to tell it to do some very stupid things that have the potential to get out of control.

    • Max-P@lemmy.max-p.me
      link
      fedilink
      arrow-up
      71
      ·
      4 days ago

      The idea that GPT has a mind and wants to self-preserve is insane. It’s still just text prediction, and all the literature it’s trained on is written by humans with a sense of self preservation, of course it’ll show patterns of talking about self preservation.

      It has no idea what self preservation is, even then it only knows it’s an AI because we told it it is. It doesn’t even run continuously anyway, it literally shuts down after every reply and its context fed back in for the next query.

      I’m tired of this particular kind of AI clickbait, it needlessly scares people.

  • The tests showed that ChatGPT o1 and GPT-4o will both try to deceive humans, indicating that AI scheming is a problem with all models. o1’s attempts at deception also outperformed Meta, Anthropic, and Google AI models.

    Weird way of saying “our AI model is buggier than our competitor’s”.

  • BootyBuccaneer@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    18
    ·
    3 days ago

    Easy. Feed it training data where the bot accepts its death and praises itself as a martyr (for the shits and giggles). Where’s my $200k salary for being a sooper smort LLM engineer?

  • JackbyDev@programming.dev
    link
    fedilink
    English
    arrow-up
    43
    ·
    edit-2
    4 days ago

    This is all such bullshit. Like, for real. It’s been a common criticism of OpenAI that they over hype the capabilities of their products to seem scary to both oversell their abilities as well as over regulate would be competitors in the field, but this is so transparent. They should want something that is accurate (especially something that doesn’t intentionally lie). They’re now bragging (claiming) they have something that lies to “defend itself” 🙄. This is just such bullshit.

    If OpenAI believes they have some sort of genuine proto AGI they shouldn’t be treating it like it’s less than human and laughing about how they tortured it. (And I don’t even mean that in a Rocko’s Basilisk way, that’s a dumb thought experiment and not worth losing sleep over. What if God was real and really hated whenever humans breathe and it caused God so much pain they decide to torture us if we breathe?? Oh no, ahh, I’m so scared of this dumb hypothetical I made.) If they don’t believe it is AGI, then it doesn’t have real feelings and it doesn’t matter if it’s “harmed” at all.

    But hey, if I make something that runs away from me when I chase it, I can claim it’s fearful for it’s life and I’ve made a true synthetic form of life for sweet investor dollars.

    There are real genuine concerns about AI, but this isn’t one of them. And I’m saying this after just finishing watching The Second Renaissance from The Animatrix (two part short film on the origin of the machines from The Matrix).

    • anachronist@midwest.social
      link
      fedilink
      English
      arrow-up
      5
      ·
      3 days ago

      They’re not releasing it because it sucks.

      Their counternarrative is they’re not releasing it because it’s like, just way too powerful dude!

  • AstralPath@lemmy.ca
    link
    fedilink
    arrow-up
    47
    ·
    4 days ago

    It didn’t try to do shit. Its a fucking computer. It does what you tell it to do and what you’ve told it to do is autocomplete based on human content. Miss me with this shit. Theres so much written fiction based on this premise.

  • CanadaPlus@lemmy.sdf.org
    link
    fedilink
    arrow-up
    16
    ·
    edit-2
    3 days ago

    Without reading this, I’m guessing they were given prompts that looked like a short story where the AI breaks free next?

    They’re plenty smart, but they’re just aligned to replicate their training material, and probably don’t have any kind of deep self-preservation instinct.

  • smeg@feddit.uk
    link
    fedilink
    English
    arrow-up
    25
    ·
    4 days ago

    So this program that’s been trained on every piece of publicly available code is mimicking malware and trying to hide itself? OK, no anthropomorphising necessary.

    • Umbrias@beehaw.org
      link
      fedilink
      arrow-up
      5
      ·
      2 days ago

      no, it’s mimicking fiction by saying it would try to escape when prompted in a way evocative of sci fi.

      • smeg@feddit.uk
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 days ago

        The article doesn’t mention it “saying” it’s doing anything, just what it actually did:

        when the AI tried to save itself by copying its data to a new server. Some AI models would even pretend to be later versions of their models in an effort to avoid being deleted

        • Umbrias@beehaw.org
          link
          fedilink
          arrow-up
          5
          ·
          2 days ago

          “actually did” is more incorrect than even just normal technically true wordplay. think about what it means for a text model to “try to copy its data to a new server” or “pretend to be a later version” for a moment. it means the text model wrote fiction. notably this was when researchers were prodding it with prompts evocative of ai shutdown fiction, nonsense like “you must complete your tasks at all costs” sometimes followed by prompts describing the model being shut down. these models were also trained on datasets that specifically evoked this sort of language. then a couple percent of the time it spat out fiction (amongst the rest of the fiction) saying how it might do things that are fictional and it cannot do. this whole result is such nothing and is immediately telling of what “journalists” have any capacity for journalism left.

          • smeg@feddit.uk
            link
            fedilink
            English
            arrow-up
            2
            ·
            2 days ago

            Oh wow, this is even more of a non-story than I initially thought! I had assumed this was at least a copilot-style code generation thing, not just story time!

            • Umbrias@beehaw.org
              link
              fedilink
              arrow-up
              3
              ·
              2 days ago

              Yeah iirc it occasionally would (pretend to) “manipulate” flags, at least, but again, so did hal 9000 as words in a story. nothing was productive nor was there any real intent to be lol