• tyler@programming.dev
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    6
    ·
    8 months ago

    I’m very confused by this, I had the same discussion with my coworker. I understand what the benchmarks are saying about these models, but have any of y’all actually used deepseek? I’ve been running it since it came out and it hasn’t managed to solve a single problem yet (70b param model, I have downloaded the 600b param model but haven’t tested it yet). It essentially compares to gpt-3 for me, which only cost OpenAI like $4-9 million to train (can’t remember the exact number right now).

    I just do not see the “efficiency” here.

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      19
      ·
      8 months ago

      what if none of it’s good, all of it’s fraud (especially the benchmarks), and having a favorite grifter in this fuckhead industry is just too precious

    • Ksin@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      12
      ·
      8 months ago

      The 70b model is a distilation of Llama3.3, that is to say it replicates the output of Llama3.3 while using the deepseekR1 architecture for better processing efficiency. So any criticism of the capability of the model is just criticism of Llama3.3 and not deepseekR1.

      • bitofhope@awful.systems
        link
        fedilink
        English
        arrow-up
        11
        ·
        8 months ago

        Thank you for shedding light on the matter. I never realized that 69b model is a pisstillation of Lligma peepee point poopoo, that is to say it complicates the outpoop of Lligma4.20 while using the creepbleakR1 house design for better processing deficiency. Now I finally realize that any criticism of Kraftwerk’s 1978 hit Das Model is just criticism of Sugma80085 and not deepthroatR1.

      • froztbyte@awful.systems
        link
        fedilink
        English
        arrow-up
        8
        ·
        8 months ago

        [to the tune of Fort Minor’s Remember The Name]

        10% senseless, 20% post
        15% concentrated spirit of boast
        5% reading, 50% pain
        and a 100% reason to not post here again