• OsrsNeedsF2P@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    7 days ago

    I work with people who work in this field. Everyone knows this, but there’s also an increased effort in improvements all across the stack, not just the final LLM. I personally suspect the current generation of LLMs is at its peak, but with each breakthrough the technology will climb again.

    Put differently, I still suspect LLMs will be at least twice as good in 10 years.

  • jpablo68@infosec.pub
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 days ago

    I just want a portable self hosted LLM for specific tasks like programming or language learning.

    • plixel@programming.dev
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 days ago

      You can install Ollama in a docker container and use that to install models to run locally. Some are really small and still pretty effective, like Llama 3.2 is only 3B and some are as little as 1B. It can be accessed through the terminal or you can use something like OpenWeb UI to have a more “ChatGPT” like interface.

      • cybersandwich@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 days ago

        I have a few LLMs running locally. I don’t have an array of 4090s to spare so I am limited to the smaller models 8B and whatnot.

        They definitely aren’t as good as anything you get remotely. It’s more private and controlled but it’s much less useful (I’ve found) than any of the other models.

  • Blackmist@feddit.uk
    link
    fedilink
    English
    arrow-up
    2
    ·
    8 days ago

    Thank fuck. Can we have cheaper graphics cards again please?

    I’m sure a RTX 4090 is very impressive, but it’s not £1800 impressive.

      • Blackmist@feddit.uk
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        7 days ago

        I just don’t get whey they’re so desperate to cripple the low end cards.

        Like I’m sure the low RAM and speed is fine at 1080p, but my brother in Christ it is 2024. 4K displays have been standard for a decade. I’m not sure when PC gamers went from “behold thine might from thou potato boxes” to “I guess I’ll play at 1080p with upscaling if I can have a nice reflection”.

        • Tywèle [she|her]@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          7 days ago

          4k displays are not at all standard and certainly not for a decade. 1440p is. And it hasn’t been that long since the market share of 1440p overtook that of 1080p according to the Steam Hardware survey IIRC.

          • Blackmist@feddit.uk
            link
            fedilink
            English
            arrow-up
            0
            arrow-down
            1
            ·
            7 days ago

            Maybe not monitors, but certainly they are standard for TVs (which are now just monitors with Android TV and a tuner built in).

            • Tywèle [she|her]@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              1
              ·
              7 days ago

              That doesn’t really matter if people on PC don’t game on it, does it?

              These are the primary display resolutions from the Steam Hardware Survey.

  • Someplaceunknown@fedia.io
    link
    fedilink
    arrow-up
    2
    ·
    9 days ago

    “LLMs such as they are, will become a commodity; price wars will keep revenue low. Given the cost of chips, profits will be elusive,” Marcus predicts. “When everyone realizes this, the financial bubble may burst quickly.”

    Please let this happen

  • LavenderDay3544@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    AI was 99% a fad. Besides OpenAI and Nvidia, none of the other corporations bullshitting about AI have made anything remotely useful using it.

  • acargitz@lemmy.ca
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    It’s so funny how all this is only a problem within a capitalist frame of reference.

    • masquenox@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 days ago

      What they call “AI” is only “intelligent” within a capitalist frame of reference, too.

      • Hazor@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 days ago

        I don’t understand why you’re being downvoted. Current “AI” based on LLM’s have no capacity for understanding of the knowledge they contain (hence all the “hallucinations”), and thus possess no meaningful intelligence. To call it intelligent is purely marketing.

  • KeenFlame@feddit.nu
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    I am so tired of the ai hype and hate. Please give me my gen art interest back please just make it obscure again to program art I beg of you

  • j4p@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    Sigh I hope LLMs get dropped from the AI bandwagon because I do think they have some really cool use cases and love just running my little local models. Cut government spending like a madman, write the next great American novel, or eliminate actual jobs are not those use cases.

  • dejected_warp_core@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    7 days ago

    Welcome to the top of the sigmoid curve.

    If you were wondering what 1999 felt like WRT to the internet, well, here we are. The Matrix was still fresh in everyone’s mind and a lot of online tech innovation kinda plateaued, followed by some “market adjustments.”

    • Hackworth@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      arrow-down
      1
      ·
      edit-2
      7 days ago

      I think it’s more likely a compound sigmoid (don’t Google that). LLMs are composed of distinct technologies working together. As we’ve reached the inflection point of the scaling for one, we’ve pivoted implementations to get back on track. Notably, context windows are no longer an issue. But the most recent pivot came just this week, allowing for a huge jump in performance. There are more promising stepping stones coming into view. Is the exponential curve just a series of sigmoids stacked too close together? In any case, the article’s correct - just adding more compute to the same exact implementation hasn’t enabled scaling exponentially.

  • randon31415@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    8 days ago

    The hype should go the other way. Instead of bigger and bigger models that do more and more - have smaller models that are just as effective. Get them onto personal computers; get them onto phones; get them onto Arduino minis that cost $20 - and then have those models be as good as the big LLMs and Image gen programs.

    • _NoName_@lemmy.ml
      link
      fedilink
      English
      arrow-up
      0
      arrow-down
      1
      ·
      8 days ago

      That would be innovation, which I’m convinced no company can do anymore.

      It feels like I learn that one of our modern innovations was already thought up and written down into a book in the 1950s, and just wasn’t possible at that time due to some limitation in memory, precision, or some other metric. All we did was do 5 decades of marginal improvement to get to it, while not innovating much at all.

  • Semi-Hemi-Lemmygod@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    9 days ago

    I wish just once we could have some kind of tech innovation without a bunch of douchebag techbros thinking it’s going to solve all the world’s problems with no side effects while they get super rich off it.

    • ohwhatfollyisman@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      9 days ago

      … bunch of douchebag techbros thinking it’s going to solve all the world’s problems with no side effects…

      one doesn’t imagine any of them even remotely thinks a technological panacaea is feasible.

      … while they get super rich off it.

      because they’re only focusing on this.