• SavvyWolf@pawb.social
    link
    fedilink
    English
    arrow-up
    63
    arrow-down
    1
    ·
    11 months ago

    Do people actually want this?

    Like, I know the megacorps that control our lives do (since it’s a cheap way of adding value to their products), but what about actual users? I think many see it as a novelty and a toy rather than a productivity tool. Especially when public awareness of “hallucinations” and the plight faced by artists rises.

    Kinda feels like the whole “voice controlled assistants” bubble that happened a while ago. Sure they are relatively commonplace nowadays, but nowhere near as universal as people thought they would be.

      • EvilMonkeySlayer@kbin.social
        link
        fedilink
        arrow-up
        12
        ·
        11 months ago

        I think it’s those stupid hard coded buttons on my remote that I accidentally press every so often then have to repeatedly try and back/exit out of the stupid thing it launched that I cannot remove/uninstall from my tv.

    • Awhiskeydrunker@kbin.social
      link
      fedilink
      arrow-up
      14
      ·
      11 months ago

      Maybe I’m a pessimist but this is going to really resonate with the people who are “looking forward to AI” because they read headlines, but haven’t actually used any LLMs yet because nobody has told them how.

    • Uranium3006@kbin.social
      link
      fedilink
      arrow-up
      6
      ·
      11 months ago

      I want a voice controlled assistant that runs locally and is fully FOSS and I can just run on my bog standard linux PC, hardware minimum requirements nonwithstanding

    • coolin@beehaw.org
      link
      fedilink
      arrow-up
      3
      ·
      11 months ago

      Current LLMs are manifestly different from Cortana (🤢) because they are actually somewhat intelligent. Microsoft’s copilot can do web search and perform basic tasks on the computer, and because of their exclusive contract with OpenAI they’re gonna have access to more advanced versions of GPT which will be able to do more high level control and automation on the desktop. It will 100% be useful for users to have this available, and I expect even Linux desktops will eventually add local LLM support (once consumer compute and the tech matures). It is not just glorified auto complete, it is actually fairly correlated with outputs of real human language cognition.

      The main issue for me is that they get all the data you input and mine it for better models without your explicit consent. This isn’t an area where open source can catch up without significant capital in favor of it, so we have to hope Meta, Mistral and government funded projects give us what we need to have a competitor.

      • chicken@lemmy.dbzer0.com
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        11 months ago

        A year ago local LLM was just not there, but the stuff you can run now with 8gb vram is pretty amazing, if not quite as good yet as GPT 4. Honestly even if it stops right where it is, it’s still powerful enough to be a foundation for a more accessible and efficient way to interface with computers.

  • const_void@lemmy.ml
    link
    fedilink
    arrow-up
    52
    ·
    edit-2
    11 months ago

    This is the dumbest fucking thing I’ve ever heard of. I’m not buying any keyboard or laptop that has this key. There’s enough Linux-first vendors these days that it’s easy to avoid (Framework, System76, Tuxedo, etc). It’s time to be done with Lenovo and Dell.

    • palordrolap@kbin.social
      link
      fedilink
      arrow-up
      16
      arrow-down
      2
      ·
      edit-2
      11 months ago

      This is the dumbest fucking thing I’ve ever heard of. I’m not buying any keyboard or laptop that has this key.

      Which is exactly what people said about the Windows key.

      Now it’s all but impossible to buy a keyboard that doesn’t have it. Worse, most of us use it without thinking.

      Sure you can call it Super if you like, and even have a Tux key-cap on it, but there used to be a literal gap between the Alt keys and their Ctrl brethren in the lateral directions away from the space bar, and those days are long gone.

      There’ll be the niche users who stick with old keyboards without this new key, just like there are the die-hards who have stuck resolutely to the old IBM keyboards and the like from pre-1995, but if you want a new keyboard?

      Gonna have to shell out a small fortune for a custom build or make do with that dumb new key.

      (Shoutout to the Context Menu key which went as unmentioned in the above as it goes unused in day to day use, despite having been included with its Super cousin since day one.)

      • brax@sh.itjust.works
        link
        fedilink
        arrow-up
        10
        ·
        11 months ago

        I don’t see an issue with a “super” key. But what would a copilot key bring that’s of any value? The super key already does everything you’d need.

        • Krzd@lemmy.world
          link
          fedilink
          arrow-up
          9
          ·
          11 months ago

          more keys for custom keybinds ¯⁠\⁠_⁠(⁠ツ⁠)⁠_⁠/⁠¯ depending on where it’s located I’ll probably just use it as a microphone toggle

          • brax@sh.itjust.works
            link
            fedilink
            arrow-up
            2
            ·
            11 months ago

            We have so many unused potential binds already, though. Knowing the way tech goes these days, they’ll find a way to hard-code the key to one macrob and that’s it lol

      • const_void@lemmy.ml
        link
        fedilink
        arrow-up
        8
        ·
        11 months ago

        Gonna have to shell out a small fortune for a custom build or make do with that dumb new key.

        I don’t think this is true. Just buy a laptop from a company that ships it with Linux. No Windows, no Windows keys. It doesn’t have to be ‘custom’.

        • Keith@lemmy.zip
          link
          fedilink
          arrow-up
          4
          ·
          11 months ago

          The post mentioned this, and argues that a super a key is basically just a windows key

      • PumaStoleMyBluff@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        11 months ago

        The article actually says the Copilot key will mostly be replacing Menu or Right Control on existing layouts. So if you’re already not using those (or are already re-binding them), it’s just a new keycap.

        • 𝘋𝘪𝘳𝘬@lemmy.ml
          link
          fedilink
          arrow-up
          1
          ·
          11 months ago

          iit’s just a new keycap

          Plus the configuration that is needed to remap the key back to the correct key code.

      • giloronfoo@beehaw.org
        link
        fedilink
        arrow-up
        2
        ·
        11 months ago

        The video made it look like this was the context menu key. This may just be a key cap change for WHQL certification of keyboards.

    • njordomir@lemmy.world
      link
      fedilink
      arrow-up
      9
      ·
      11 months ago

      Same, I think I might give the System76 Darter a try when I eventually have to replace my Xps 9370. It’s bad enough that my computer comes with a windows logo on the super-key and often windows preinstalled. Shipping with a non-ANSI/ISO layout is a no-buy for me.

    • BaldProphet@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      11 months ago

      I fully agree with you, but Framework is definitely not Linux-first. The only OS they offer preloaded on their laptops is Windows. You have to install Linux yourself if you want it.

  • Lvxferre@lemmy.ml
    link
    fedilink
    arrow-up
    44
    ·
    11 months ago

    Oh “great”, more crap between Ctrl and Alt.

    [Grumpy grandpa] In my times, the space row only had five keys! And we did more than those youngsters do with eight, now nine keys!

  • Apollo2323@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    40
    ·
    11 months ago

    So you can pressed accidentally activating the fucking AI and make the numbers go up so Microsoft can then go and say to investors look millions are using my AI. So annoying.

    • floofloof@lemmy.ca
      link
      fedilink
      English
      arrow-up
      29
      ·
      11 months ago

      They’ve learned from their mistakes, and concluded that Clippy failed because there was no Clippy key.

  • Eager Eagle@lemmy.world
    link
    fedilink
    English
    arrow-up
    27
    ·
    11 months ago

    this kind of shit is what gives AI a bad rep

    no one needs this

    almost no one wants it

    and they’ll kill it in a couple of years like they did it with Cortana

  • TommySoda@lemmy.world
    link
    fedilink
    arrow-up
    23
    arrow-down
    1
    ·
    edit-2
    11 months ago

    I have nothing against the people that are working on AI and appreciate the work they do. However every time I see an article about a company using AI like this I just get the vibe that it’s a bunch of middle aged men trying desperately to make things like the “future” they saw when they were a kid. I’ve seen amazing implementations of AI in a lot of different ways but I’m so sick of dumb ideas like this because some guy that used to watch Star Trek as a kid wants to feel like they live in the future while piggybacking on someone else’s work. It’s like the painted tunnel in cartoons where it looks like a real tunnel but in reality it’s just a very convincing lie. And that’s all that it is. Complexity does not mean sophistication when it comes to AI and never has and to treat it as such is just a forceful way to make your ideas come true without putting in the real effort.

    Sorry, I had to get that out. Also I have nothing against Star Trek and I used to watch it as a kid because my parents watched it all the time.

    • Lvxferre@lemmy.ml
      link
      fedilink
      arrow-up
      3
      arrow-down
      2
      ·
      11 months ago

      Complexity does not mean sophistication when it comes to AI and never has and to treat it as such is just a forceful way to make your ideas come true without putting in the real effort.

      It’s a bit off-topic, but what I really want is a language model that assigns semantic values to the tokens, and handles those values instead of directly working with the tokens themselves. That would be probably far less complex than current state-of-art LLMs, but way more sophisticated, and require far less data for “training”.

      • njordomir@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        11 months ago

        I’m not sure I understand. Do you mean hearing codewords triggering actions as opposed to trying to understand the users intent through language? Or is are there a few more layers to this whole thing than my moderate nerd cred will allow me to understand?

        • Lvxferre@lemmy.ml
          link
          fedilink
          arrow-up
          4
          ·
          11 months ago

          Not quite. I’m focusing on chatbots like Bard, ChatGPT and the likes, and their technology (LLM, or large language model).

          At the core those LLMs work like this: they pick words, split them into “tokens”, and then perform a few operations on those tokens, across multiple layers. But at the end of the day they still work with the words themselves, not with the meaning being encoded by those words.

          What I want is an LLM that assigns multiple meanings for those words, and performs the operations above on the meaning itself. In other words the LLM would actually understand you, not just chain words.

          • Kogasa@programming.dev
            link
            fedilink
            arrow-up
            5
            ·
            11 months ago

            Semantic embeddings are a thing. LLMs “work with tokens” but they associate them with semantic models internally. You can externalize it via semantic embeddings so that the same semantic models can be shared between LLMs.

            • Lvxferre@lemmy.ml
              link
              fedilink
              arrow-up
              2
              ·
              edit-2
              11 months ago

              The source that I’ve linked mentions semantic embedding; so does further literature on the internet. However, the operations are still being performed with the vectors resulting from the tokens themselves, with said embedding playing a secondary role.

              This is evident for example through excerpts like

              The token embeddings map a token ID to a fixed-size vector with some semantic meaning of the tokens. These brings some interesting properties: similar tokens will have a similar embedding (in other words, calculating the cosine similarity between two embeddings will give us a good idea of how similar the tokens are).

              Emphasis mine. A similar conclusion (that the LLM is still handling the tokens, not their meaning) can be reached by analysing the hallucinations that your typical LLM bot outputs, and asking why that hallu is there.

              What I’m proposing is deeper than that. It’s to use the input tokens (i.e. morphemes) only to retrieve the sememes (units of meaning; further info here) that they’re conveying, then discard the tokens themselves, and perform the operations solely on the sememes. Then for the output you translate the sememes obtained by the transformer into morphemes=tokens again.

              I believe that this would have two big benefits:

              1. The amount of data necessary to “train” the LLM will decrease. Perhaps by orders of magnitude.
              2. A major type of hallucination will go away: self-contradiction (for example: states that A exists, then that A doesn’t exist).

              And it might be an additional layer, but the whole approach is considerably simpler than what’s being done currently - pretending that the tokens themselves have some intrinsic value, then playing whack-a-mole with situations where the token and the contextually assigned value (by the human using the LLM) differ.

              [This could even go deeper, handling a pragmatic layer beyond the tokens/morphemes and the units of meaning/sememes. It would be closer to what @njordomir@lemmy.world understood from my other comment, as it would then deal with the intent of the utterance.]

  • Aurenkin@sh.itjust.works
    link
    fedilink
    arrow-up
    10
    ·
    11 months ago

    Can they just make the copilot shortcut on my taskbar permanently fuck off? It appears erratically and I don’t seem to be able to get rid of it when it’s there.