• SSUPII@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    32
    arrow-down
    3
    ·
    edit-2
    8 hours ago

    Investment is done really to train models for ever more miniscule gains. I feel like the current choices are enough to satisfy who is interested in such services, and what really is lacking is now more hardware dedicated to single user sessions to improve quality of output with the current models.

    But I really want to see more development on offline services, as right now it is really done only by hobbyists and only occasionally large companies with a little dripfeed (Facebook Llama, original Deepseek model [latter being pretty much useless as no one has the hardware to run it]).

    I remember seeing the Samsung Galaxy Fold 7 (“the first AI phone”, unironic cit.) presentation and listening to them talking about all the AI features instead of the real phone capabilities. “All of this is offline, right? A powerful smartphone… makes sense to have local models for tasks.” but it later became abundantly clear it was just repackaged always-online Gemini for the entire presentation on $2000 of hardware.

    • Taldan@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 hours ago

      what really is lacking is now more hardware dedicated to single user sessions to improve quality of output with the current models

      That is the exact opposite of my opinion. They’re throwing tons of computing at the current models. It has produced little improvement. The vast majority of investment is in compute hardware, rather than R&D. They need more R&D to improve the underlying models. More hardware isn’t going to get the significant gains we need

    • mcv@lemmy.zip
      link
      fedilink
      English
      arrow-up
      30
      ·
      12 hours ago

      They’re investing this much because they honestly seem to think they’re on the cusp of super intelligent AGI. They’re not, but they really seem to think they are, and that seems to justify these insane investments.

      But all they’re really doing is the same thing as before but even bigger. It’s not going to work. It’s only going to make things even more expensive.

      I use Copilot and Claude at work, and while it’s really impressive at what it can do, it’s also really stupid and requires a lot of hand holding. It’s not on the brink of AGI super intelligence. Not even close. Maybe we’ll get there some day, but not before all these companies are bankrupt.

    • artyom@piefed.social
      link
      fedilink
      English
      arrow-up
      21
      ·
      12 hours ago

      I knew it was a bubble since Computex January 2024 when Derb8uer showed an “AI PC case”. He asked “What’s AI about this PC case?” and they replied that you could put an AI PC inside it.

      • SSUPII@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        7
        ·
        11 hours ago

        You are talking more about the term here being used everywhere out of context.

        • artyom@piefed.social
          link
          fedilink
          English
          arrow-up
          21
          ·
          11 hours ago

          I am talking about companies slapping “AI” on their products and systems and raising their value, in the same way that companies in the 90s slapped “dotcom” on their branding and raised their value.

    • ferrule@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      ·
      8 hours ago

      The problem is there is little continuous cash flow for on prem personal services. Look at Samsung’s home automation, its nearly all online features and when the internet is out you are SOL.

      To have your own Github Copilot in a device the size and power usage of a Raspberry Pi would be amazing. But then they won’t get subscriptions.

    • humanspiral@lemmy.ca
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 hours ago

      more development on offline services

      There is absolutely massive development on open weight models that can be used offline/privately. Minimax M2, most recent one, has comparable benchmark scores to the private US megatech models at 1/12th the cost, and at higher token throughput. Qwen, GLM, deepseek have comparable models to M2, and have smaller models more easily used on very modest hardware.

      Closed megatech datacenter AI strategy is partnership with US government/military for oppressive control of humanity. Spending 12x more per token while empowering big tech/US empire to steal from and oppress you is not worth a small fraction in benchmark/quality improvement.