• skuzz@discuss.tchncs.de
    link
    fedilink
    arrow-up
    1
    ·
    1 month ago

    It will be once the bubble pops. Small local tuned models for specific tasks that the user powers are much less expensive for the tech companies than tech companies powering and watering datacenters.

    Right now the tech bros genuinely think people will be cool paying hundreds of dollars a month to rent a GPU for all their Internet tasks. AI fatigue is already setting in.

    The tech bros’ investors will pull funding once they realize how asinine that is long-term. Probably already starting to, with the likes of Zuck trying to use green charity money to fund his LLMs.

  • TrippinMallard@lemmy.ml
    link
    fedilink
    arrow-up
    1
    ·
    1 month ago

    OpenAI/Anthropic is incentivized to prevent this.

    They are also big enough and unregulated enough that they could use their power & political/industry relationships to drive up the price of local AI ownership (RAM, GPUs, etc)

    • orc girly@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      1 month ago

      I’m not sure of how much they can actually prevent us from just running foss Chinese alternatives locally though

      • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
        link
        fedilink
        arrow-up
        2
        ·
        1 month ago

        Exactly, and a lot of big companies in US are heavily reliant on Chinese models already. For example, Airbnb uses Qwen cause they can self host it and customize it. Cursor built their latest composer model on top of Kimi, and so on. There are far more companies using these tools than making them, so while open models hurt companies that want to sell them as a service, they’re lowering the cost for everyone else.

  • FEIN@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    1 month ago

    That would be preferable. If ML optimization open sources and progresses greatly that would be good for the little guy

  • brucethemoose@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    1 month ago

    Also, if y’all are interested, run local models!

    It’s not theoretical.

    The cost of hybrid inference is very low; You can squeeze Qwen 35B on a 16GB RAM machine as long as it has some GPU. Check out ik_llama.cpp and ubergarm’s quants in particular:

    https://huggingface.co/ubergarm/models#repos

    But if you aren’t willing to even try, I think that’s another bad omen for local models. Like the Fediverse, it won’t be served to you on a silver platter, you gotta go out and find it.

    • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
      link
      fedilink
      arrow-up
      1
      ·
      1 month ago

      It’s really unfortunate how a lot of people have a knee jerk reaction towards anything LLM related right now. While you can make good arguments for avoiding proprietary models offered as a service, there’s really no rational reason to shun open models. If anything, it’s important to develop them into a viable alternative to corporate offerings.

      • orc girly@lemmy.ml
        link
        fedilink
        arrow-up
        1
        ·
        1 month ago

        I think it’s an extension of people only conceiving these things within capitalism (although they might call it techno feudalism or some shit), I remember the phrase “if you aren’t paying for something you’re the product” and thinking that so many people don’t realize we already have things that fall outside of that like so much of the FOSS ecosystem including Linux. It doesn’t help that this kind of messaging is so amplified by liberals on social media who refuse to see the real cause behind our current issues with AI and instead focusing on idealism.

    • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
      link
      fedilink
      arrow-up
      2
      ·
      1 month ago

      Do elaborate. The tech industry has gone through many cycles of going from mainframe to personal computer over the years. As new tech appears, it requires a huge amount of computing power to run initially. But over time people figure out how to optimize it, hardware matures, and it becomes possible to run this stuff locally. I don’t see why this tech should be any different.