Hey :) For a while now I use gpt-oss-20b on my home lab for lightweight coding tasks and some automation. I’m not so up to date with the current self-hosted LLMs and since the model I’m using was released at the beginning of August 2025 (From an LLM development perspective, it feels like an eternity to me) I just wanted to use the collective wisdom of lemmy to maybe replace my model with something better out there.

Edit:

Specs:

GPU: RTX 3060 (12GB vRAM)

RAM: 64 GB

gpt-oss-20b does not fit into the vRAM completely but it partially offloaded and is reasonably fast (enough for me)

  • Jozzo@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    9 hours ago

    It’s done by software using an LLM, not just a raw LLM. They do only work with text, but you can get it to output the text “get_weather(mylocation)”, and instead of just outputting that directly to the user, the software running on top of the LLM runs a " get_weather" function that calls some weather API. The result of that function is then output to the user.

    Any time you see an “AI” taking “actions”, this is what happens in the background for every action.

    • SuspciousCarrot78@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      7 hours ago

      ^ exactly that.

      Also, I suspect that’s the reason for Claude famously telling everyone to “go to bed” all the time. That bastich cannot run time and date as a background check reliably…it wings it based on start of conversation. Bitch I type a lot and fast…stop tellling me to go to bed at 9pm.

      I expect it will get patched soon.

      An endearing quirk…but it exposes the wiring if you know. Still, doesn’t make the trick any less impressive when it hits.