Im using Ollama on my server with the WebUI. It has no GPU so its not quick to reply but not too slow either.

Im thinking about removing the VM as i just dont use it, are there any good uses or integrations into other apps that might convince me to keep it?

    • 1rre@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      3
      ·
      5 months ago

      Yeah even gpt4o couldn’t keep track of encounters, run battles etc. in my case…

      I think if you wanted to do it mechanically consistently you’d probably need to integrate it into a vtt where you give it context and potentially fine-tune it to give quest related summaries & gming rather than just “stuff”

      • Bluesheep@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 months ago

        I don’t know how tech savvy you are, but I’m assuming since your on lemmy it’s pretty good :)

        The way we’ve solved this sort of problem in the office is by using the LLM’s JSON response, and a prompt that essentially keeps a set of JSON objects alongside the actual chat response.

        In the DND example, this would be a set character sheets that get returned every response but only changed when the narrative changes them. More expensive, and needing a larger context window, but reasonably effective.

    • RachelRodent@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      2
      ·
      edit-2
      5 months ago

      the answer is very spesific to ur pc and amount of vram you have availşble to you. But anything lama 3 even 8b models finetuned to DM or write stories should theoritically work. The other reply that reccomends connecting to another program to make sure rules are consistent sounds like a great idea whşch I have not tried. I use silly tavern as the ui whşch has lots of options and shit to mske thşngs wkrk well. I would reccomend goşng şnto the “KoboldAI” discord and askşng şn the support sectşon folk there are very helpfull sorry for not beşng able to gşve a strsight answer Also boost the context size way up that shit makes dşfference I habe like 16k or sumthin. good luck!

      • RandomLegend [He/Him]@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        5 months ago

        What on earth is going on with your keyboad?!

        Besides that, i have 20GB of VRAM and 64GB or RAM. I can run the mixtral 8x7b model relatively usable. Currently i use oobabooga the most.

        • RachelRodent@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          edit-2
          5 months ago

          I type very poorly on my phone. with that much vram ypu csn get somethşng lşke a 70b model defineyly ask around in the koboldai community that shşt’s crszy