“cool”
Comment on Uses for local AI?
WeLoveCastingSpellz@lemmy.dbzer0.com 3 months ago
playing dnd alone is pretty cool
badcommandorfilename@lemmy.world 3 months ago
Comment on Uses for local AI?
WeLoveCastingSpellz@lemmy.dbzer0.com 3 months ago
playing dnd alone is pretty cool
“cool”
RandomLegend@lemmy.dbzer0.com 3 months ago
Any model recommendation for that?
The ones i tried get stuck in a loop at some point due to the small context windows.
WeLoveCastingSpellz@lemmy.dbzer0.com 3 months ago
the answer is vey spesific to ur pc and amount of vram you have availşble to you. But anything lama 3 even 8b models finetuned to DM or write stotied should theoritically work. The other reply that reccomends connecting to another program to make sure rules are consistent dounds like a great idea whşch I have not tried. I use silly tavern as the ui whşch has lots of options and shit to mske thşngs wkrk well. I would reccomend goşng şnto the “KoboldAI” discord and askşng şn the support sectşon folk there are very helpfull sorry for not beşng able to gşve a strsight answer. good luck!
RandomLegend@lemmy.dbzer0.com 3 months ago
What on earth is going on with your keyboad?!
Besides that, i have 20GB of VRAM and 64GB or RAM. I can run the mixtral 8x7b model relatively usable. Currently i use oobabooga the most.
WeLoveCastingSpellz@lemmy.dbzer0.com 3 months ago
I tupe very poorly on my phone. with that much vram ypu csn get somethşng lşke a 70b model defineyly ask around in the koboldai community that shşt’s crszy
1rre@discuss.tchncs.de 3 months ago
Yeah even gpt4o couldn’t keep track of encounters, run battles etc. in my case…
I think if you wanted to do it mechanically consistently you’d probably need to integrate it into a vtt where you give it context and potentially fine-tune it to give quest related summaries & gming rather than just “stuff”
Bluesheep@lemmy.world 3 months ago
I don’t know how tech savvy you are, but I’m assuming since your on lemmy it’s pretty good :)
The way we’ve solved this sort of problem in the office is by using the LLM’s JSON response, and a prompt that essentially keeps a set of JSON objects alongside the actual chat response.
In the DND example, this would be a set character sheets that get returned every response but only changed when the narrative changes them. More expensive, and needing a larger context window, but reasonably effective.
RandomLegend@lemmy.dbzer0.com 3 months ago
VTT integration would be one hell of a job to do.