RimWorld

RimWorld

Local AI Social Interactions
13 Comments
tw_boring Feb 15 @ 3:43am 
Thank so much.:steamhappy:
DealingLace Feb 6 @ 11:30am 
This looks sick! Keen to check out the repo, thanks man!
gavin.blair  [author] Feb 3 @ 9:27am 
@Deankiller Turnupseed haha that might be overkill. it is on github for anyone that wants to take it to the next level
Deankiller Turnupseed Feb 2 @ 11:03pm 
deepseek when
Szczygiel Jan 26 @ 10:17am 
Well its fun, pity
gavin.blair  [author] Jan 26 @ 9:45am 
@Szczygiel thanks for the feedback. it is on github for anyone that wants to do something more with it: https://github.com/gavinblair/SocialInteractions

it was just a fun weekend project i thought people might enjoy. I likely won’t develop it further beyond maybe adding some settings for Ollama.
Szczygiel Jan 25 @ 3:19am 
Not only here to whine, really like the idea. But you can do better and I am sure you will do better, looking forward to that. I have to say I am even excited. And hyped, pretty sure mod will be popular or have adaptations.
First thing that comes to my mind is some kind of restriction, how many convos mod makes into this notification by using LLM. It heavily spams the console, not sure if this won't be a problem for logfile size, idk.
As rimworld player you should know rimworld alone is really... well, not going peacefully with number of mods and performance. This mod adress LLM so outside the game, thus need some optimisation on performance, as LLM alone takes some of it. It could be done - as far as I imagine possible - done by mod options you could only think of.
Szczygiel Jan 25 @ 3:18am 
I will make it one message to avoid clutter
1. something to ask : Does it takes into account only vanilla backrounds and traits (or definied, pointed by mod author)? Or it takes whatever it is in pawn bio? One AI mod that I tried completely ignored things added by other mods sadly.
2. So, this instruction in description on model version, this is example? Or this is the model?
3. Will it work with any other model from ollama, or any other llama?
4. If no, do you plan to add support to newer versions or other variants, like this uncensored one? The newest is 3.3 I believe.
5. How hard it would be for you to add support or somebody to patch it?
6. From very first look, it is really performance heavy. Could be good in scenario where you have 3 colonists, but if you got 33 good luck with that.
7. Messages are barely readable and dissappear too fast.
8. Not only lags game pretty much, but also lagging itself. You will get notifications from past convos on paused game.
gavin.blair  [author] Jan 24 @ 10:55am 
@GRUFOLA I tried that but it became difficult to read and opted to put it at the top. Perhaps it can be an option in the future.
Giorgio Dibitonto Jan 24 @ 10:38am 
This seems amazing!

Could you consider to make it look a little more like "speak up" mod? It use "bubble interaction" mod to make the sentence appear near the pawn as a bubble? It would look a lot better.
gavin.blair  [author] Jan 24 @ 10:37am 
@Captain_Tess you will need to download Ollama from ollama.com . It is free and open source. Once it is running run the command in the instructions and it will download and start up the LLM (it is a 2gb file that downloads). You will need 2gb of ram available. The mod will eventually have settings but it currently only works with the default Ollama settings. You may have to find instructions on starting LLM’s with Ollama on Windows, I only have a Mac, but it should work. Please let me know how you fare.
Captain_Tess Jan 24 @ 9:50am 
This sounds like an amazingly interesting idea, but is there a chance you could provide any addition instructions, will be attempting to use with the provided instructions but I feel a little confused by them. Also depending on how this works the ability to write a prompt in game for each pawn to be sent to the ai, a little player written bio, would be a good idea.
gavin.blair  [author] Jan 24 @ 8:16am 
Please let me know how it works for you. This is my first mod.