RimWorld

RimWorld

EchoColony
LM Studio setup?
Good day, I am trying to get Echo Colony set up with my local LM Studio.

I have multiple models from gemma to Deepseek but I am unable to connect it successfully to Echo Colony.

Settings
Local Endpoint: http://localhost:1234/v1/embeddings
Name Model: gemma-3-4b-it-qat

The above setup seems to connect to LM Studio, load the model (even if it is already loaded), and then fails. The in-game chat then returns "Error contacting local model: HTTP/1.1 4040 not found"

I have tried every listed connection from HTTP:/127.0.0.1:1234 to http://localhost:1234/v1/embeddings

The http://localhost:1234/v1/embeddings seems to connect and almost work but might be bugged.

Can you provide any guidance as to how to connect to LM Studio? Has anyone gotten it working?

Thank you!
< >
Showing 1-15 of 20 comments
Gerik Uylerk  [developer] Apr 19 @ 11:25pm 
Hey there!
Great news for everyone who had issues with LM Studio! 🎉
After installing the model and dedicating quite a few hours to research and testing, I finally got it working with my mod.

🔧 I’ve added a new setting where you can select LM Studio as your local language model. Once selected, EchoColony will adapt automatically, and your colonists will respond properly 💬.

⚠️ Important tip:
I highly recommend setting the token limit to at least 5500, and ideally around 6000, if possible.
This ensures the colonist has enough context to reply coherently, based on everything they know and what surrounds them.

👉 Let me know if it’s working for you now — or if there’s anything else I should fix!
And here’s the exact endpoint/model I used, just in case:
📸 https://imgur.com/4J6N0TY

Thanks for your support! 🫶
SandFox Apr 20 @ 12:14pm 
I got an error http/1.1 400.
The settings are the same as the picture you uploaded.
I downloaded Meta-Llama-3.1-8B-Instruct-GGUF and applied it right away.
Is there anything I need to change in the settings??
Gerik Uylerk  [developer] Apr 20 @ 8:00pm 
Sorry for the inconvenience. The issue was that the mod was sending too much information, which caused the model to become overloaded. I’ve created a lighter version so the colonist can now respond and hold a conversation with all the essential context. Please let me know how the experience goes. I’m testing the model’s capabilities, so if it responds well, I might start giving colonists more context again. Thanks for your support and patience.
James009 Apr 21 @ 12:13am 
Originally posted by Gerik Uylerk:
Hey there!
Great news for everyone who had issues with LM Studio! 🎉
After installing the model and dedicating quite a few hours to research and testing, I finally got it working with my mod.

🔧 I’ve added a new setting where you can select LM Studio as your local language model. Once selected, EchoColony will adapt automatically, and your colonists will respond properly 💬.

⚠️ Important tip:
I highly recommend setting the token limit to at least 5500, and ideally around 6000, if possible.
This ensures the colonist has enough context to reply coherently, based on everything they know and what surrounds them.

👉 Let me know if it’s working for you now — or if there’s anything else I should fix!
And here’s the exact endpoint/model I used, just in case:
📸 https://imgur.com/4J6N0TY

Thanks for your support! 🫶
I've been using this model with 7000 tokens and it's working brilliantly, quickly, the formatting is like 95% perfect, and the colonists are like 95% aware of everything.

Thank you for all your work, this is mind blowing!
SandFox Apr 21 @ 1:32am 
It works well! But it is so different from when I used "gemini".
My speech style has changed a lot, and most of all, my language (Korean) expression skills have gotten worse.
Is there a good solution??
If you have a good model or setting method, please let me know.

Not sure, but..
I think you also need to set the "system prompt" of the model.
How should I write it so that it reacts like gemini?
Last edited by SandFox; Apr 21 @ 3:07am
Gerik Uylerk  [developer] Apr 21 @ 8:31am 
Hey! Thanks for your feedback — and I'm really sorry about the drop in quality.
While I was calibrating the mod for local models, I ended up removing too much context from the colonist's prompt. That definitely affected the speech style and language quality. My bad!

I've just improved the prompt structure for LM Studio models — the colonist should now be much more aware of their situation, emotions, tasks, and surroundings, which should help restore a more natural conversation style.

If this version works better, we can continue refining it together! I plan to gradually reintroduce even more context while staying under each model’s token limits.

It's a bit tricky for me to get the balance right since I'm not super familiar with LM Studio or other local models yet — but I’m fully committed to making this work great for everyone, including Korean speakers
there is lumimaid-magnum-v4-12b or something alike if someone tries to find more uncensored version.
it kinda works better than standard lama but maybe need some tuning
Originally posted by Gerik Uylerk:
Hey! Thanks for your feedback — and I'm really sorry about the drop in quality.
While I was calibrating the mod for local models, I ended up removing too much context from the colonist's prompt. That definitely affected the speech style and language quality. My bad!

I've just improved the prompt structure for LM Studio models — the colonist should now be much more aware of their situation, emotions, tasks, and surroundings, which should help restore a more natural conversation style.

If this version works better, we can continue refining it together! I plan to gradually reintroduce even more context while staying under each model’s token limits.

It's a bit tricky for me to get the balance right since I'm not super familiar with LM Studio or other local models yet — but I’m fully committed to making this work great for everyone, including Korean speakers


Sir, thank you so much for this mode — it's sooo awesome!

I just tested it with neuraldaredevil-8b-abliterated in LM Studio, and it works like a charm. I don’t even know what else to add — it’s so good!

https://i.imgur.com/2IDw6b0.png
Last edited by /dev/null; Apr 22 @ 10:59am
Auster Apr 25 @ 4:06pm 
lumimaid-magnum-v4-12 by quantfactory is more NSFW, it works but the response almost takes 10 seconds for a reply. I will say that they remember their history when i used the cloud gemini.

neuraldaredevil-8b-abliterated works well. Thanks for everyone for their comments on the modules, i was installing kobold but had no idea how it worked.

<----not a programmer. I couldn't even program my way out of a paper bag nor the bag itself. xD

@Gerik Uylerk
Thank you for having LM Studio added to this mod list. I wanted something mostly offline. The cloud works well, however gets very heavy handed when you try and roll play. One day it is fine the other it is a nono.

Edit 1: Tried another version of lumimaid-magnum-v4-12. Seemed to work then crashed.

neuraldaredevil-8b-abliterated seems to still work well. Responses seem wordier then Google Gemini API. I think you can do more with it however.
Last edited by Auster; Apr 25 @ 5:24pm
Acorn Apr 26 @ 6:35pm 
For the life me I can't figure this out, I keep getting Error contacting local model: HTTP/1.1 400 bad request on LM studio and I've not used it before so idk. Gemini is the only one that works perfectly for me even if it can't really do NSFW. I'd need a guide or something.

Keep up your great work with this mod though!
Auster Apr 27 @ 7:47am 
@Acorn
The mod dev posted an image with the info, however for ease of access

For the first text box labled Local Model Endpoint: http://localhost:1234/v1/chat/completions

Name model: (use whatever one you want)

Once you do find the model you want you need to put the name in the name model text box.

On the most lefthand side there are icons, speech, Dev mode, my models then the search icon. When chats is highlighted you can see the settings icon to the left of your loaded model. The fist thing you see is Context Length. This is what you wish to increase. You will need to load the model and reset this every time you restart LM Studio. In Dev mode you will see whatever model is ready, you can use the copy link to copy the LMM that is already loaded and paste this information in the Name Model.

I use neuraldaredevil-8b-abliterated. The AI can be directed ingame but it seems your FIRST interactions mold the pawn for the rest of the game. Some pawns end up almost poetic and it must be guided to the choice you want.

To everyone else, if you have a fairly strong/robust pc setup you can run this. CPU isn't really used much, nor is ram (~2-3 GB). TBH for things of this nature I do recommend 32gb ram(2x16) for like 50 bucks. GPU is highly used, more so after you have been RPing heavily for days(RL days not ingame), you will hear your gpu fans, however this could also be from switching from the Gemini Cloud to LM Studio. There are other models that are not as powerful.

@Gerik Uylerk

I usually never go this in depth with any guide. Thank you SOOOOOO much for including LM Studio, could not get that working for the life of me. And please feel free to use this miniguide or whatever you call it if it helps anyone else out.
Last edited by Auster; Apr 27 @ 7:56am
Gerik Uylerk  [developer] Apr 27 @ 9:27am 
Thank you very much for the support and the detailed guidance!
Auster Apr 27 @ 9:45am 
@Gerik Uylerk
You are a wizard programmer, I am not. I tried using KoboldAI but could not get it to work, LM studio is much easier to setup. I have noticed that when the AI is active Rimworld's ram almost doubles. I understand why.

As a note with your most recent update, it erased my past conversations with my pawns. I just wanted to let you know and put a warning if people update midgame. I know the high risk but wanted to check out the update.

I usually add comments to other mods. I really do not add much to discussions but your mod shows real progress.

As a possible suggestion, I might suggest adding another Discussions thread possibly with PC specs/general specs if others wish to use LM Studio or KoboldAI. I know mileage will very between PCS ect......
Acorn Apr 27 @ 12:58pm 
Thanks! I'll try again but I definitely think there could be another discussion thread as well and include pictures for a type by type sort of thing for LM Studio.

Edit - I got it working! It just takes a few minutes for the colonist to respond. Althought my colonist mentions Bella and April. Bella was a dino that died a long time ago that was bonded to Dove so that's a little confusing xD And lord they give you a VERY wordy response.

https://gyazo.com/b0fbc780285f8831c9d5d4b47b986282 https://gyazo.com/d53242b5cd94663f8dc8869f73b62c4a

Hopefully if you chat with a colonist long enough they'll actually start remembering things - I ended up using neuraldaredevil-8b-abliterated as well and guess I'll test it out some more to figure out if there's a way to get a faster reply? LM Studio seems good so far, Gemini is likely the best right now with how the colonists react and not being overly wordy.

Edit - a hour later it stopped working again and giving the same error xD
Last edited by Acorn; Apr 27 @ 2:54pm
Auster Apr 27 @ 5:21pm 
@Acorn
If you mean Gemini, give it some time before non-stop chatting ect.... I did the same thing and got the error, since you are prolly using the free version there is a limit, and if you think of it since it is probably later in the day more people are on using Gemini as well.
< >
Showing 1-15 of 20 comments
Per page: 1530 50