Report bugs here. Ideally, your report should include the steps necessary to reproduce the bug.
Three Eyes Software
Creator of
Recent community posts
Currently a sort of game master reads the conversation and translates RP actions into game actions.
You can use (parentheses) after your message for OOC comments, but the NPC will probably be confused because, barring a few exceptions, the NPC itself isn't informed of what game actions it can take and what game actions have been taken.
NPCs can't place status effects on you as of now.
The game communicates with the locally hosted KoboldCPP server in the same way it would with the demo server or OpenRouter.
To me it seems like the backend is failing in some way without the game noticing, and then the request errors out.
KoboldCPP outputs are currently sent to /dev/null on Linux because I had some worries about compatibility with different distros, so there's no way to see why this is happening either. Since it's model specific, maybe the download got corrupted somehow. Try deleting the Gemma-4-Sparse .gguf and .meta files in Silverpine 1.6.6b\Silverpine_Data\StreamingAssets to redownload it.
Another thing it could be is that Gemma 4 uses SWA, which might be an issue with AMD somehow?
If anything, I will look into adding better logging for the outputs on Linux in 1.7.
Doing this breaks the AI in many ways because the entire pipeline outside of the LLM requires everything to be in English to work properly. For example, the AI won't be able to properly retrieve memories anymore, and the NPC action system stops accurately translating RP actions to game actions. The LLM will also be dumber in general because it went through English RL training. I specifically added a language check to the first user input to deter people from doing what you describe.
Try recording a video of the command prompt and getting a screenshot of the error that way. Sadly the game can't capture the error by itself because of how Windows handles capturing process outputs.
One reason I think this might be happening is because, at least on Linux, or perhaps specifically the GPU I use for testing, the backend sometimes crashes when loading a large model multiple times. This is fixed by restarting the PC. Other than that, I can only imagine that the model file itself got corrupted while downloading somehow. You can redownload it by deleting the model file in Silverpine 1.6.5d\Silverpine_Data\StreamingAssets.
Basically, if I understand right the model uses 16.7 GB (gigabyte? gibibyte?), and it all has to fit into your VRAM + RAM. The backend uses an algorithm that's not transparent to me to fit a certain amount of that into your VRAM. 6 + 16 GB = 22 GB - the OS and other software using a certain amount - your laptop's iGPU using a certain amount, is making it a really tight fit, if it's even possible.
Like I said, you need 8 GB of VRAM for this model. It might fit if you update to the "c" version I just uploaded, since it uses an updated backend that better squeezes the layers of the models into VRAM. Try closing your web browser too, and make sure the iGPU has as little VRAM as possible assigned to it in the BIOS, since it draws from your normal system RAM which most of the model spills into.

