At the start of this year we installed a high-performance server (512 GB memory, two high-end GPUs). This machine now hosts local large-language models LLMs, meaning the AI runs entirely on‑premise and your data never leaves our network.

What you can do today

  • Access the web UI at chat.science.ru.nl. Log in with your Science RU username (e.g., jdoe, not jdoe@science.ru.nl).
  • The service is powered by Ollama and Open WebUI, providing a clean, responsive chat interface.
  • Early tests show reliable speech‑to‑text (including Dutch) and the ability to review source code. A user experience that feels very close to the larger commercial models, even though we are currently using open‑source models a few generations older.

Current status - experimental

  • We don’t know how many concurrent users the system can support and how much persistent storage will be required.
  • No automated backups are in place; please treat the data you store there as temporary.
  • Uptime is not guaranteed. The service may need to be restarted or the server rebooted when we apply configuration changes or updates.
  • The setup will evolve as we gather feedback, so your input is valuable.

Which models are available?

Browse the full catalog on ollama.com/search. At the moment we run two modes: gpt‑oss and deepseek-r1 , but others can be added on request. We ran deepseek-r1:671b, but this model is too large for the machine we have.

How to help

  • Report bugs, performance issues, or feature requests to C&CZ
  • Share use‑cases that work well – they help us prioritise future improvements.

We look forward to seeing how you put this on‑premise AI capability to work in your projects!