At the start of this year we installed a high-performance server (512 GB memory, two high-end GPUs). This machine now hosts local large-language models LLMs, meaning the AI runs entirely on‑premise and your data never leaves our network.
What you can do today Access the web UI at chat.science.ru.nl. Log in with your Science RU username (e.g., jdoe, not jdoe@science.ru.nl). The service is powered by Ollama and Open WebUI, providing a clean, responsive chat interface. Early tests show reliable speech‑to‑text (including Dutch) and the ability to review source code. A user experience that feels very close to the larger commercial models, even though we are currently using open‑source models a few generations older. Current status - experimental We don’t know how many concurrent users the system can support and how much persistent storage will be required. No automated backups are in place; please treat the data you store there as temporary. Uptime is not guaranteed. The service may need to be restarted or the server rebooted when we apply configuration changes or updates. The setup will evolve as we gather feedback, so your input is valuable. Which models are available? Browse the full catalog on ollama.com/search. At the moment we run two gpt‑oss models, but many others can be added on request.
...