Your mission
We are looking for a motivated student assistant to support our R&D team in setting up and maintaining large language model (LLM) inference environments and related API services. The role involves hands-on work with modern inference frameworks and GPU-based infrastructures, both cloud-hosted and on-premises.
1. Setting up, configuring, and maintaining LLM inference frameworks such as vLLM, TensorRT-LLM,, Ollama, and SGLang.
2. Deploying and managing API endpoints for model inference on self-hosted GPU servers and cloud GPU instances (e.g., RunPod, Hetzner, AWS).
3. Performing DevOps-related activities such as container setup, port forwarding, reverse proxy configuration, and HTTPS endpoint deployment.
Your profile
* Enrolled student in Computer Science, Electrical Engineering, Data Science or related field.
* Solid knowledge of Linux environments and shell scripting.
* Experience with Docker, Python, and basic networking and SSH concepts (e.g., ports, reverse proxies, secure connections).
* Experience with local LLM serving frameworks such as, vLLM, Ollama, or TensorRT-LLM as well as familiarity with GPU-based computation, including CUDA, driver management, and hardware resource monitoring would be a strong plus.
About us
LUBIS is a fast-growing German startup redefining how the semiconductor industry works. We tackle one of its hardest challenges — ensuring complex chips work flawlessly before
they're built.
Our mission is simple: to transform verification from a craft into a system. By structuring how teams work and automation we make chip design faster, reliable, and bug-free.
LUBIS isn't just improving the process — we're defining how verification is done.