Systems Research Engineer
SKILLS
FULL DESCRIPTION
[Employer hidden — view at passion-project.co.uk] is seeking Systems Research Engineers with a strong interest in computer systems, distributed AI infrastructure, and performance optimization. The role involves architecting, implementing, and evaluating distributed system components for AI and data-centric workloads, performance optimization, and research and publications.
Systems Research Engineer
Job Vision In an era where LLM are rebuilding the foundational software stack, [Employer hidden]’s CloudMatrix super-node clusters and AI-native infrastructure are reshaping how large-scale models are trained, served, and deployed. The Edinburgh Research Centre plays a key role in this transformation, driving new AI Infra & Agentic Serving architectures and helping define [Employer hidden]’s next-generation large-scale data centre and AI infrastructure systems. Positioned at the intersection of advanced systems research and industrial-scale engineering, our team turns innovative system designs into deployable, real-world technologies. We are seeking Systems Research Engineers with a strong interest in computer systems, distributed AI infrastructure, and performance optimization. These roles are ideal for recent PhD graduates or exceptional BSc/MSc engineers looking to build research-driven engineering experience in areas such as operating systems, distributed systems, AI model serving, and machine learning infrastructure. You will work closely with senior architects on real-world projects, helping to prototype and optimize next-generation AI infrastructure.
Key Responsibilities
- · Distributed Systems Research & Development: Architect, implement, and evaluate distributed system components for emerging AI and data-centric workloads. Drive modular design and scalability across CPU, GPU, and NPU clusters, building highly efficient serving and scheduling systems.
- · Performance Optimization & Profiling: Conduct in-depth profiling and performance tuning of large-scale inference and data pipelines, focusing on KV cache management, heterogeneous memory scheduling, and high-throughput inference serving using frameworks like vLLM, Ray Serve, and modern PyTorch Distributed systems.
- · Scalable Model Serving Infrastructure: Develop and evaluate frameworks that enable efficient multi-tenant, low-latency, and fault-tolerant AI serving across distributed environments. Research and prototype new techniques for cache sharing, data locality, and resource orchestration and scheduling within AI clusters.
- · Research & Publications:Translate innovative research ideas into publishable contributions at leading venues (e.g., OSDI, NSDI, EuroSys, SoCC, MLSys, NeurIPS, ICML, ICLR) while driving internal adoption of novel methods and architectures.
- · Cross-Team Collaboration: Communicate technical insights, research progress, and evaluation outcomes effectively to multidisciplinary stakeholders and global [Employer hidden] research teams.
Person Specification
Required Qualifications and Skills:
- · Bachelor’s or Master’s degree in Computer Science, Electrical Engineering, or related field.
- · Strong knowledge of distributed systems, operating systems, machine learning systems architecture, Inference serving, and AI Infrastructure.
- · Hands-on experience with LLM serving frameworks (e.g., vLLM, Ray Serve, TensorRT-LLM, TGI) and distributed KV cache optimization.
- · Proficiency in C/C++, with additional experience in Python for research prototyping.
- · Solid grounding in systems research methodology, distributed algorithms, and profiling tools.
- · Team-oriented mindset with effective technical communication skills.
Desired Qualifications and Experience:
- · PhD in systems, distributed computing, or large-scale AI infrastructure.
- · Publications in top-tier systems or ML conferences (NSDI, OSDI, EuroSys, SoCC, MLSys, NeurIPS, ICML, ICLR).
- · Understanding of load balancing, state management, fault tolerance, and resource scheduling in large-scale AI inference clusters.
- · Prior experience designing, deploying, and profiling high-performance cloud or AI infrastructure systems.