OUR SECTORS
At European Tech Recruit, our sectors cover a wide range of industries within the field of technology.
tech jobs in the US?
Looking for
tech jobs in the US?
At European Recruitment, our sectors cover a wide range of industries within the field of technology
At European Recruitment, our sectors cover a wide
range of industries within the field of technology
At European Recruitment, our sectors cover a wide
range of industries within the field of technology
Client services
Learn about the range of client services we offer at European Tech Recruit, and browse through our case sudies.
tech jobs in the US?
Looking for
tech jobs in the US?
At European Recruitment, our sectors cover a wide range of industries within the field of technology
About us
Learn about European Tech Recruit's mission, values, our team, and our commitment to DE&I.
tech jobs in the US?
Looking for
tech jobs in the US?
At European Recruitment, our sectors cover a wide range of industries within the field of technology
Systems Research Engineer
Systems Research Engineer – AI Infrastructure & Distributed Systems
Department
Systems Infrastructure Research
Job Vision
As large language models reshape the foundational software stack, next-generation AI-native infrastructure is redefining how large-scale models are trained, served, and deployed.
An advanced systems research center in the UK is driving innovation in AI infrastructure and agent-oriented serving architectures, helping to define future large-scale data center and distributed AI systems. Positioned at the intersection of cutting-edge systems research and industrial-scale engineering, the team translates novel architectural concepts into deployable technologies.
We are seeking Systems Research Engineers with strong interest in computer systems, distributed AI infrastructure, and performance optimization. These roles are well suited to recent PhD graduates or outstanding BSc/MSc engineers aiming to develop research-driven engineering expertise in operating systems, distributed systems, AI model serving, and machine learning infrastructure.
Key Responsibilities
Distributed Systems Research & Development
-
Architect, implement, and evaluate distributed system components for emerging AI and data-intensive workloads.
-
Design modular and scalable infrastructure spanning heterogeneous clusters (CPU, GPU, accelerators).
-
Develop efficient serving and scheduling systems optimized for large-scale AI workloads.
Performance Optimization & Profiling
-
Conduct deep profiling and performance tuning of large-scale inference and data pipelines.
-
Optimize key-value cache management and heterogeneous memory scheduling.
-
Improve high-throughput inference serving using modern distributed ML frameworks.
-
Apply systematic performance analysis methodologies to identify bottlenecks and scalability constraints.
Scalable Model Serving Infrastructure
-
Develop frameworks enabling multi-tenant, low-latency, and fault-tolerant AI serving across distributed environments.
-
Research techniques for:
-
Cache sharing
-
Data locality optimization
-
Resource orchestration
-
Cluster-level scheduling
-
-
Prototype and evaluate new serving and inference architectures.
Research & Publications
-
Translate novel system designs into publishable research contributions at leading systems and ML venues.
-
Drive internal adoption of innovative methods and architectural improvements.
Cross-Team Collaboration
-
Communicate technical insights and evaluation results clearly to multidisciplinary engineering and research teams.
-
Collaborate across global research groups to align on long-term infrastructure strategy.
Required Qualifications
-
Bachelor’s or Master’s degree in Computer Science, Electrical Engineering, or related field.
-
Strong knowledge of:
-
Distributed systems
-
Operating systems
-
Machine learning systems
-
AI inference serving infrastructure
-
-
Hands-on experience with LLM serving frameworks and distributed cache optimization.
-
Proficiency in C/C++ for systems development.
-
Experience using Python for research prototyping.
-
Solid understanding of distributed algorithms and systems research methodology.
-
Familiarity with profiling and performance analysis tools.
-
Strong communication skills and collaborative mindset.
Preferred Qualifications
-
PhD in systems, distributed computing, or large-scale AI infrastructure.
-
Publications in top-tier systems or ML conferences.
-
Experience with:
-
Load balancing
-
State management
-
Fault tolerance
-
Resource scheduling in inference clusters
-
-
Practical experience designing, deploying, or profiling high-performance cloud or AI infrastructure.
Apply Now
By applying to this role, you acknowledge that we may collect, store, and process your personal data on our systems.
For more information, please refer to our
Privacy
Notice