Artificial intelligence scaled on a bet - that bigger models, more identical chips, and more data would keep delivering. As problems grow more complex and the requirements of intelligence more diverse, that bet is breaking down. The next era belongs to heterogeneous intelligence: diverse models on diverse chips, each with distinct strengths, co-evolving into systems of capability unreachable by any single model or accelerator.
Callosum is the Intelligent Systems company. We built the infrastructure to make that possible. Our co-evolution engine optimises simultaneously across workflows, agents, and silicon. We launched in early 2026 showing orders of magnitude improvements in performance and a shift in the cost-performance frontier that no single chip or model provider can provide.
We believe intelligence comes from the system, not the model.
We are scientists and engineers solving what others consider impossible. If you thrive on hard problems, and are passionate and energised by the scale of the challenge, we'd love to hear from you.
Heterogeneous intelligence runs on clusters of diverse silicon. You will own the infrastructure layer underneath Callosum's inference platform, ensuring heterogeneous clusters remain operational and observable. You will set up cluster provisioning and lifecycle management, high-performance networking and storage, systems-level performance work to the platform inside its latency and throughput envelopes as the hardware beneath it changes. You will work closely with the hardware and orchestration teams that bring new accelerators into the co-evolution engine. When new silicon arrives, you are one of the people who makes it usable.
Who We Are Looking For
Our teams vary across expertise but are united in high agency, and a track record of driving hard problems from ambiguous to solved.
Strong systems engineering background with deep Linux and networking expertise
Proven debugging, profiling, and performance optimization skills
Experience operating distributed systems and infrastructure platforms
Strong programming skills in Go, Python, C++, Rust, or similar
Comfortable working across hardware/software boundaries
What Stands Out
Operating large-scale Linux, HPC, or GPU clusters
Container orchestration, cluster management, virtualization and provisioning
High-performance networking and storage systems (RDMA, InfiniBand, RoCE)
AI systems performance optimization, especially distributed inference, memory efficiency, and networking performance (CUDA, NCCL)
Competitive Salary, determined by skills and experience
Equity & Ownership
Private healthcare
We offer Visa sponsorship and relocation benefits to hire the best in the world
We work in person at our London office. You'll have the tools, space and setup to do your best work, and if you have specific needs, just tell us
We're committed to building an inclusive workplace where everyone feels welcome, and believe in equal opportunities for all.