Staff Machine Learning Performance Engineer, Inference Optimisation
Wayve(2 months ago)
About this role
A Staff/Principal ML Performance Engineer leading high-impact projects to optimise ML inference for edge accelerators and GPUs, enabling Wayve’s first driving product. The role focuses on running large transformer-based models efficiently on low-cost, low-power edge devices and operates at the intersection of ML compilers, kernels, and ML engineering. It involves building technical roadmaps, collaborating across teams, and contributing to early-stage product development.
Required Skills
- Inference Optimization
- MLIR
- TensorRT
- CUDA
- Qualcomm QNN
- OpenCL
- Triton
- Kernel Development
- Compiler Optimization
- Target Platforms
+4 more
About Wayve
firststage.coFirst is an AI-powered hiring platform that helps companies automate recruiting with intelligent candidate screening, fair assessments, and a strong focus on world-class candidate experience. The product includes AI pipeline management and hiring workflows to surface top talent, reduce bias, and cut down administrative work for recruiters. Designed for industry-leading teams, First integrates quickly and can be deployed in minutes with options tailored to scaling organizations.
Apply instantly with AI
Let ApplyBlast auto-apply to jobs like this for you. Save hours on applications and land your dream job faster.
More jobs at Wayve
Similar Jobs
Head of Inference Kernels
Etched(3 months ago)
Compiler Architect
d-Matrix(1 month ago)
Architecture Intern - Inference
Etched(1 month ago)
Performance Engineer - Inference
Cerebras Systems(13 days ago)
Senior Engineer, AI Systems
Samsung Research America(7 days ago)
ML Engineer, Large Language Models (LLM Training & Inference Optimization)
Nebius(10 months ago)