AI/ML Software Engineer
About the Role
As a Senior Software Engineer on the AI Platform team, you'll design and build infrastructure that supports model training, validation, deployment, and serving at scale. You will work with modern AWS-native technologies, focusing on low-latency microservices, automated pipelines, and robust deployment workflows to enable safe and efficient delivery of machine learning models into production.
This role is ideal for someone who enjoys building platforms and tools that simplify complexity for ML and data science teams and thrives in fast-paced environments where engineering excellence and reliability are paramount.
What You'll Do
- Build and maintain scalable systems and infrastructure for deploying and serving ML models.
- Design low-latency, fault-tolerant model inference systems using Amazon SageMaker.
- Implement safe deployment strategies like blue/green deployments and rollbacks.
- Create and manage CI/CD pipelines for ML workflows.
- Monitor model performance and system health using AWS observability tools.
- Develop internal tools and APIs to help ML teams deploy and monitor models easily.
- Collaborate with ML engineers, data scientists, and DevOps to productionize new models.
- Participate in code reviews, system design, and platform roadmap discussions.
- Continuously improve deployment reliability, speed, and usability of the ML platform.
What You Bring
- 4+ years of software engineering experience, with at least 2 years focused on low-latency, highly available backend systems.
- Bachelor's or Master's in Computer Science, Data Science, AI, Machine Learning, or related field.
- Strong fundamentals in data structures, algorithms, and distributed computing.
- Proficiency in Python; familiarity with Go or Rust is a plus.
- Hands-on experience with model serving systems, registries, and pipeline orchestration (preferably SageMaker).
- Solid understanding of MLOps best practices, including versioning, testing, deployment, and reproducibility.
- Experience building and maintaining CI/CD pipelines for ML workflows.
- Familiarity with ML frameworks like TensorFlow, PyTorch, or Scikit-learn.
- Experience with SQL/NoSQL databases or data warehouses like Snowflake or Redshift.
Preferred Qualifications
- Experience building internal ML platform services or self-service tooling for model deployment and monitoring.
- Knowledge of model optimization techniques (TorchScript, ONNX, quantization, batching).
- Experience with feature stores, real-time feature serving, or caching systems for ML workloads.
- Background in deploying ML models into high-availability, mission-critical environments.
FAQs
Congratulations, we understand that taking the time to apply is a big step. When you apply, your details go directly to the consultant who is sourcing talent. Due to demand, we may not get back to all applicants that have applied. However, we always keep your CV and details on file so when we see similar roles or see skillsets that drive growth in organisations, we will always reach out to discuss opportunities.
Yes. Even if this role isn’t a perfect match, applying allows us to understand your expertise and ambitions, ensuring you're on our radar for the right opportunity when it arises.
We also work in several ways, firstly we advertise our roles available on our site, however, often due to confidentiality we may not post all. We also work with clients who are more focused on skills and understanding what is required to future-proof their business.
That's why we recommend registering your CV so you can be considered for roles that have yet to be created.
Yes, we help with CV and interview preparation. From customised support on how to optimise your CV to interview preparation and compensation negotiations, we advocate for you throughout your next career move.
