Lead MLOps Engineer (Databricks/AWS/SageMaker/Python) - Fully remote - OUTSIDE IR35 at Urban Connect, United Kingdom, £Contract Rate

Contract Description

Digital Transformation & Software Engineering consultancy immediately require a Lead MLOps Engineer on a contract basis who has expertise in Databricks and AWS SageMaker.


Contract Details:


  • Job Title: Lead MLOps Engineer
  • Day Rate: £625
  • Location: Fully Remote
  • Determination: OUTSIDE IR35
  • Duration: 3 months initially (extension highly likely)
  • Start Date: ASAP


Project Scope:


Time-bound delivery programme focused on migrating production ML workloads from Databricks to AWS SageMaker to a client within a regulated environment. The immediate priority is a container-first migration of existing Databricks-hosted ML workloads to AWS, with SageMaker as the default execution platform and a hard commercial deadline. In parallel, you will help define the future MLOps operating model on SageMaker, which will become business-as-usual once the migration completes.


Responsibilities:


  • This is a hands-on technical leadership role where you will set patterns, review work, unblock delivery, and personally handle the most complex migrations.
  • Act as the Lead MLOps Engineer delivering the migration from Databricks to AWS SageMaker.
  • Own the technical direction, delivery integrity, and coordination across all technical workstreams.
  • Lead and coordinate work across multiple streams (standardised migrations, complex/edge-case workloads, platform foundations), working closely with Data Engineers, Cloud Engineers, Delivery Management, and Data Science SMEs.


What you’ll be doing:


You will lead and contribute across the following areas:


  • AWS SageMaker-based ML execution - Designing and operating batch processing, training, and (where appropriate) inference workloads on SageMaker.
  • Databricks to SageMaker migration - Migrating Databricks notebooks, jobs, and ML workloads into containerised execution on AWS, ensuring behavioural parity and production stability.
  • Python-based ML workloads - Working directly with Python-based ML codebases (e.g. sklearn, XGBoost, and similar libraries), refactoring only where required to support containerised execution.
  • Containerised ML runtimes - Using containers to replicate Databricks runtimes, manage Python dependencies, and stabilise legacy workloads.
  • ML pipelines & automation - Orchestrating end-to-end ML workflows on AWS, including batch execution, retraining, and validation.
  • Monitoring, validation & governance - Implementing monitoring, logging, and validation patterns suitable for regulated production ML environments.


Essential skills & experience (must-haves):


  • Proven, hands-on experience migrating ML workloads from Databricks to AWS SageMaker (this is non-negotiable).
  • Strong experience building and operating Python-based ML workloads in production environments.
  • Solid understanding of container-based ML execution and Python dependency management.
  • Experience leading or owning technical delivery across multiple engineers and workstreams.
  • Comfort working in regulated or high-governance environments where validation, auditability, and controlled change are required.