Posted on Monday, 2nd February 2026
We’re seeking an experienced Databricks specialist to design, build, and optimise scalable data pipelines. You’ll work closely with business and analytics teams to deliver high‑quality, reliable, and well‑governed data solutions, while championing modern DataOps practices. This role is remote but you must be UK based.
Build and maintain Databricks-based data pipelines (PySpark, Delta Lake, SQL).
Integrate data from APIs, databases, and file systems.
Work with stakeholders to translate requirements into engineering solutions.
Ensure data quality, reliability, security, and governance.
Implement Git, CI/CD, automated testing, and IaC.
Optimise Databricks jobs, clusters, and workflows.
Document processes and best practices.
Strong experience with Databricks, PySpark, Delta Lake, and SQL.
Advanced Python and SQL for data transformation.
Solid understanding of ETL/ELT, data warehousing, and distributed processing.
Hands‑on with DataOps: Git, CI/CD, testing, IaC.
Experience with AWS, Azure, or GCP.
Strong troubleshooting and performance optimisation skills.
Self‑sufficient contractor comfortable leading technical decisions.
Strong communicator who collaborates well with business and analytics teams.
Delivery‑focused, pragmatic, and experienced in fast‑moving environments
