cover image
bp

Senior Data Engineer

Remote

United kingdom

Senior

Freelance

29-01-2026

Share this job:

Skills

Python SQL dbt (data build tool) CI/CD AWS Analytics Spark Databricks PySpark

Job Specifications

Job Title: CDI Senior Analytics Engineer

Job Location: Remote (UK)

Contract Length: 3 Months Contract

Industry: Oil and Energy, Tech

Working Hours: 8 hours per day/ 40 per week

Role Overview:

We are looking for a hands-on engineering heavy-hitter to join our Customer Data & Intelligence (CDI) function immediately. We have a rich dataset covering multiple global markets, but our legacy codebase (monolithic SQL and Python scripts) is fragile.

Your immediate mission is to triage, refactor, and stabilize our critical data pipelines. You will take "God Queries" and break them down into modular, testable, and performant dbt models.

Immediate Deliverables (First 30-60 Days):

The "Code Rescue": Audit and patch critical queries currently causing data corruptions. Fix logic errors.
Modularization Pilot: Implement dbt (Data Build Tool) within our AWS/Databricks environment. Migrate the most critical reporting tables from stored procedures/scripts into dbt models.
Automated Quality Gates: Deploy automated tests (using dbt tests or Great Expectations) to check for identity uniqueness and any data errors on critical columns. Stop bad data before it hits the dashboard.

What you will do:

Refactoring: Rewrite inefficient legacy SQL to improve performance and readability.
Pipeline Repair: Fix error handling in existing AWS Glue/PySpark jobs .
Standardization: Establish the "Gold Standard" for what good code looks like. Create the Pull Request template and SQL linting rules that the rest of the team must follow.
Mentorship: Act as the "Bar Raiser" in code reviews, establishing standards and teaching the existing team how to write modular, defensive code.

What you will have:

Expert SQL: You can read a multi thousand-line spaghetti query, understand the intent, and rewrite it into 5 clean, modular steps without losing logic.
dbt (Data Build Tool): Proven experience setting up dbt from scratch. You know how to structure a project (Staging -> Intermediate -> Marts).
Python & Spark: Ability to read and fix PySpark syntax errors and optimize Spark execution plans (Databricks/AWS Glue).
AWS Ecosystem: Comfortable with S3, Athena, and IAM permissions.
CI/CD: Experience setting up and running tests automatically on commit.

Why You?:

You hate "Toil." You refuse to check data manually; you write scripts to check it for you.
You are not afraid of legacy codes. You see a messy codebase as a puzzle to be solved, not a reason to run away.
You care about Truth. You understand that "mostly correct" data is useless to a business.

We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, colour, national origin, sex, gender, gender expression, sexual orientation, age, marital status, veteran status, or disability status.

About the Company

We believe the world wants and needs a better and more balanced energy system that delivers secure, affordable and lower-carbon energy. We're playing our part by investing in today's energy system, which is mainly oil and gas – and, not or – in our transition and the energy transition. While today we’re mostly in oil and gas, we increased the proportion of our global annual investment that went into our lower carbon & other transition businesses from around 3% in 2019 to around 23% in 2023. Increasing investment in these ... Know more