Back to jobs

Data Engineer

Intellect Group
London Area, United Kingdom
Full-time

🚀 Are you a Data Engineer (5–7+ years) who enjoys owning production-grade pipelines end-to-end, optimising performance, and working with modern Python tooling on time-series datasets?

I’m supporting a London-based fintech in their search for a hands-on Data Engineer to help build and improve the data infrastructure powering a unified data + analytics API for financial markets participants.

You’ll sit in a small engineering/analytics team and take ownership of pipelines end-to-end — from onboarding new datasets through to reliability, monitoring and data quality in production. Finance experience is a bonus, but not essential.

Note: they use cloud infrastructure but deploy services on their own servers, so a strong production/ops mindset is important.

In this role, you’ll:

* Build, streamline and improve ETL/data pipelines (prototype → production)

* Ingest and normalise high-velocity, time-series datasets from multiple external sources

* Work heavily in Python with a modern columnar stack (Polars + Parquet/Arrow/PyArrow; DuckDB is a nice-to-have)

* Orchestrate workflows and improve reliability (they use Temporal — similar orchestration experience is fine)

* Own production readiness: validations, automated checks, backfills/reruns, monitoring/alerting, incident/RCA mindset

* Work independently and help drive delivery forward — including providing practical technical guidance to shape solutions

What’s in it for you?

* Modern Python stack – Polars + Parquet/Arrow (DuckDB a plus)

* Ownership & impact – high visibility; you’ll influence performance and reliability directly

* Market/time-series exposure – complex financial datasets; learn the domain as you go

* Hybrid London – London preferred, 2–3 days in the office

* Start ASAP – interviewing now

What my client is looking for:

* 5–7+ years hands-on data engineering experience

* Strong Python + SQL fundamentals (ETL, pipelines, data modelling, performance)

* Hands-on experience with Polars and Parquet/Arrow/PyArrow

* Proven ability to operate pipelines in production (monitoring, backfills, data quality, incidents)

* Able to work independently and drive things forward without heavy oversight

* Interest in financial data (experience helpful but not required)

Nice to have:

* DuckDB experience

* Time-series experience (market data, telemetry, pricing, events)

* Streaming exposure (Kafka/Event Hubs/Kinesis)

* Experience with Temporal (or similar orchestrators like Airflow/Dagster/Prefect)

* Any exposure to AI agents / automation tooling

👉 Apply now!

Applications go to the hiring team directly