Available for contract · Netherlands

Zhe Ren, or call me Brian

Data Platform Engineer & Data Engineer

6+ years building large-scale data platforms in Europe — pipelines, lakehouse architecture, and production reliability at scale.

Profile

Data Engineer with 6+ years of experience designing and operating large-scale data platforms, specialising in scalable pipeline development, ETL/ELT architecture, and production reliability. Hands-on experience with Python, dbt, Airflow, Snowflake, and AWS, with a strong track record of end-to-end ownership across ingestion, transformation, data quality, and observability.

Co-architected a streaming-first market data platform operating ~200 pipelines in a regulated, high-volume European energy trading environment. Currently transitioning to freelance consulting, available for contract engagements in the Netherlands.

Skills

Cloud — AWS

S3, Lambda, DynamoDB, MSK, EC2

Cloud — GCP

GKE, BigQuery, Pub/Sub, GCS

Languages

Python, SQL, Bash

Data Tools

Spark, dbt, Airflow, Dagster, Databricks, Snowflake

Infra & DevOps

Terraform, Helm, Docker, GitLab CI/CD, GitHub Actions, Jenkins

Observability

Grafana, Prometheus, Sentry

Python Apache Spark AWS Google Cloud Apache Airflow Dagster dbt Databricks Snowflake Apache Kafka Redis Terraform Helm Docker Kubernetes GitLab GitHub Actions Grafana Prometheus Sentry

All product names, logos, and trademarks are the property of their respective owners and are used here for identification purposes only.

Experience

09/2023 – 03/2026 Amsterdam, NL

Data Platform Engineer / Data Engineer

Dexter Energy

  • Co-led and built the Market Data Platform (MDP) from scratch — Medallion + Kappa architecture, ~200 pipelines, ~90% legacy migration completed.
  • Designed a connector-based ingestion framework (polling and push) with YAML data contracts as the single source of truth for schema validation.
  • Implemented 7–9 production connectors (EPEX MATS, EEX, Volue, APG, GME, IPTO, DAMAS) with custom rate-limit handling.
  • Built core platform mechanisms: sanitisation, idempotent ingestion, Redis-backed distributed deduplication.
  • Owned production reliability — triaging data gaps, schema drift, and outages; improved observability via Grafana, Prometheus, and Sentry.
  • Maintained dbt consolidation layer feeding DWH, analytics, and model pipelines.
11/2020 – 09/2023 Hoofddorp, NL

Data Platform Engineer

FedEx

  • Designed scalable ELT pipelines ingesting multi-domain enterprise datasets and SaaS platforms (Salesforce, Google Ads) into a centralised data platform.
  • Orchestrated ELT pipelines using Apache Airflow — self-hosted via Terraform, later migrated to Cloud Composer.
  • Delivered self-service PySpark/SQL templates enabling 50+ analysts to build and schedule jobs independently.
  • Implemented CI/CD pipelines for automated testing and deployment.
  • Enabled lakehouse data exploration through Apache Superset.
  • Operated platform infrastructure on GKE including orchestration, compute runtime, and monitoring.
05/2020 – 11/2020 Hoofddorp, NL

Data Engineer

FedEx via Xccelerated

  • Built data pipelines and microservices for the O2P migration (TNT → FedEx).
  • Developed orchestration tooling for hourly migration across users.
  • Built an MVP with scheduled PySpark SQL jobs + microservices for automated campaign engagement (Salesforce / Adobe Target).

Education

2014 – 2016 Groningen, NL

MSc Biomedical Engineering

University of Groningen

2010 – 2014 Guangzhou, CN

BSc Biomedical Engineering

South China University of Technology

Contact

Available for contract engagements in the Netherlands.