Senior Data Engineer
Remote (Poland or Romania) | Monthly travel to Geneva | Full-time | Employment via Employer of Record
DLG builds LuxuryIQ, a market intelligence platform used by 80+ luxury brands. The platform ingests data from 50+ sources — social platforms, search engines, secondary markets, advertising networks — through BigQuery into client-facing analytics and AI-powered tools.
The data infrastructure exists and runs in production. It needs an experienced engineer to take ownership, improve reliability, expand coverage, and enforce quality standards across the full pipeline.
What makes this different: your clean, documented data will power a conversational AI layer (via Model Context Protocol) that lets luxury brand CEOs query market intelligence in natural language. You’re not building another analytics pipeline — you’re building the foundation of an AI product.
SCOPE ●
As our most senior individual contributor on the data team, you will set the technical direction for pipeline architecture and data quality standards. You will take end-to-end ownership of mission-critical pipelines — from raw ingestion through to clean, documented delivery in BigQuery — and be the go-to technical authority on data behaviour across the platform.
Responsibilities
- Design and develop data processing, cleansing, transformation, and QA scripts using Python and SQL
- Own the operation, maintenance, and enhancement of existing mission-critical data pipelines and ETLs
- Monitor pipeline reliability and data quality standards; identify and close coverage gaps systematically
- Optimise ETL performance and migrate legacy pipelines into a more scalable architecture
- Define and enforce data quality standards across the full pipeline — you treat data quality as non-negotiable
- Contribute to architectural decisions on ingestion, transformation, and delivery layer design
- Own the full data lifecycle documentation: every transformation rule, edge case, and business logic decision must be spec'd, not held in someone's memory
- Fetch datasets from external data sources: REST APIs, JSON, CSV
- Perform data migrations and batch data updates
- Conduct exploratory analysis to support new business requirements and data source onboarding
- Develop proof-of-concept pipelines for new data initiatives
REQUIREMENTS
Technical
- Graduate in Computer Science, Information Systems, Statistics, Mathematics, or related field
- 7+ years of relevant experience in data engineering, database or ETL development
- Excellent Python skills (pandas, numpy); able to write production-grade pipeline code
- Excellent SQL skills and strong database design and development experience
- Proven ETL development and long-term maintenance experience
- Apache Airflow experience
- Entity resolution and deduplication across heterogeneous, multi-source datasets
- Web scraping, crawler development, and API integration experience
- Working knowledge of AI/LLM tooling (Claude Code, Copilot, or equivalent)
Preferred
- BigQuery (primary stack) and/or GCP experience — strongly preferred
- PostgreSQL experience
- dbt experience
- Experience with AliCloud, AWS, or similar cloud platforms
- Big Data processing technologies: Databricks, Spark, Redis, Kafka
Behavioural
- You take full ownership of your scope and drive it independently
- You identify gaps in pipeline coverage and close them systematically
- You document thoroughly — undocumented systems are unreliable systems
- You treat data quality as non-negotiable — you've been burned by bad data before and it shows
IDEAL BACKGROUND ●
You’ve worked with scraped, unstructured, messy real-world data at scale. You know what it’s like to reconcile the same entity across dozens of sources with different naming conventions, missing fields, and inconsistent formats.
Strong-fit industries: Price comparison and travel aggregation platforms, marketplace intelligence and web analytics, real estate platforms, alternative data providers for finance, or competitive intelligence companies.
The key differentiator: Have you built a data product that external customers pay for? That’s a different mindset than running internal data ops. Interest in the luxury industry is a plus but not required — the data engineering challenges are universal.
Top Skills
What We Do
With offices in Geneva, New York and Shanghai, DLG (Digital Luxury Group) is the leading strategic partner for luxury & lifestyle brands.
We know what sophisticated audiences want and how to engage them through meaningful digital experiences.
In August 2021, DLG Geneva was awarded the Great Place to Work© certification and ranked among Europe's Top 10 Best Workplaces









