About Cerberus
Founded in 1992, Cerberus is a global leader in alternative investing with approximately $65 billion in assets across complementary credit, private equity, and real estate strategies. We invest across the capital structure where our integrated investment platforms and proprietary operating capabilities create an edge to improve performance and drive long-term value. Our tenured teams have experience working collaboratively across asset classes, sectors, and geographies to seek strong risk-adjusted returns for our investors. For more information about our people and platforms, visit us at www.cerberus.com.
Role Summary
The Data Engineer will be working closely with the Business Optimization Team which is a team based in the Baarn office and responsible for all offices in Europe. This team advances the business through the implementation of solution sets using a combination of technology, data and process changes. The Data Engineer will report into the Global technology Infrastructure team which is based in New York.
The Data Engineer role is responsible for the design, development, implementation, and operational support of Cerberus data tier within the Global technology Infrastructure team. The ideal candidate will be able to use appropriate programming languages to build robust, scalable pipelines that ingest, transform, and deliver financial data across domains. They’ll work with structured and semi-structured data from disparate sources, communicating in different modalities, using tools from the Microsoft stack of technologies and an emphasis on the Fabric platform. This engineer must be able to work autonomously in a collaborative team environment. Written and oral communication skills will be paramount in understanding the complexity of the data landscape, as well as managing deliverables from different stakeholders.
Responsibilities
1. Lakehouse/Warehouse Design and Maintenance
Architect and maintain Lakehouses using Microsoft Fabric’s OneLake and Delta Lake technologies.
Implement medallion architecture (Bronze, Silver, Gold layers) for scalable and modular data processing.
Optimize storage formats (e.g., Parquet, Delta) for performance and cost-efficiency.
2. Pipeline Development and Orchestration
Build robust Data Factory pipelines or Fabric Dataflows Gen2 to ingest data from diverse sources (SQL, APIs, cloud storage, etc.).
Schedule and orchestrate ETL/ELT workflows using Fabric Pipelines or notebooks.
Implement incremental loads, schema evolution, and error handling mechanisms.
3. Data Cleansing and Transformation
Apply data quality rules and transformations using Spark notebooks, SQL scripts, or Dataflows.
Standardize, enrich, and deduplicate data across ingestion layers.
Ensure consistency and referential integrity across medallion layers.
4. Monitoring and Observability
Monitor pipeline execution, performance, and failures using Fabric’s built-in telemetry and alerts.
Set up automated alerts for pipeline failures, latency breaches, or data anomalies.
Track resource utilization and optimize compute workloads.
5. Governance and Compliance
Implement data quality checks at ingestion and transformation stages.
Maintain data lineage using Microsoft Purview integration for traceability and auditability.
Enforce access controls, sensitivity labels, and data retention policies.
Document data assets and transformations for transparency and collaboration.
6. Collaboration and DevOps
Use Infrastructure as Code (IaC) tools like Terraform to provision Fabric resources.
Collaborate with analysts, data scientists, and business stakeholders to align data models with business needs.
Integrate with Azure DevOps for CI/CD of pipelines and notebooks.
7. Performance Tuning and Optimization
Profile and tune Spark jobs and SQL queries for optimal performance.
Partition and index data for faster access in Lakehouse queries.
Manage compute capacity and scale workloads efficiently.
Qualifications:
7+ years of experience in T-SQL and Python Development.
3+ years in Fabric/Databricks or related lakehouse platform with Apache Spark engine.
Hands on experience with Data Lakehouse/Warehouse development and principles, including development in notebooks using Pyspark/SparkSQL.
Excellent troubleshooting skills and ability to apply them under pressure
Strong experience in logical/physical database designs
Knowledge of cloud platform services for ETL/ELT
Excellent written and oral communication skills
Ability to learn and apply new technologies to the existing environment
Ability to perform work with minimal guidance from the other team members
Willingness to communicate and work with business stakeholders and understand their requirements
Preferred Skills:
Experience with building pipelines and notebooks using PySpark/SparkSQL.
Experience with Source Control platforms such as Git/Azure dev ops, their usage and principles of CI/CD.
Knowledge of Microsoft Azure services including Azure SQL Databases, Managed SQL, Azure Synapse Analytics, Cosmos DB, Azure Data Factory, Databricks, Fabric and Power BI.
Prior work experience in the financial sector, especially Capital Markets or Asset Management.
Knowledge of the Real Estate sector and European markets is a plus.
Certifications such as Azure Data Engineer Associate (DP-203), Fabric Data Engineer Associate (DP-700), Fabric Analytics Engineer Associate (DP-600) are a big plus.
Location and Hours
The role is based in the Baarn office, the Netherlands. Flexible Hybrid working possible.
Employment Condition
The position is for 40 hours per week
Salary range: EUR 80,000 – EUR 120,000 gross per year, depending on relevant experience, plus a bonus of approximately 15%
In addition to the gross salary, the employee is entitled to reimbursement of travel expenses (EUR 0.23 per kilometer), a working-from-home allowance (EUR 2.45 per day), and a net expense allowance based on job level (EUR 1,200 per year)
25 vacation days per year
Disability insurance fully paid by the employer
Employer-paid pension contribution (8%), including survivor’s pension and ANW gap insurance
The employment contract will initially be entered into for a fixed term of one year
Restrictions apply with regard to personal investments
As part of the recruitment process, a background screening must be successfully completed before the employment contract becomes effective
Other contractual terms and conditions apply
Top Skills
What We Do
About Cerberus Capital Management, L.P.
Founded in 1992, Cerberus is a global leader in alternative investing with approximately $60 billion in assets across complementary credit, private equity, and real estate strategies. We invest across the capital structure where our integrated investment platforms and proprietary operating capabilities create an edge to improve performance and drive long-term value. Our tenured teams have experience working collaboratively across asset classes, sectors, and geographies to seek strong risk-adjusted returns for our investors.
About Cerberus Operations and Advisory Company, LLC
The proprietary operations affiliate of Cerberus Capital Management, L.P., Cerberus Operations and Advisory Company, LLC (COAC) is one of the largest and most experienced teams of operating executives and functional experts in alternative investing. COAC consists of approximately 110 professionals from a wide spectrum of industries and business disciplines that assist Cerberus by sourcing opportunities, conducting highly informed due diligences, executing acquisitions, and improving the performance of portfolio investments. COAC executives often hold senior leadership and director positions at Cerberus’ portfolio companies, and COAC teams are regularly deployed to provide practical assistance to help drive operational and strategic transformations.
About Cerberus Technology Solutions, LLC
Cerberus Technology Solutions, LLC (CTS) is an operating company and subsidiary of Cerberus Capital Management focused exclusively on leveraging emerging technology, data, and advanced analytics to drive transformations. CTS employs expert technologists that work closely with Cerberus investment and operating professionals to apply new technologies, realize new sources of revenue and value creation, and accelerate technological transformation and differentiation.
For more information about our people and platforms, visit us at www.cerberus.com.









