What you will do :
- Technical
- Data testing & validation: Design, implement and maintain unit, integration, regression, pipeline, acceptance and data-quality tests for ETL/ELT jobs, streaming pipelines and data services (batch and real-time).
- Automation: Build and own test automation frameworks and test harnesses for data pipelines (e.g., Python/pytest, dbt tests, Great Expectations), including synthetic dataset generators, golden datasets and fixture management.
- Data systems: Deep experience testing data platforms and tooling: cloud data systems (Snowflake, BigQuery, Redshift, Microsoft Fabric), relational databases (MS SQL) and cloud providers like Azure , orchestration tools (Airflow), and columnar file formats (Parquet, Avro, ORC).
- SQL & code: Expert in SQL for analytical validation and in at least one programming language used for data engineering such as Python. Build scripts, tools and automation to validate schema, data lineage, transformations, aggregation correctness, null/missing value handling and performance characteristics.
- Observability & monitoring: Implement automated checks and alert logic for data freshness, schema drift, volume anomalies and metric regressions using observability/data quality tools (e.g., Great Expectations, Monte Carlo, custom monitoring).
- Performance & scale testing: Design and run performance, throughput and scalability tests for pipelines and data services; profile and tune ETL jobs and queries to identify bottlenecks.
- Data contracts & governance: Work with engineering and product teams to enforce data contracts, contract tests for data APIs, and validate PII handling, access controls and compliance requirements.
- Debugging & triage: Investigate, reproduce and document data defects, root causes and remediation plans; apply production debugging skills against ETL jobs, queries and logs.
- Process
- Follow SDLC and agile practices, including writing and reviewing tests as code, peer code reviews, CI/CD pipeline integration of tests, and release gating based on automated data quality checks.
- Maintain and follow coding and test case management standards; ensure test suites are deterministic, reproducible and fast enough to run in CI and nightly/rolling regression schedules.
- Document testing approaches, known limitations and runbooks for operational incidents involving data.
- Impact
- Own quality for analytics deliverables and data platform features within your area. Establish test strategies and define appropriate test coverage for pipelines, models and metrics.
- Drive improvements in data quality and reliability through automation, test architecture, and proactive detection of data problems.
- Contribute to data documentation (data dictionaries, lineage diagrams, assumptions) and raise the bar for data governance and observability across teams.
- Communication
- Collaborate with Data Engineers, Analytics Engineers, Data Scientists, Product Managers and Platform Engineers to design tests and validate business logic in analytical pipelines.
- Explain technical tradeoffs, testing coverage and risk to non-technical stakeholders and suggest pragmatic compromises when necessary.
- Produce clear, evidence-based defect reports and remediation plans; participate in post-mortems focused on data incidents.
Education & Experience :
- Required: Bachelor’s degree in Computer Science or related field, or equivalent years’ experience
- Required: 4+ years of testing or SDET experience with a strong emphasis on data systems and analytics. Demonstrated experience writing tests for data pipelines, validating analytical outputs, or building test frameworks for ETL/ELT processes.
- Required: Proficiency in SQL and at least one general-purpose programming language used for data processing (Python preferred; Scala/Java acceptable).
- Required: Experience with cloud data warehouses (e.g., Snowflake, BigQuery, Redshift), and familiarity with at least one orchestration or streaming technology (Airflow, Kafka, Spark).
- Preferred: Experience with dbt, data observability platforms (e.g., Monte Carlo), data modeling, and analytical/BI tools (Looker, Tableau, etc.). Experience testing ML models, A/B test validation, or statistical methods is a plus.
- Required : Experience within Travel industry
What we are looking for :
- Strong SQL skills for complex analytical validation: joins, window functions, aggregation correctness and performance tuning.
- Proficient in Python for test automation, data manipulation (pandas/pyarrow), and building test harnesses.
- Demonstrable experience designing automated data quality checks and defining acceptance criteria for analytics deliverables.
- Experience with CI/CD systems (GitHub Actions, Jenkins, CircleCI) and integrating test runs into pipelines.
- Familiarity with data formats (JSON, Avro, Parquet) and schema evolution strategies.
- Solid understanding of distributed data processing, consistency, eventual consistency tradeoffs, and data lineage.
- Ability to reason statistically about datasets—detecting outliers, sampling strategies, and validating model/metric correctness.
- Excellent debugging skills across code, SQL, job logs and metadata; ability to produce reproducible test cases and remediation paths.
- Strong collaboration and written communication skills; experience conducting design/code reviews and mentoring peers.
Top Skills
What We Do
Emburse humanizes work by empowering business travelers, finance professionals and CFOs to eliminate manual, time-consuming tasks so they can focus on what matters most.
Emburse brings together some of the world’s most powerful and trusted expense and AP automation solutions, including Abacus, Captio, Certify, Chrome River, Nexonia and Tallie. The company’s innovative offerings, which are uniquely tailored for specific industries, company sizes, and geographies, are trusted by more than 4.5 million users in more than 120 countries. Over 14,000 customers, from start-ups to global enterprises, including Boot Barn, Grant Thornton, Telefónica, Lufthansa Systems, and Toyota rely on Emburse to make faster, smarter decisions, empower business travelers to recapture lost nights and weekends spent doing tedious expense management, and help make users’ lives -- and their businesses -- better.







