Our Data team consists of highly skilled senior software and data professionals who collaborate to solve complex data challenges. We process billions of records daily from multiple sources using multi-stage pipelines with intricate data structures and advanced queries.
We are responsible for building data pipelines end to end—from raw data ingestion to the creation of actionable datasets—following the bronze, silver, and gold paradigm. This includes business logic, infrastructure, ETLs, optimization, and ongoing maintenance.
The data we deliver drives insights and decision-making across the organization and enhances our product offerings. We leverage technologies such as AWS, Snowflake, Iceberg, Airflow, Spark, and more.
What You’ll DoLead the translation of business and product requirements into scalable data models, transformations, and pipelines.
Design and own datasets across bronze, silver, and gold layers, including defining grain, aggregations, and data contracts.
Develop and maintain SQL-heavy data pipelines and Airflow DAGs (workflow logic, dependencies, backfills, python, and lots of SQL).
Own data correctness for key business metrics (e.g., ARR), including deep root cause analysis and resolution of data issues.
Define and drive best practices for SQL, data modeling, and pipeline design across the team.
Optimize queries and data models for performance, scalability, and cost efficiency.
Collaborate closely with product managers, analysts, and BI developers to refine requirements and ensure high-quality data delivery.
Develop AI-agents to accelerate data analysis by internal and external users.
Work with complex data inputs (e.g., JSON, schemas, logs) and incorporate them into robust data pipelines.
Requirements
7+ years of experience as a Data Engineer or Architect or in a similar data-focused role, with clear ownership of end-to-end data solutions.
Strong expertise in writing and optimizing complex SQL queries (advanced joins, aggregations, performance tuning).
Proven experience building and maintaining Airflow DAGs (or similar orchestration tools), focused on workflow logic, code, and SQL rather than infrastructure.
Deep understanding of data modeling principles, including designing datasets at the correct grain and preventing data inconsistencies. Use of the medallion model.
Strong ability to understand business needs and translate them into scalable, maintainable data solutions.
Demonstrated experience debugging data issues and tracing discrepancies in critical business metrics across pipelines.
Proficient in Python for orchestration and data workflows.
Comfortable reading and reasoning about existing code, SQL, DAGs, schemas, and input data formats (e.g., JSON).
Experience working with cloud data warehouses such as Snowflake, BigQuery, or Databricks.
Experience in Snowflake and its extended SQL and nuances is a strong advantage.
Skills Required
- 7+ years of experience as a Data Engineer or Architect
- Strong expertise in writing and optimizing complex SQL queries
- Proven experience building and maintaining Airflow DAGs
- Deep understanding of data modeling principles
- Proficient in Python for orchestration and data workflows
- Experience working with cloud data warehouses like Snowflake
What We Do
Cloudinary’s mission is to empower companies to deliver visual experiences that inspire and connect by unleashing the full potential of their media. With more than 50 billion assets under management and 7,500 customers worldwide, Cloudinary is the industry standard for developers, creators and marketers looking to upload, store, transform, manage, and deliver images and videos online. As a result, leading brands like Atlassian, Bleacher Report, Grubhub, Hinge, NBC, Mediavine, Peloton, Petco and Under Armour are seeing significant business value in using Cloudinary, including faster time to market, higher user satisfaction, and increased engagement and conversions. For more information, visit www.cloudinary.com.








