Data Engineer, Discover Data Infrastructure
Snap Inc. is a camera company. We believe that reinventing the camera represents our greatest opportunity to improve the way people live and communicate. Our products empower people to express themselves, live in the moment, learn about the world, and have fun together.
We’re looking for a Data Engineer to join the Discover Data Infrastructure team. Working from one of our offices in Santa Monica, CA, Mountain View, CA, Seattle, WA, or New York, NY, you’ll collaborate with Software Engineers, Data Scientists, and Product Managers to help Snapchat grow across the globe. The Discover Data Infrastructure team builds and operates key components for metrics reporting, business analysis, machine learning signals and training data for key content products in Snap. In this role, you will build the data infrastructure and tools that will deliver insights to broaden and deepen user engagement and improve product experience for our hundreds of millions of passionate users. You will have an opportunity to tackle large-scale engineering and product challenges while working alongside kind, smart, and creative colleagues. Come grab a front row seat to witness and influence how Snapchat grows to become the world's camera!
What you’ll do:
Define data models for instrumentation and reporting in partnership with Engineering, Data Science, and Product Management to support product analytics.
Build scalable aggregation pipelines to deliver performant datasets that can be consumed through surfaces such as Looker, Tableau, Superset, and Jupyter.
Drive data quality end-to-end from instrumentation to reporting. Build automated controls and processes to prevent and fix regressions.
Democratize data access amongst engineers, PMs, and scientists with well-documented and extensible pipelines and datasets.
Partner with Snap’s Data Governance and Insights teams to make high-quality datasets in the Growth domain available for external and partner reporting.
Knowledge, Skills & Abilities:
Experience in building data pipelines to serve reporting needs
Experience owning all or part of a team roadmap
Experience with data visualization tools like Looker and Tableau
Ability to prioritize requests from multiple stakeholders in disparate domains
Ability to effectively communicate complex projects to non-technical stakeholders
Minimum Qualifications:
BS/BA degree in Computer Science, Math, Physics, or a related field, or equivalent years of experience in a relevant field
5+ year experience in SQL or similar languages
5+ years development experience in at least one object-oriented or scripting language (Python, Java, Scala, etc), Python preferred
Experience in ETL / Data application development
Preferred Qualifications
Hands on experience with Google BigQuery
Experience using and sharing notebook solutions like Jupyter
Experience in version control systems such as Git
Data architecture and warehousing experience
Experience with Airflow and Druid