Sr Software Engineer - Hadoop Platform at Uber (San Francisco, CA)
Sorry, this job was removed at 8:38 p.m. (CST) on Thursday, February 10, 2022
By clicking continue you agree to Built In’s Privacy Policy and Terms of Use.
The Hadoop Platform team aims to build the next generation of intelligent data services that will profoundly impact how we march towards the vision of Uber's data-driven, real-time marketplace. Our team builds libraries and distributed services around the Hadoop stack (Hive, Presto, HDFS, Kafka, ....) to ease user interaction with the Big Data world. We share them with the open-source community (e.g., Hudi or Marmaray).
The Data Ingestion team owns the platform that moves hundreds of TBs of data a day from thousands of data sources to Uber's data lake. The hundreds of PBs data ingested by the Data Ingestion platform is the source of truth for business analytics and insights at Uber, powering various data-driven decisions. The platform consists of Marmaray, built from the ground-up at Uber and various services and tooling to operate the Data ingestion platform at scale and to deliver new data to customers at lower latencies.
What you will do:
We are currently looking for a strong engineer to join the Data Ingestion team to
Basic Qualifications ----
Preferred Qualifications ----
The Data Ingestion team owns the platform that moves hundreds of TBs of data a day from thousands of data sources to Uber's data lake. The hundreds of PBs data ingested by the Data Ingestion platform is the source of truth for business analytics and insights at Uber, powering various data-driven decisions. The platform consists of Marmaray, built from the ground-up at Uber and various services and tooling to operate the Data ingestion platform at scale and to deliver new data to customers at lower latencies.
What you will do:
We are currently looking for a strong engineer to join the Data Ingestion team to
- Design and implement distributed solutions to make new data available faster for business analytical needs.
- Drive reliability and operational scalability improvements through design and automation - scaling with the ever-increasing data volume flowing through the systems and the number of data sources producing data, observability, and monitoring.
- Drive efficiency efforts for one of the largest consumers of Uber's Data Infrastructure.
- Collaborate with multiple partner teams to achieve these multi-faceted goals.
Basic Qualifications ----
- Bachelor's degree with 4+ years of experience.
- Strong problem solving and coding skills in at least one of Java / C++ / Python.
- Experience with developing, debugging, and shipping software products on large codebases.
Preferred Qualifications ----
- Masters or Ph.D. in Computer Science or related subject area.
- Experience designing and executing large-scale distributed applications.
- Experience with big data infrastructure tools and software such as Spark, HDFS, Yarn.
- Demonstrated experience working collaboratively in cross-functional teams.
- Passion for learning new technologies, domains, and challenging the status quo.
Read Full Job Description