Key Responsibilities : Data Engineering
- Build and optimize scalable data pipelines for ingestion, transformation, and storage.
- Work with structured and unstructured data, handling diverse file formats such as CSV, Excel, JSON, and PDFs.
- Extract, parse, and integrate emission factor databases (EFDBs) from external sources, ensuring they are structured for ingestion, efficient retrieval and analysis.
- Leverage on scheduler tools such as Apache Airflow, Dagster, or equivalent to manage and automate data workflows.
- Ensure data integrity, consistency, and governance while working with high-volume datasets.
- Develop and optimize queries in MongoDB and SQL for high-performance querying.
- Manage and deploy data infrastructure on cloud providers such as AWS.
- Consolidate data models written in Node.js and Python across different systems, ensuring a single source of truth for internal applications.
Analytics Engineering
- Design and implement robust data models to support analytics, business intelligence (BI), and data science.
- Advise on data modeling strategies to optimize performance, maintainability, and scalability.
- Enable BI reporting and self-service analytics by preparing analytics-ready datasets.
- Work with BI tools such as Tableau, GoodData, Power BI, Looker, Metabase, or equivalent to build visualizations and dashboards.
- Optimize query performance, materialized views, and aggregation strategies for efficient reporting.
- Collaborate closely with data scientists, Product, Implementations and Sales teams to provide actionable insights.
- Ensure that datasets are properly indexed and structured for fast and efficient access.
What We Are Looking For
- 5 to 8 years of experience in Data Engineering and/or Analytics Engineering roles.
- Strong knowledge of Python (including pandas library) and/or Node.js, SQL.
- Proven experience with SQL (PostgreSQL) and/or NoSQL databases (MongoDB or AWS DocumentDB), including indexing, partitioning, and query optimization.
- Experience with data modeling (designing database schemas, tables, entity-relationship diagrams, etc)
- Strong data structures and algorithms knowledge and understanding of time and space complexity.
- Experience working with scheduler tools like Apache Airflow, Dagster, or similar frameworks.
- Ability to parse and process unstructured data, including PDFs, Excel, and other file formats.
- Experience working with large-scale data systems and optimizing query performance.
- Hands-on experience with AWS, GCP, or Azure, for data storage, compute, databases and security.
- Possess knowledge in DevOps and Infrastructure-as-Code (Terraform, GitOps, Kubernetes, CI/CD tools like GitHub Actions, Argo CD).
- Passion or willingness to learn about sustainability and carbon emissions data.
Nice to Have
- Experience with other NoSQL and SQL databases beyond MongoDB.
- Hands-on experience with streaming data architectures such as Kafka.
- Exposure to distributed computing frameworks such as Spark.
- Knowledge and practical experience in AWS EC2, S3, Glue, IAM and CloudWatch.
- Experience in data security, governance, and compliance best practices.
- Background in carbon accounting methodologies or sustainability-related data processing.
- Experience working in a SaaS-product company and/or startups, and comfortable with change and ambiguity.
Top Skills
What We Do
Terrascope is a enterprise grade, end to end, smart carbon measurement and management SaaS platform. And we are on a mission to empower companies to build a credible pathway to net zero.
By combining data science, machine learning and sustainability expertise, our platform provides the data, analytics and digital tools to help large companies decarbonise their business operations and supply chains.
Powered by technology, data science and deep sustainability expertise, Terrascope is on mission is to drive decarbonisation at scale by helping enterprises:
- Measure with confidence. Terrascope increases the speed, accuracy, and confidence of scope 1, 2, and 3 emissions measurement, while ensuring compliance with GHG protocol, reporting frameworks and assurance standards.
- Manage complexity. Terrascope helps enterprises focus decarbonisation efforts where it matters the most by identifying emission hotspots and defining next best actions to make tangible progress towards net zero.
- Collaborate seamlessly. Terrascope enables collective action and shared accountability by allowing internal and external stakeholders to centralise data