Data Engineer

Reposted 6 Hours Ago
Bangalore, Bengaluru Urban, Karnataka, IND
In-Office
Mid level
Fintech • Consulting
The Role
As a Data Engineer, you'll develop and maintain scalable data pipelines, utilize Azure technologies, and ensure data quality while collaborating with cross-functional teams.
Summary Generated by Built In

Equifax is where you can power your possible. If you want to achieve your true potential, chart new paths, develop new skills, collaborate with bright minds,  and make a meaningful impact, we want to hear from you.

Synopsis of the role 

As a Data Engineer, you will be a key builder within our data ecosystem, responsible for developing and maintaining the scalable data pipelines that power our business. Working closely with Lead Engineers and Architects, you will use Azure Databricks, PySpark, and Azure Data Factory to transform raw data into actionable insights. You will apply software engineering best practices to data processing, ensuring our Medallion Architecture remains performant, reliable, and secure.

What you’ll do

As a Data Engineer, you will focus on the development, automation, and optimization of our cloud data platform. Your core responsibilities include:

  • Pipeline Development: Build and deploy robust ETL/ELT workflows using Azure Data Factory (ADF) to ingest data from diverse internal and external sources.
  • Spark Engineering: Write clean, efficient PySpark code to perform complex data transformations, ensuring optimal resource utilization on Databricks clusters.
  • Lakehouse Maintenance: Develop and manage Delta Lake tables across Bronze, Silver, and Gold layers, implementing schema enforcement and data quality checks.
  • Data Modeling: Translate business requirements into physical data models, implementing Star Schemas and dimensional modeling to support BI tools like Power BI.
  • SQL Optimization: Author and tune sophisticated SQL queries for data validation, ad-hoc analysis, and reporting layer performance.
  • Data Governance Support: Work within Unity Catalog to manage data assets, ensuring proper tagging, documentation, and adherence to access control policies.
  • Automated Testing & CI/CD: Participate in the full DevOps lifecycle, writing unit tests for Spark logic and using Azure DevOps for continuous integration and deployment.
  • Monitoring & Troubleshooting: Proactively monitor pipeline health, identify bottlenecks, and resolve production issues to maintain high data availability.

What experience you need

  • Total Data Engineering Experience: 3–6 years of hands-on experience in data engineering, ETL development, or backend software engineering with a data focus.
  • Azure Foundations: 4+ years of experience working within the Azure cloud environment (Storage Accounts, Key Vault, Resource Groups).
  • Databricks & PySpark: 4+ years of experience building data transformation logic specifically using Databricks and Spark (Python preferred).
  • Relational Mastery: 3+ years of strong SQL skills, with a deep understanding of joins, window functions, and query execution plans.
  • Orchestration: Proven experience building multi-stage pipelines in Azure Data Factory or similar tools (e.g., Airflow, Synapse Pipelines).
  • Data Modeling Basics: Solid understanding of data warehousing concepts, including slowly changing dimensions (SCD) and Fact/Dimension table design.
  • Education/Certifications: Bachelor’s degree in CS or a related field. An Azure Data Engineer Associate (DP-203) certification is highly preferred.

 What could set you apart

Modern Data Stack Features

  • Delta Live Tables (DLT): Experience using DLT to simplify streaming and batch ETL development.

  • Databricks SQL: Familiarity with configuring SQL Warehouses for analyst self-service.

Software Engineering Rigor

  • Testing Frameworks: Experience with pytest or chispa for validating Spark transformations.

  • Python Proficiency: Strong general-purpose Python skills beyond just Spark (API integrations, automation scripts).

Performance & Scaling

  • Partitioning & Z-Ordering: Deep understanding of how to optimize Delta tables for large-scale query performance.

  • Streaming: Experience with Structured Streaming for real-time data ingestion from Event Hubs or Kafka.

Security & Compliance

  • Networking: Understanding of Azure VNet integration, Private Links, and secure data transit.

  • Data Privacy: Experience implementing data masking or encryption at rest/in transit.

Advanced Data Governance & Security
  • Unity Catalog Implementation: Experience configuring and managing Unity Catalog for fine-grained access control (Row-Level Security and Column-Level Masking) and tracking end-to-end data lineage.

  • Data Quality Frameworks: Expertise in building automated data validation using frameworks like Great Expectations or Databricks Expectations (DLT) to ensure data integrity before it reaches the Gold layer.

  • Metadata Management: Ability to maintain a searchable data catalog, ensuring all assets are tagged for PII (Personally Identifiable Information) and comply with GDPR/CCPA regulations.

Sophisticated CI/CD & DataOps
  • Infrastructure as Code (IaC): Proficiency in using Terraform or Bicep to deploy and manage Azure resources (Databricks workspaces, Key Vaults, Storage Accounts) as code.

  • Automated Testing Suites: Experience implementing a "Test-Driven Development" (TDD) approach for data, using pytest or chispa to run unit tests on PySpark transformations within the build pipeline.

  • Azure DevOps Integration: Mastery of YAML-based Azure Pipelines for automated deployment, including specialized tasks for Databricks Asset Bundles (DABs) or the Databricks CLI.

  • Environment Parity & Promotion: Proven ability to manage complex deployment patterns (Dev > QA > Prod) ensuring seamless promotion of code, ADF triggers, and Databricks job configurations.

  • Monitoring & Alerting: Setting up proactive monitoring using Azure Monitor and Log Analytics to track pipeline failures and cluster performance in real-time.

#India

We offer a hybrid work setting, comprehensive compensation and healthcare packages, attractive paid time off, and organizational growth potential through our online learning platform with guided career tracks.

Are you ready to power your possible?  Apply today, and get started on a path toward an exciting new career at Equifax, where you can make a difference!

Primary Location:

IND-Bangalore-Equifax-Analytics

Function:

Function - Data and Analytics

Schedule:

Full time

Top Skills

Azure Data Factory
Azure Databricks
Azure Devops
Azure Monitor
Ci/Cd
Delta Lake
Pyspark
SQL
Terraform
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: Atlanta, GA
16,742 Employees

What We Do

At Equifax (NYSE: EFX), we believe knowledge drives progress. As a global data, analytics, and technology company, we play an essential role in the global economy by helping financial institutions, companies, employers, and government agencies make critical decisions with greater confidence. Our unique blend of differentiated data, analytics, and cloud technology drives insights to power decisions to move people forward. Headquartered in Atlanta and supported by nearly 15,000 employees worldwide, Equifax operates or has investments in 24 countries in North America, Central and South America, Europe, and the Asia Pacific region. For more information, visit Equifax.com.

Similar Jobs

Optum Logo Optum

Data Engineer

Artificial Intelligence • Big Data • Healthtech • Information Technology • Machine Learning • Software • Analytics
In-Office
Bengaluru, Bengaluru Urban, Karnataka, IND
160000 Employees

Definitive Healthcare Logo Definitive Healthcare

Data Engineer

Big Data • Healthtech • Software
Hybrid
Bengaluru, Bengaluru Urban, Karnataka, IND
800 Employees

Boeing Logo Boeing

Platform Engineer

Aerospace • Information Technology • Software • Cybersecurity • Design • Defense • Manufacturing
In-Office
Bengaluru, Bengaluru Urban, Karnataka, IND
170000 Employees

Samsara Logo Samsara

Senior Data Engineer

Artificial Intelligence • Cloud • Computer Vision • Hardware • Internet of Things • Software
Easy Apply
Hybrid
Bengaluru, Bengaluru Urban, Karnataka, IND
4000 Employees

Similar Companies Hiring

Rain Thumbnail
Blockchain • Fintech • Payments • Financial Services • Cryptocurrency • Web3 • Infrastructure as a Service (IaaS)
New York, NY
100 Employees
Scotch Thumbnail
Artificial Intelligence • eCommerce • Fintech • Payments • Retail • Software • Analytics
US
35 Employees
Kepler  Thumbnail
Fintech • Software
New York, New York
6 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account