Data Engineer

Reposted 11 Days Ago
Be an Early Applicant
Hiring Remotely in Hyderabad, Telangana, IND
In-Office or Remote
Senior level
Database • Analytics
The Role
The Data Engineer will design and optimize data pipelines and workflows to support a Media Mix Optimization platform, ensuring data integrity and governance while collaborating with Data Science and BI teams.
Summary Generated by Built In
Company Description

Blend360 is a data and AI services company specializing in data engineering, data science, MLOps, and governance to build scalable analytics solutions. It partners with enterprise and Fortune 1000 clients across industries including financial services, healthcare, retail, technology, and hospitality to drive data-driven decision making. Headquartered in Columbia, Maryland, the company is recognized for rapid growth and global delivery of AI solutions through the integration of people, data, and technology.

We are seeking a hands-on Data Engineer with deep expertise in distributed systems, ETL/ELT development, and enterprise-grade database management. The engineer will design, implement, and optimize ingestion, transformation, and storage workflows to support the MMO platform. The role requires technical fluency across big data frameworks (HDFS, Hive, PySpark), orchestration platforms (NiFi), and relational systems (Postgres), combined with strong coding skills in Python and SQL for automation, custom transformations, and operational reliability.

Job Description

We are implementing a Media Mix Optimization (MMO) platform designed to analyze and optimize marketing investments across multiple channels. This initiative requires a robust on-premises data infrastructure to support distributed computing, large-scale data ingestion, and advanced analytics. The Data Engineer will be responsible for building and maintaining resilient pipelines and data systems that feed into MMO models, ensuring data quality, governance, and availability for Data Science and BI teams. The environment integrates HDFS for distributed storage, Apache NiFi for orchestration, Hive and PySpark for distributed processing, and Postgres for structured data management.

This role is central to enabling seamless integration of massive datasets from disparate sources (media, campaign, transaction, customer interaction, etc.), standardizing data, and providing reliable foundations for advanced econometric modeling and insights.

Responsibilities:

 

Data Pipeline Development & Orchestration
o Design, build, and optimize scalable data pipelines in Apache NiFi to

automate ingestion, cleansing, and enrichment from structured, semi-structured, and unstructured sources.

Ensure pipelines meet low-latency and high-throughput requirements for distributed processing.

Data Storage & Processing
o Architect and manage datasets on HDFS to support high-volume,

fault-tolerant storage.
o Develop distributed processing workflows in PySpark and Hive to

handle large-scale transformations, aggregations, and joins across

petabyte-level datasets.
o Implement partitioning, bucketing, and indexing strategies to

optimize query performance.

Database Engineering & Management
o Maintain and tune Postgres databases for high availability, integrity,

and performance.
o Write advanced SQL queries for ETL, analysis, and integration with

downstream BI/analytics systems.

Collaboration & Integration
o Partner with Data Scientists to deliver clean, reliable datasets for

model training and MMO analysis.
o Work with BI engineers to ensure data pipelines align with reporting

and visualization requirements.

Monitoring & Reliability Engineering
o Implement monitoring, logging, and alerting frameworks to track

data pipeline health.
o Troubleshoot and resolve issues in ingestion, transformations, and

distributed jobs.

Data Governance & Compliance
o Enforce standards for data quality, lineage, and security across

systems.
o Ensure compliance with internal governance and external

regulations.

Documentation & Knowledge Transfer
o Develop and maintain comprehensive technical documentation for

pipelines, data models, and workflows.
o Provide knowledge sharing and onboarding support for cross-

functional teams.

 

Qualifications

  • Bachelor’s degree in Computer Science, Information Technology, or related field (Master’s preferred).

  • Proven experience as a Data Engineer with expertise in HDFS, Apache NiFi, Hive, PySpark, Postgres, Python, and SQL.

  • Strong background in ETL/ELT design, distributed processing, and relational database management.

  • Experience with on-premises big data ecosystems supporting distributed computing.

  • Solid debugging, optimization, and performance tuning skills.

  • Ability to work in agile environments, collaborating with multi-disciplinary

    teams.

  • Strong communication skills for cross-functional technical discussions.

    Preferred Qualifications:

  • Familiarity with data governance frameworks, lineage tracking, and data cataloging tools.

  • Knowledge of security standards, encryption, and access control in on- premises environments.

  • Prior experience with Media Mix Modeling (MMM/MMO) or marketing analytics projects.

  • Exposure to workflow schedulers (Airflow, Oozie, or similar).

  • Proficiency in developing automation scripts and frameworks in Python for

    CI/CD of data pipelines.

Top Skills

Apache Nifi
Hdfs
Hive
Postgres
Pyspark
Python
SQL
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: Columbia, MD
390 Employees
Year Founded: 2016

What We Do

Our Vision is to build a company of world-class people that helps our clients optimize business performance through data, technology and analytics. Blend360 has two divisions: Data Science Solutions: We work at the intersection of data, technology and analytics. Talent Solutions: We live and breathe the digital and talent marketplace.

Similar Jobs

MetLife Logo MetLife

Big Data Engineer

Fintech • Information Technology • Insurance • Financial Services • Big Data Analytics
Remote or Hybrid
India
43000 Employees

Kroll Logo Kroll

Data Engineer

Big Data • Security • Software • Analytics • Cybersecurity
Remote or Hybrid
India
5001 Employees

Zoom Logo Zoom

Data Engineer

Artificial Intelligence • Information Technology • Software
Remote
IND
11053 Employees
6-6 Annually

Cytel Logo Cytel

Data Engineer

Software • Analytics • Biotech
In-Office or Remote
2 Locations
1395 Employees

Similar Companies Hiring

Northslope Technologies Thumbnail
Software • Information Technology • Generative AI • Consulting • Artificial Intelligence • Analytics
Denver, CO
88 Employees
Scotch Thumbnail
Software • Retail • Payments • Fintech • eCommerce • Artificial Intelligence • Analytics
US
25 Employees
Milestone Systems Thumbnail
Software • Security • Other • Big Data Analytics • Artificial Intelligence • Analytics
Lake Oswego, OR
1500 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account