Senior Data Engineer

Reposted 25 Days Ago
Be an Early Applicant
Bengaluru, Bengaluru Urban, Karnataka
In-Office
Senior level
Information Technology • Consulting
The Role
The Data Engineer will design and maintain cloud-based data lakes and pipelines on AWS, utilizing Databricks and Python for data processing, while collaborating with teams to optimize performance and manage data architecture.
Summary Generated by Built In
Senior Data Engineer

Req number:

R6655

Employment type:

Full time

Worksite flexibility:

RemoteWho we are

CAI is a global technology services firm with over 8,500 associates worldwide and a yearly revenue of $1 billion+. We have over 40 years of excellence in uniting talent and technology to power the possible for our clients, colleagues, and communities. As a privately held company, we have the freedom and focus to do what is right—whatever it takes. Our tailor-made solutions create lasting results across the public and commercial sectors, and we are trailblazers in bringing neurodiversity to the enterprise.

Job Summary

We are looking for a motivated Data Engineer ready to take us to the next level! If you have strong experience in building cloud-based data lake and analytics architectures using AWS and Databricks, and is proficient in Python programming for data processing and automation and are looking for your next career move, apply now.

Job Description

We are looking for a Data Engineer that has experience in building data products using Databricks and related technologies. This position will be Full-time and Remote position.What You’ll Do
  • Design, develop, and maintain data lakes and data pipelines on AWS using ETL frameworks and Databricks.
  • Integrate and transform large-scale data from multiple heterogeneous sources into a centralized data lake environment.
  • Implement and manage Delta Lake architecture using Databricks Delta or Apache Hudi.
  • Develop end-to-end data workflows using PySpark, Databricks Notebooks, and Python scripts for ingestion, transformation, and enrichment.
  • Design and develop data warehouses and data marts for analytical workloads using Snowflake, Redshift, or similar systems.
  • Design and evaluate data models (Star, Snowflake, Flattened) for analytical and transactional systems.
  • Optimize data storage, query performance, and cost across the AWS and Databricks ecosystem.
  • Build and maintain CI/CD pipelines for Databricks notebooks, jobs, and Python-based data processing scripts.
  • Collaborate with data scientists, analysts, and stakeholders to deliver high-performance, reusable data assets.
  • Maintain and manage code repositories (Git) and promote best practices in version control, testing, and deployment.
  • Participate in making major technical and architectural decisions for data engineering initiatives.
  • Monitor and troubleshoot Databricks clusters, Spark jobs, and ETL processes for performance and reliability.
  • Coordinate with business and technical teams through all phases of the software development life cycle.
What You'll NeedRequired
  • 5+ years of experience building and managing Data Lake Architecture on AWS Cloud
  • 3+ years of experience with AWS Data services such as S3, Glue, Lake Formation, EMR, Kinesis, RDS, DMS, and Redshift.
  • 3+ years of experience building Data Warehouses on Snowflake, Redshift, HANA, Teradata, or Exasol.
  • 3+ years of hands-on experience working with Apache Spark or PySpark, on Databricks.
  • 3+ years of experience implementing Delta Lakes using Databricks Delta or Apache Hudi.
  • 3+ years of experience in ETL development using Databricks, AWS Glue, or other modern frameworks.
  • Proficiency in Python for data engineering, automation, and API integrations.
  • Experience in Databricks Jobs, Workflows, and Cluster Management.
  • Experience with CI/CD pipelines and Infrastructure as Code (IaC) tools like Terraform or CloudFormation is a plus.
  • Bachelor’s degree in computer science, Information Technology, Data Science, or related field.
  • Experience working on Agile projects and methodology in general.

Preferred
  • Strong SQL, RDBMS, and data modeling skills.
  • Experience with Databricks Unity Catalog, Delta Live Tables (DLT), and MLflow for data governance and model lifecycle.
  • AWS or Databricks Cloud Certifications (e.g., AWS Data Analytics Specialty, Databricks Certified Data Engineer Professional) are a big plus.
  • Understanding data security, access control, and compliance in cloud environments.
  • Strong analytical, problem-solving, and communication skills.
Physical Demands
  • This role involves mostly sedentary work, with occasional movement around the office to attend meetings, etc.
  • Ability to perform repetitive tasks on a computer, using a mouse, keyboard, and monitor.

Reasonable accommodation statement

If you require a reasonable accommodation in completing this application, interviewing, completing any pre-employment testing, or otherwise participating in the employment selection process, please direct your inquiries to [email protected] or (888) 824 – 8111.

Top Skills

Apache Hudi
AWS
Ci/Cd
CloudFormation
Databricks
Delta Lake
ETL
Pyspark
Python
Redshift
Snowflake
SQL
Terraform
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
Chicago, , Illinois
2,689 Employees

What We Do

CAI is a global services firm with over 8,700 associates worldwide and a yearly revenue of $1 billion+. We have over 40 years of excellence in uniting talent and technology to power the possible for our clients, colleagues, and communities. As a privately held company, we have the freedom and focus to do what’s right—whatever it takes. Our tailor-made solutions create lasting results across the public and commercial sectors, and we are trailblazers in bringing neurodiversity to the enterprise

Similar Jobs

CrowdStrike Logo CrowdStrike

Senior Software Engineer

Cloud • Computer Vision • Information Technology • Sales • Security • Cybersecurity
Remote or Hybrid
KA, IND
10000 Employees

Allica Bank Logo Allica Bank

Technical Incident Lead

Fintech • Software • Financial Services
In-Office or Remote
2 Locations
502 Employees

Elsevier Logo Elsevier

Account Manager

Artificial Intelligence • Healthtech • Information Technology • Other • Analytics
In-Office or Remote
3 Locations

Allica Bank Logo Allica Bank

Full-stack Engineer

Fintech • Software • Financial Services
In-Office or Remote
2 Locations
502 Employees

Similar Companies Hiring

Scrunch AI Thumbnail
Software • SEO • Marketing Tech • Information Technology • Artificial Intelligence
Salt Lake City, Utah
Amplify Platform Thumbnail
Fintech • Financial Services • Consulting • Cloud • Business Intelligence • Big Data Analytics
Scottsdale, AZ
62 Employees
Standard Template Labs Thumbnail
Software • Information Technology • Artificial Intelligence
New York, NY
15 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account