Databricks Consultant

Reposted 11 Days Ago
Be an Early Applicant
Mumbai, Maharashtra, IND
Hybrid
Senior level
Database
The Role
Develop scalable ETL/ELT pipelines in Databricks, optimize Delta Lake tables, implement data governance, performance tuning, and collaborate with teams to ensure efficient data workflows.
Summary Generated by Built In

Job Title: Senior Associate Developer - Databricks, PySpark, and Spark SQL

Education: Any Graduate

Experience: 5+years

Location: Mumbai

 

Key Skills:

  • Strong hands-on experience with Databricks, PySpark, and Spark SQL.

  • Expertise in Delta Lake, Bronze–Silver–Gold architecture, and Lakehouse patterns.

  • Strong experience with cloud platforms (AWS/Azure/GCP).

  • Solid understanding of data warehousing, dimensional modeling, and bigdata concepts.

 

Job Description:

  • Build scalable ETL/ELT pipelines using Databricks (PySpark, SQL, Spark Streaming).

  • Develop and optimize Delta Lake tables, ACID transactions, schema evolution, and time travel.

  • Implement Unity Catalog, data governance, and access control.Optimize cluster configurations, job workflows, and performance tuning in Databricks.

  • Design and implement batch and streaming pipelines using Spark Structured Streaming.

  • Integrate Databricks with multiple data sources (RDBMS, APIs, cloud storage, message queues).Develop reusable, modular, and automated data processing frameworks.

  • Implement CI/CD pipelines for Databricks using GitHub Actions / Azure DevOps / GitLab.Automate cluster management and job orchestration using Databricks REST APIs.

  • Maintain code quality, unit tests, and documentation. 

  • Write and optimize complex SQL queries and statements to ensure high performance and efficient data retrieval.

  • Strong database design including normalization, data modelling, and relational schema creation.

  • Conduct performance analysis, troubleshoot database issues like slow queries or deadlocks and implement solutions

  • Design and implement database structures, including tables, schemas, views, stored procedures, functions, and triggers.

  • Optimize database performance through query tuning, indexing, and performance analysis.

  • Ensure data integrity, security, and compliance standards

  • Need strong Python skills combined with expertise in Apache Spark for large scale data processing. Core abilities include building efficient ETL pipelines, optimizing distributed jobs, and handling large-scale data transformations

  • Expertise in Python programming, Spark APIs, and parallel processing.

  • Proficiency in Python (including Pandas, NumPy) for data manipulation and scripting

  • Deep knowledge of PySpark APIs like DataFrames, RDDs, Spark SQL for querying and processing.

  • Familiarity with RESTful APIs, batch processing, CI/CD, and monitoring data jobs.

  • Optimize Spark jobs for performance, troubleshoot issues, and ensure data quality across systems.

  • Collaborate with data engineers and scientists to implement workflows, conduct code reviews, and integrate with cloud platforms like AWS or Azure.

  • Design, develop, and maintain scalable data pipelines and ETL processes using Azure Databricks

  • Build data transformation workflows using Python or Scala.

  • Work with data lakes using Delta Lake.

  • Integrate data from multiple sources such as APIs, databases, and cloud storage.

  • Monitor and optimize data workflows for performance and reliability.

  • Collaborate with data scientists, analysts, and business teams.

 

About UsDatavail is a leading provider of data management, application development, analytics, and cloud services, with more than 1,000 professionals helping clients build and manage applications and data via a world-class tech-enabled delivery platform and software solutions across all leading technologies. For more than 17 years, Datavail has worked with thousands of companies spanning different industries and sizes, and is an AWS Advanced Tier Consulting Partner, a Microsoft Solutions Partner for Data & AI and Digital & App Innovation (Azure), an Oracle Partner, and a MySQL Partner. About the Team
Datavail’s Data Management Services: 
Datavail’s Data Management and Analytics practice is made up of experts who provide a variety of data services including initial consulting and development, designing and building complete data systems, as well as ongoing support and management of database, data warehouse, data lake, data integration, and virtualization and reporting environments. Datavail’s team is comprised of not just excellent BI & analytics consultants, but great people as well. Datavail’s data intelligence consultants are experienced, knowledgeable and certified in the best in breed BI and analytics software applications and technologies. We ascertain your business objectives, goals and requirements, assess your environment, and recommend the tools which best fit your unique situation. Our proven methodology can help your project succeed, regardless of stage. With the combination of a proven delivery model and top-notch experience ensures that Datavail will remain the Data Management experts on demand you desire. Datavail’s flexible and client focused services always add value to your organization.
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: Broomfield, CO
263 Employees
Year Founded: 2007

What We Do

A premiere data services company serving clients in North America, Datavail has 1,000 data professionals, data engineers, developers, project managers, consultants, and business experts, supported by industry-leading automation and intellectual property. For more than 17 years, Datavail has worked with thousands of companies spanning different industries and sizes. At Datavail, we look for more than smarts, experience and proficiency. On top of those requirements, we seek people who mesh with our corporate values. We seek brilliance without bravado and know-how without a know-it-all attitude. We hold low ego in high regard, embrace problem-solving as a passion and welcome every day as a new opportunity to learn. We’re flexible and hard working. We’re committed to our clients and colleagues. We help our people grow so they can help our clients grow. That makes us grow so we can help even more customers leverage organizational data for business value. Our Core Values: 1. We desire to serve. 2. We embody flexibility for availability 3. We exemplify low ego. 4. We work hard. 5. We strive for continuous improvement. 6. We are growth-oriented.

Similar Jobs

Rapid7 Logo Rapid7

Senior Salesforce Engineer

Artificial Intelligence • Cloud • Information Technology • Sales • Security • Software • Cybersecurity
Remote or Hybrid
Pune, Maharashtra, IND
2400 Employees

Rapid7 Logo Rapid7

Artificial Intelligence Engineer

Artificial Intelligence • Cloud • Information Technology • Sales • Security • Software • Cybersecurity
Remote or Hybrid
Pune, Maharashtra, IND
2400 Employees

Rapid7 Logo Rapid7

Technical Support

Artificial Intelligence • Cloud • Information Technology • Sales • Security • Software • Cybersecurity
Remote or Hybrid
Pune, Maharashtra, IND
2400 Employees

Morningstar Logo Morningstar

Consultant

Artificial Intelligence • Big Data • Enterprise Web • Fintech • Software • Financial Services
Hybrid
Navi Mumbai, Thane, Maharashtra, IND
11500 Employees

Similar Companies Hiring

Apollo.io Thumbnail
Software • Sales • Productivity • Information Technology • Enterprise Web • Database • Artificial Intelligence
US
850 Employees
Perchwell Thumbnail
Mobile • Real Estate • Software • Database • Analytics
New York City, NY
60 Employees
Jellyfish Thumbnail
Big Data • Cloud • Productivity • Software • Database • Analytics • Automation
Boston, MA
225 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account