We are seeking a highly skilled Senior Data Engineer to join our Data Engineering team. The ideal candidate will be responsible for designing, building, and optimizing scalable data pipelines and enabling advanced analytics and machine learning use cases. This role requires strong expertise in big data technologies, data processing frameworks, and feature engineering to support ML models in a production environment.
Key Responsibilities- Gather and analyze business and technical requirements for data solutions
- Perform Exploratory Data Analysis (EDA) to understand data patterns and quality
- Design, build, and maintain scalable and high-performance data pipelines
- Ingest structured and unstructured data from multiple sources
- Transform and process large datasets using PySpark and Python
- Implement feature engineering techniques for machine learning models
- Optimize Spark jobs for performance and cost efficiency
- Ensure data quality, integrity, and security across pipelines
- Collaborate with Data Scientists, Analytics Teams, and Delivery Leads
- Work in an Agile environment with cross-functional teams
- Communicate effectively with stakeholders and provide technical insights
- Maintain version control and CI/CD workflows using Git
- Strong experience in Python and PySpark
- Hands-on expertise in Apache Spark (performance tuning & optimization)
- Solid knowledge of Hadoop ecosystem
- Advanced proficiency in SQL
- Experience in building end-to-end data pipelines and ML pipelines
- Strong understanding of data modeling and data warehousing concepts
- Experience with feature engineering for ML use cases
- Familiarity with Git and version control systems
- Exposure to cloud platforms (AWS/Azure/GCP) is a plus
- Bachelor’s or Master’s degree in Computer Science, Engineering, or related field
- 8–10+ years of experience in Data Engineering / Big Data
- Prior experience working in banking or financial services is preferred
- Strong problem-solving and analytical thinking
- Excellent communication and stakeholder management skills
- Ability to work independently and in a team environment
- Adaptability to fast-paced Agile workflows
Skills Required
- Strong experience in Python and PySpark
- Hands-on expertise in Apache Spark
- Solid knowledge of Hadoop ecosystem
- Advanced proficiency in SQL
- Experience in building end-to-end data pipelines and ML pipelines
- Strong understanding of data modeling and data warehousing concepts
- Experience with feature engineering for ML use cases
- Familiarity with Git and version control systems
- Exposure to cloud platforms (AWS/Azure/GCP)
What We Do
Global Software Solutions Group Veracious product line is a series of robust banking platforms that provide core banking, payment systems, custom process automation, and document management solutions for banks and financial institutions in Middle East & Africa. This cutting-edge product line features the Veracious Payments Hub, Digital Banking and the DMS, all built on the Torus Lowcode development platform software. Global Software Solutions Group is a software solutions provider that aims to solve mission-critical problems that financial institutions face today. Our software solutions bring together our Low Code platform, the payments product line and customized service offerings to solve mission-critical statements in core banking, payments, process automation, and document management. The Payments Hub is GSS's flagship product.








