At Caseware, we are evolving our cloud platform to deliver intelligent, data-driven experiences that help our customers work smarter, faster, and with greater confidence. As we expand Caseware Cloud with advanced analytics capabilities, we are investing in a modern, scalable data platform that serves as the foundation for insight, reporting, interoperability, and trust.
As a Senior Software Developer - Data Engineer, you will play a key role in shaping and operating the data infrastructure that powers Caseware Cloud. You will design and build reliable data pipelines that move data from our transactional systems into our data lake and Gold-tier analytics datasets, enable real-time and event-driven data flows, and help evolve our lakehouse architecture to support secure, standards-based data sharing with external consumers.
In this role, you will take ownership of complex data engineering solutions end-to-end, influence architectural direction through technical leadership and proof-of-concepts, and help ensure our data platform is secure, scalable, and operationally excellent. You will collaborate closely with platform, DevOps, and product teams to deliver durable data capabilities that directly impact customers across Caseware’s cloud ecosystem.
📍 Location: This is a fully remote position located in Colombia.
Contact
Maira Russo - Senior Talent Acquisition Partner
What you will be doing:
Design, build, and operate scalable, highly available ETL/ELT pipelines that move data from OLTP systems into the data lake and Gold-tier analytics tables, optimized for reliability, performance, and cost efficiency.
Design and evolve data lake and lakehouse foundations, including storage layout, partitioning strategies, schema management, lineage tracking, data cataloging, versioning, and delta synchronization to ensure data is reliable, analytics-ready, and consumable through open table standards.
Build and operate Apache Iceberg-based datasets and sharing patterns for external consumers, including row-level change data capture, incremental sync workflows, and companion change tables for downstream consumption.
Implement consumer-facing data sharing capabilities using AWS-native catalog and access patterns, enabling secure integration with external analytics platforms and open ecosystem tools.
Design and operate orchestration workflows for batch publication, watermarking, dependency-aware release of shared datasets, and engine-agnostic processing across Athena and Glue.
Apply data quality, validation, monitoring, and testing frameworks in production pipelines, ensuring governance, tenant isolation, access control, lineage, and security standards are met.
Monitor, troubleshoot, and optimize data pipelines and lakehouse workflows for reliability, performance, scale, and cost, including query concurrency and compute-engine tradeoffs.
Lead pragmatic platform evolution by defining clear data contracts, reducing coupling between systems, and improving operability and developer experience for data workflows.
Provide technical mentorship to junior engineers through code reviews, pair programming, and collaborative solution design.
Drive operational excellence by identifying recurring issues, eliminating root causes, improving system reliability, and participating in a production support rotation with thorough post-mortems and root cause analysis.
What you’ll bring:
- 5+ years of experience in data engineering or software engineering, with at least 3 years designing and operating cloud-based AWS data platforms.
- Demonstrated experience building and owning end-to-end ETL/ELT pipelines, including ingestion from OLTP systems into analytics-ready data lakes or lakehouse architectures.
- Experience designing and operating production-grade data systems for customer-facing SaaS applications with high availability, reliability, and operational rigor.
- Hands-on experience with modern table formats or lakehouse patterns, such as Apache Iceberg, including schema evolution, partitioning, versioning, and metadata-driven access.
- Experience building incremental data processing and CDC-style workflows, including watermarking, replay safety, and downstream synchronization patterns.
- Experience with AWS-native analytics and lake technologies, especially S3, Athena, Glue, Glue Catalog, and Lake Formation.
- Familiarity with distributed data processing and compute tradeoffs across serverless query engines and Spark-based pipelines.
- Experience implementing secure data access patterns, including tenant isolation, scoped credentials, role-based access, and governance controls.
- Experience enabling external or cross-platform consumption of data through open standards, shared catalogs, or interoperable data access patterns.
- Strong software engineering fundamentals, including designing maintainable, testable systems and participating across the full software development lifecycle.
- Experience operating data systems in production, including monitoring, incident response, post-mortems, and continuous improvement.
- Proven ability to mentor junior engineers and contribute to technical leadership through code reviews and collaborative design.
- Strong English language communication and collaboration skills.
- Operate with an AI-first mindset, leveraging tools to enhance productivity, decision-making, and day-to-day work.
- Champion a culture of AI experimentation — proactively explore, adopt, and share emerging tools relevant to your role.
- Integrate AI tools into daily workflows to automate repetitive tasks, accelerate outputs, and improve overall efficiency.
- Use AI-assisted tools to enhance drafting, summarization, research, and data analysis tasks.
- Apply critical thinking to validate and refine AI-generated outputs, ensuring accuracy, relevance, and alignment with quality standards.
- Recognize the limitations of AI tools and exercise independent judgment on all AI-assisted work before sharing or acting on results.
- Continuously build AI fluency — including foundational prompt engineering skills — to maximize the quality and reliability of AI outputs.
Nice to have
Experience building platforms in regulated or compliance-heavy domains.
Familiarity with accounting, auditing, or financial workflows.
Experience designing multi-tenant data sharing platforms or customer-facing analytics data products.
Experience with orchestration using Step Functions and EventBridge.
The Tech Stack You’ll Work With:
Backend & Platform: Java (21+) Spring microservices, Python 3, AWS Lambda
Data Lake & Table Formats: Amazon S3, Apache Iceberg, S3 Tables
Data Processing & Query: AWS Athena, AWS Glue, Glue Spark
Metadata & Catalog: AWS Glue Data Catalog, Glue Iceberg REST Catalog
Governance & Security: AWS Lake Formation, IAM, STS
Orchestration & Control Plane: AWS Step Functions, Amazon EventBridge, DynamoDB
Messaging & Integration: AWS SNS/SQS
Developer Tooling: GitHub, GitHub Actions, Nx Monorepo, IntelliJ IDEA
Collaboration: Atlassian Cloud (Jira/Confluence), Microsoft Teams, Outlook, Microsoft
Perks & Benefits
- ¨Contrato a termino Indefinido¨ with all the legal benefits
- Prepaid Medicine
- Life insurance and funeral assistance
- Internet allowance
- Home office stipend
- Competitive compensation — above the market average
- 100% remote work environment and an excellent work-life balance
- Opportunity to work for a growing global SaaS leader company
- A culture that promotes independence, innovation, trust, and accountability
- Open space to be creative, innovative and strategize for the future
- Mentorship by highly experienced professional
- Budget for training, we want you to grow
- 5 Personal Time Off days per year
- Sick Leave Top up to total 100% of salary paid by the employer from Day 3 to 90.
- Recognition Award, additional paid time off in recognition of the corresponding year of service
- Upgrade vacation starting at 5 years of service
What We Do
Caseware is the leading global provider of cloud-enabled audit, financial reporting and data analytics solutions for accounting firms, corporations and government regulators. Caseware’s innovative tools and platforms help more than half a million customers in 130 countries work smarter, dig deeper and see further as they transform insights into impact.

.png)







