Senior Software Developer
TLDR
Shape and operate the data infrastructure for Caseware Cloud, designing scalable ETL/ELT pipelines and ensuring secure data sharing while leading technical direction.
-
Design, build, and operate scalable, highly available ETL/ELT pipelines that move data from OLTP systems into the data lake and Gold-tier analytics tables, optimized for reliability, performance, and cost efficiency.
-
Design and evolve data lake and lakehouse foundations, including storage layout, partitioning strategies, schema management, lineage tracking, data cataloging, versioning, and delta synchronization to ensure data is reliable, analytics-ready, and consumable through open table standards.
-
Build and operate Apache Iceberg-based datasets and sharing patterns for external consumers, including row-level change data capture, incremental sync workflows, and companion change tables for downstream consumption.
-
Implement consumer-facing data sharing capabilities using AWS-native catalog and access patterns, enabling secure integration with external analytics platforms and open ecosystem tools.
-
Design and operate orchestration workflows for batch publication, watermarking, dependency-aware release of shared datasets, and engine-agnostic processing across Athena and Glue.
-
Apply data quality, validation, monitoring, and testing frameworks in production pipelines, ensuring governance, tenant isolation, access control, lineage, and security standards are met.
-
Monitor, troubleshoot, and optimize data pipelines and lakehouse workflows for reliability, performance, scale, and cost, including query concurrency and compute-engine tradeoffs.
-
Lead pragmatic platform evolution by defining clear data contracts, reducing coupling between systems, and improving operability and developer experience for data workflows.
-
Provide technical mentorship to junior engineers through code reviews, pair programming, and collaborative solution design.
-
Drive operational excellence by identifying recurring issues, eliminating root causes, improving system reliability, and participating in a production support rotation with thorough post-mortems and root cause analysis.
-
5+ years of experience in data engineering or software engineering, with at least 3 years designing and operating cloud-based AWS data platforms.
-
Demonstrated experience building and owning end-to-end ETL/ELT pipelines, including ingestion from OLTP systems into analytics-ready data lakes or lakehouse architectures.
-
Experience designing and operating production-grade data systems for customer-facing SaaS applications with high availability, reliability, and operational rigor.
-
Hands-on experience with modern table formats or lakehouse patterns, such as Apache Iceberg, including schema evolution, partitioning, versioning, and metadata-driven access.
-
Experience building incremental data processing and CDC-style workflows, including watermarking, replay safety, and downstream synchronization patterns.
-
Experience with AWS-native analytics and lake technologies, especially S3, Athena, Glue, Glue Catalog, and Lake Formation.
-
Familiarity with distributed data processing and compute tradeoffs across serverless query engines and Spark-based pipelines.
-
Experience implementing secure data access patterns, including tenant isolation, scoped credentials, role-based access, and governance controls.
-
Experience enabling external or cross-platform consumption of data through open standards, shared catalogs, or interoperable data access patterns.
-
Strong software engineering fundamentals, including designing maintainable, testable systems and participating across the full software development lifecycle.
-
Experience operating data systems in production, including monitoring, incident response, post-mortems, and continuous improvement.
-
Proven ability to mentor junior engineers and contribute to technical leadership through code reviews and collaborative design.
-
Strong English language communication and collaboration skills.
-
Experience building platforms in regulated or compliance-heavy domains.
-
Familiarity with accounting, auditing, or financial workflows.
-
Experience designing multi-tenant data sharing platforms or customer-facing analytics data products.
-
Experience with orchestration using Step Functions and EventBridge.
-
Backend & Platform: Java (21+) Spring microservices, Python 3, AWS Lambda
-
Data Lake & Table Formats: Amazon S3, Apache Iceberg, S3 Tables
-
Data Processing & Query: AWS Athena, AWS Glue, Glue Spark
-
Metadata & Catalog: AWS Glue Data Catalog, Glue Iceberg REST Catalog
-
Governance & Security: AWS Lake Formation, IAM, STS
-
Orchestration & Control Plane: AWS Step Functions, Amazon EventBridge, DynamoDB
-
Messaging & Integration: AWS SNS/SQS
-
Developer Tooling: GitHub, GitHub Actions, Nx Monorepo, IntelliJ IDEA
-
Collaboration: Atlassian Cloud (Jira/Confluence), Microsoft Teams, Outlook, Microsoft
Benefits
Health Insurance
Prepaid Medicine
Home Office Stipend
Training budget
Budget for training, we want you to grow
Sick leave top-up
Sick Leave Top up to total 100% of salary paid by the employer from Day 3 to 90.
CaseWare develops sophisticated software solutions tailored for accounting firms, corporations, and government entities. With a global reach and over 30 years of expertise, we empower users to optimize audits and financial reporting, transforming complex data into actionable insights.
- Founded
- Founded 1988
- Employees
- 201-500 employees
- Industry
- Internet Software & Services