About Us
DeepSea is a maritime technology leader, providing a unified optimisation platform that helps shipping companies cut fuel consumption, reduce emissions, and transform vessel performance across entire fleets.
Built on the most advanced maritime AI data pool in the world, DeepSea’s platform powers everything from vessel monitoring to advanced automation - helping clients achieve measurable results with clarity, speed and scale.
We are trusted by some of the world’s most forward-thinking ship owners and operators, and our team spans Greece, the UK, Armenia, Romania, Singapore and Japan. As the shipping industry undergoes enormous change, we are at the centre of that transition - helping our customers take control, reduce risk, and build competitive advantage.
Main Responsibilities
Design, build, and maintain the data infrastructure and pipelines to support AI model training and analytics.
Help us migrate redundant legacy systems into a structured, high-performance environment where data integrity is the baseline.
Have a "cost-first" mindset, prioritizing efficient code and smart joins over expensive resource scaling.
Fine-tune our performance at the storage level, focusing on data layout and engine optimization rather than simply scaling hardware.
Design and implement high-throughput streaming architectures capable of processing granular, event-level data in real-time while maintaining low latency and high availability.
Architect the full data lifecycle, ensuring everything we build is resilient for the source and optimized for the end user.
Collaborate with cross-functional teams (Architects, Software Engineers, AI specialists) to design data solutions based on business needs and best practices.
Main Requirements
Bachelor’s or Master’s degree in Engineering, Mathematics, Physics, Computer Science or related field.
Proven expertise in designing, developing, and rolling out production scalable data processing systems for ML/AI and analytics workloads, preferably using Databricks or similar.
Strong understanding of relational and non-relational database design, including normalisation, indexing, and partitioning strategies.
Expertise in schema modeling for structured and semi-structured data (e.g., Delta Lake, Parquet, JSON).
Strong understanding of data security, compliance, and governance, including data privacy regulations, secure data handling, auditability, and enterprise-grade data management.
Hands-on expertise with Spark-based data engineering, Delta Lake, warehouses, ETL/ELT data integration, MLflow integration, collaborative notebooks, and scalable model deployment.
Ability to optimize Lakehouse performance using Delta Lake features
Solid grasp of data warehousing concepts and performance tuning in cloud environments (e.g., Azure, AWS).
Programming skills in Python, SQL, Pandas, PySpark, NumPy, SciPy or similar frameworks.
Experience in mentoring cross-functional teams.
Strong communication and collaboration skills in an agile, dynamic and cross-disciplinary environment.
Nice to Have
Experience in Databricks Structured Streaming.
Experience in using Kafka and Flink or similar.
Familiarity with Databricks on AWS, including integration with S3, APIs, and secure data access patterns.
Knowledge of Dimensional Modeling (Star/Snowflake schemas) for BI and reporting use cases.
What we offer:
You will be part of a dynamic team focused on delivering results and continuous improvement while disrupting the industry, and will also receive:
Competitive remuneration package: Skill & experience-based salary and eligibility for additional employee benefits.
Health Package: Private health insurance coverage and mental health - therapist benefit
Paid leaves: emergency and medical leaves.
Learning and Development Package: Access to courses platform and eligibility for seminars, conferences and workshops
Remote Flexibility: Great office space in the heart of Athens with hybrid option.
At DeepSea, we are looking for people who share our values and are aligned with our mission. It is important to us to ensure that no-one who is eager and capable of contributing constructively to our team is excluded because of ethnic or social origin, gender or sexuality, age or family status, disability or medical conditions etc. Diversity is well-proven to be a vital characteristic of teams that succeed, so we do everything we can to make our environment welcoming and safe for everyone.
Data Privacy
The company ensures that the personal data of candidates is handled with care and in compliance with GDPR regulations. Your personal data will be stored securely and only for the duration necessary under the law. If the recruitment process is unsuccessful, your data will be retained for 5 (five) years to consider you for future opportunities. After this period, your data will be deleted or destroyed. In case you do not wish for the company to keep your CV and personal data, please send an email to [email protected].
Please mention you found this job on AI Jobs. It helps us get more startups to hire on our site. Thanks and good luck!
Understand the required skills and qualifications, anticipate the questions you may be asked, and study well-prepared answers using our sample responses.
Senior Data Engineer Q&A's