Data Engineer (Financial Analytics focus)
Capital
Posted: March 18, 2026
Interested in this position?
Create a free account to apply with AI-powered matching
Quick Summary
Develop and maintain data pipelines, transformation logic, and data quality checks for a complex multi-jurisdiction financial data platform.
Required Skills
Job Description
We are a leading trading platform that is ambitiously expanding to the four corners of the globe. Our top-rated products have won prestigious industry awards for their cutting-edge technology and seamless client experience. We deliver only the best, so we are always in search of the best people to join our ever-growing talented team.
We are seeking a Data Engineer to join our Financial Analytics team developing and maintaining data pipelines, transformation logic, and data quality checks in a complex multi-jurisdiction financial data platform (PostgreSQL DWH + Airflow orchestration).
Responsabilities:
• Implement enhancements and changes to existing reporting processes to improve accuracy, performance, and usability
• Design and develop new reporting pipelines and datasets aligned with business requirements
• Automate of data delivery processes
• Identify and implementation of automated data quality checks
• Resolve of issues related to data quality
• Collaborate with business stakeholders to gather reporting requirements, clarify logic, and ensure outputs meet expectations
Requirements:
• 4+ years in analytics engineering or similar data-focused roles
• Advanced PostgreSQL: stored procedures and functions, complex CTEs, window functions, SCD2 patterns, query plan analysis and optimisation
• Strong understanding of data warehouse architecture: staging, core, and data mart layers; incremental load patterns; slowly changing dimensions
• Hands-on experience with Apache Airflow: DAG authoring, scheduling, dependency management, and failure handling
• Proficiency with Git (GitLab or GitHub) and JIRA
• Experience designing and evolving data warehouse architecture and data models
• Track record of building robust, maintainable ELT/ETL pipelines in production
• Experience implementing automated data quality checks
• Domain fluency in financial and trading concepts, with the ability to understand requirements and clearly explain implemented logic to business stakeholders
• High degree of autonomy: able to reverse-engineer undocumented systems, identify root causes, and take end-to-end ownership of pipelines and calculation logic
• Comfortable using AI-assisted development tools (e.g., Claude, Copilot, Cursor) to improve productivity
Nice to have:
• Hands-on experience with dbt, particularly in the context of migration or adoption initiatives
• Exposure to Snowflake or strong interest in working with it as part of a target data architecture
• Proficiency in Python for scripting, automation, and data pipeline tooling
• Background in fintech or financial services in any capacity
What you will get in return:
• Competitive compensation
• A generous paid leave policy, supporting a healthy work–life balance
• Two additional paid days per year dedicated to volunteering and giving back
• Private medical insurance for your peace of mind
• An additional flexible benefits budget, allowing you to tailor benefits to your needs
• Flexible working mode
• The opportunity to work from almost anywhere in the world for up to 30 days per year
• Annual company-wide events held in locations around the globe
• In-office massages to support wellbeing
Be a key player at the forefront of the digital assets movement, propelling your career to new heights! Join a dynamic and rapidly expanding company that values and rewards talent, initiative, and creativity. Work alongside one of the most brilliant teams in the industry.