Senior Data Quality Engineer
Bridgenext
Posted: March 2, 2026
Interested in this position?
Create a free account to apply with AI-powered matching
Quick Summary
We're seeking a Senior Data Quality Engineer to join our team in our global consulting and delivery arms, where we help clients innovate with intention and realize their digital aspirations.
Required Skills
Job Description
Company Overview
Bridgenext is a digital consulting services leader that helps clients innovate with intention and realize their digital aspirations by creating digital products, experiences, and solutions around what real people need. Our global consulting and delivery teams facilitate highly strategic digital initiatives through digital product engineering, automation, data engineering, and infrastructure modernization services, while elevating brands through digital experience, creative content, and customer data analytics services.
Don't just work, thrive. At Bridgenext, you have an opportunity to make a real difference - driving tangible business value for clients, while simultaneously propelling your own career growth. Our flexible and inclusive work culture provides you with the autonomy, resources, and opportunities to succeed.
Position Description
Bridgenext is seeking a Senior Data Quality Engineer to join a modern, cloud-based data platform team. This role is critical to ensuring the accuracy, completeness, reliability, and trustworthiness of enterprise data assets across the organization.
The ideal candidate is deeply passionate about data integrity and brings hands on experience designing and automating robust data quality frameworks in a cloud environment, with strong exposure to Azure and Databricks. You will embed quality checks across the entire data lifecycle—from ingestion through consumption—ensuring data is fit for business critical and client-facing use cases.
Responsibilities include but are not limited to:
Data Quality Framework & Automation
Design, build, and maintain a scalable, automated data reconciliation framework using Python, PySpark, and Databricks
Develop and implement automated data quality checks, validation rules, and monitoring processes within ETL / ELT pipelines on Azure
Embed data quality controls directly into ingestion and transformation workflows, enabling a “quality by design” approach
Data Profiling & Analysis
Write advanced SQL and Python (PySpark) scripts to profile datasets and identify anomalies
Diagnose data quality issues and perform root cause analysis across complex data pipelines
Collaborate with source system owners and data stewards to drive remediation and long term quality improvements
API Automation & Testing
Develop Python-based automation to validate data ingested from internal and external APIs
Ensure data integrity, schema compliance, and adherence to defined data contracts
Monitoring, Metrics & Reporting
Define, track, and monitor Data Quality Metrics (DQMs)
Build dashboards and reports using Power BI and/or Databricks SQL to provide visibility into data health for both technical and business stakeholders
Governance & Documentation
Support implementation of data governance, lineage, and security controls
Maintain comprehensive documentation for data quality rules, standards, processes, and frameworks
Workplace: This is a remote role requiring full-time Eastern timezone working hours and can be based in Ontario, Alberta, or British Columbia
Must Have Skills:
5+ years of experience in data quality, data engineering, or QA focused roles
Strong programming expertise in Python for automation, scripting, and data manipulation
Hands on experience building and scheduling PySpark jobs on Databricks
Advanced SQL skills for complex querying, profiling, and analysis
Solid understanding of ETL / ELT pipelines and large scale data processing
Proven experience with Azure data services, including:
Azure Data Lake Storage (ADLS)
Azure Data Factory (ADF)
Strong understanding of data warehousing concepts, including dimensional modeling
Experience working with large, enterprise scale data pipelines
Preferred Skills:
Experience automating tests for RESTful APIs using Python libraries such as requests or pytest
Familiarity with CI/CD and DevOps practices for data platforms (e.g., Jenkins, GitHub Actions)
Experience with data quality frameworks and tools such as Great Expectations or Amazon Deequ
Exposure to Infrastructure as Code (IaC) tools such as Terraform or ARM templates
Working knowledge of Agile delivery methodologies
Exposure to insurance and banking domain terminology and regulated data environments
Professional Skills:
Solid written, verbal, and presentation communication skills
Strong team and individual player
Maintains composure during all types of situations and is collaborative by nature
High standards of professionalism, consistently producing high quality results
Self-sufficient, independent requiring very little supervision or intervention
Demonstrate flexibility and openness to bring creative solutions to address issues
Bridgenext is an Equal Opportunity Employer
Canadian citizens and those authorized to work in Canada are encouraged to apply
Compensation varies depending on a wide array of factors, which may include but are not limited to location, role, skill set, and level of experience. As required by local law, Bridgenext provides a reasonable range of compensation, based on full-time employment, for roles that may be hired as described above. The current salary range for this position is $135,000 - $155,000 CAD annually. Our comprehensive total rewards program goes way beyond a competitive salary.
#LI-MA1
#LI-REMOTE