Integrity logo

Data Engineer in Quality Assurance team (Remote: CET ±3 hours)

Integrity

remote

Salary: 21 Pa

Posted Oct 27, 2025

Role & seniority: QA Data Engineer; mid-level to senior (3+ years programming experience; back-end/data engineering focus)

Stack/tools: Python, .NET C#, SQL; data formats (JSON, XML, CSV, SAS); databases (MongoDB, SQL); ETL pipelines; Docker, Linux, Bash; Azure DevOps or similar; familiarity with data warehouses/l lakes; CDISC SDTM/ADaM (plus); synthetic/de-identified data generation

Top 3 responsibilities

  1. Develop, maintain, and optimize structured and unstructured datasets with auditability and traceability

  2. Design data mappings from raw clinical data to internal formats; implement automated data quality checks and monitoring

  3. Collaborate with R&D, Data Science, and QA; support ETL/integration; validate metrics with synthetic/de-identified datasets; ensure regulatory alignment

Must-have skills

  • 3+ years software programming (data engineering/back-end)

  • 2+ years with Python, .NET or scripting; strong SQL/data profiling

  • Experience with ETL, data transformation, automated data quality checks, and testing

  • Familiarity with agile development; domain knowledge of clinical trials data standards (CDISC SDTM/ADaM) is a plus

  • Troubleshooting, communication, problem-solving; ability to work with legacy and new architectures

Nice-to-haves

  • Azure cloud experience/certifications; CFR 21 Part 11, ICH GCP, FDA validation familiarity; GDPR/data privacy awareness

  • Experience generating synthetic/de-identified datasets; healthcare/clinical data d

Full Description

Join Cyntegrity, one of the leading clinical trial risk management technology providers in the industry. We're proud of our highly rated MyRBQM® Portal and MyRBQM® Academy brands.

If you’re passionate about data, technology, and innovation in healthcare — and you’re looking to join a fast-growing, purpose-driven company — this is an exciting opportunity for you.

Our international and collaborative team is looking for a Quality Assurance (QA) Data Engineer to support the development and optimization of our data pipelines, datasets, and KRI (Key Risk Indicator) systems that power clinical trial insights and analytics. You will own the data quality lifecycle—from ingestion and mapping through validation, lineage, and audit—working closely with R&D, Data Science, and QA.

Key Responsibilities Develop, maintain, and optimize structured and unstructured datasets with strong traceability and audit readiness. Design and implement data mappings from raw clinical data to Cyntegrity’s internal formats. Enhance data quality, consistency, and efficiency across systems by implementing automated data quality checks and monitoring. Collaborate with R&D, Data Science, and QA teams to ensure alignment with business and regulatory requirements. Support data transformation, ETL, and integration processes. Work with both legacy and new code to improve scalability and reliability of data systems. Contribute to risk management and compliance efforts across regulated environments. Prepare realistic test datasets (synthetic and de-identified) to validate metrics across potential clinical scenarios; design scenario-based validation test dataset and acceptance criteria.

Essential Criteria

Education Bachelor’s or Master’s degree in Computer Science, Software Engineering, or a related field.

Experience / Skills 3+ years of software programming experience (Data engineering or back-end focus preferred). 2+ years of experience with .NET, Python, or other scripting languages. Hands-on SQL expertise (query performance, modeling, and data profiling). Experience with agile product development lifecycle. Proficiency in data transformation, ETL, and pipeline development. Proven implementation of automated data quality checks, validation rules, and monitoring (e.g., unit/integration tests for data, expectations-based testing). Familiarity with clinical data standards (e.g., CDISC SDTM/ADaM) and common source formats from EDC, labs, and safety systems is a plus. Experience generating synthetic or de-identified datasets. Familiarity with risk management concepts and process-oriented environments. Comfortable working remotely and delivering measurable results. Strong troubleshooting, communication, and problem-solving skills. Willingness to work with both legacy systems and new architectures. Working knowledge of written and spoken English.

Knowledge (Technical Proficiency)

Programming Languages: Python, R, JavaScript, .NET C#

Data Formats: XML/XPATH, JSON, SAS, CSV, Excel

Databases: MongoDB, SQL (familiarity with data warehouses/lakehouses is a plus).

Technologies: Docker, Linux, Bash, Azure DevOps (or similar tools), ETL pipelines

Bonus: Azure cloud experience or certifications

Plus: Domain knowledge in clinical trials, data standards, and applied statistics Familiarity with CFR 21 Part 11, ICH GCP, and FDA Software Validation Guidance; Understanding of GDPR and data privacy best practices

Compensation Full-time employment Competitive salary Remote work flexibility

About Cyntegrity We support sponsors and CROs with clinical trial risk management technology and educational programs to bring therapies to market faster, safer, and more efficiently. Cyntegrity is a fast-growing, international company driven by innovation and quality. We aim to be the first-choice provider of specialized cloud and SaaS solutions for ICH GCP E6(R3)-compliant risk-based quality management and clinical research. Learn more about us and our brands at www.cyntegrity.com

If this is a position of your interest, and you would like to know more or have any additional questions, please submit your application.

Data EngineeringPythonSQLETLData QualityData TransformationAgile DevelopmentClinical Data StandardsTroubleshootingCommunicationProblem SolvingRisk ManagementData ProfilingAutomated TestingAzure DevOpsDockerreview:company

Cookies & analytics consent

We serve candidates globally, so we only activate Google Tag Manager and other analytics after you opt in. This keeps us aligned with GDPR/UK DPA, ePrivacy, LGPD, and similar rules. Essential features still run without analytics cookies.

Read how we use data in our Privacy Policy and Terms of Service.