EPAM Systems logo

Senior Data Quality Engineer

EPAM Systems Argentina

onsite
Posted Feb 5, 2026

Role & seniority: Senior Data Quality Engineer

Stack/tools

  • Programming: Python (advanced)

  • Big Data: Hadoop ecosystem (HDFS, Hive, Spark); streaming (Kafka, Flume, Kinesis)

  • Data stores: NoSQL (Cassandra, MongoDB, HBase); RDBMS (PostgreSQL, MSSQL, MySQL, Oracle)

  • Data viz/analytics: Tableau, Power BI, Tibco Spotfire

  • Cloud: AWS, Azure, GCP; multi-cloud architectures

  • ETL/MDM: Talend, Informatica; MDM tooling

  • Testing/CI: TDD/DT/BDT; JMeter; Git/GitHub Actions/Jenkins; automation/scripting

  • Governance/QA: data quality governance frameworks; automated validation pipelines; documentation

Top 3 responsibilities

  1. Define and enforce end-to-end data quality strategy and enterprise-level testing frameworks

  2. Develop, scale, and maintain automated data quality validation pipelines for production systems

  3. Establish governance, ensure regulatory/business compliance, mentor engineers, and collaborate cross-functionally to optimize data pipelines

Must-have skills

  • 3+ years in Data Quality Engineering

  • Advanced Python; deep experience with Hadoop/Spark and streaming platforms

  • Proficiency with NoSQL and relational databases; strong SQL

  • Experience with ETL tooling and MDM integration; CI/CD practices

  • Expertise in modern testing frameworks (TDD/DT/BDT) and data-focused validation

  • Cloud experience (AWS/Azure/GCP); multi-cloud architectures

  • Strong problem-solving and communication (English, B2+)

Nice-to-haves

  • Ja

Full Description

We are seeking an experienced and accomplished Senior Data Quality Engineer to join our team, driving the reliability, accuracy, and efficiency of our data systems and processes at scale. If you are passionate about shaping high-impact data quality initiatives and are eager to work with advanced technologies, this role will empower you to influence the future of our data landscape. Responsibilities Oversee end-to-end data quality strategy, ensuring rigorous testing and reliability of data products and processes Drive data quality initiatives while instilling best practices across multiple teams and projects Define and enforce advanced testing methodologies and frameworks to ensure enterprise-level data quality Prioritize and manage complex data quality tasks, optimizing efficiency under tight timelines and competing demands Architect and maintain robust testing strategies tailored to evolving system architectures and data pipelines Advise on resource allocation, setting priorities for testing aligned with regulatory and business standards Establish and continuously enhance a robust data quality governance framework, overseeing compliance with industry standards Develop, scale, and refine automated data quality validation pipelines for production systems Collaborate at a high level with cross-functional teams to resolve infrastructure challenges and optimize performance Mentor junior engineers while maintaining comprehensive documentation, including versions of test strategies and advanced test plans Requirements 3+ years of professional experience in Data Quality Engineering Advanced programming skills in Python Deep expertise in Big Data platforms, including Hadoop ecosystem tools (HDFS, Hive, Spark) and modern streaming platforms (Kafka, Flume, Kinesis) Strong practical knowledge of NoSQL databases such as Cassandra, MongoDB, or HBase, with a track record of handling enterprise-scale datasets Advanced skills in data visualization and analytics tools (e.g., Tableau, Power BI, Tibco Spotfire) to support decision-making Extensive hands-on experience with cloud ecosystems like AWS, Azure, and GCP, including a strong understanding of complex multi-cloud architectures Demonstrated expertise with relational databases and SQL (PostgreSQL, MSSQL, MySQL, Oracle) in high-volume, real-time environments Proven ability to implement, troubleshoot, and scale ETL processes using tools like Talend, Informatica, or similar platforms Experience deploying and integrating MDM tools into existing workflows, with knowledge of performance testing tools such as JMeter Advanced experience in version control systems (Git, GitLab, or SVN) and automation/scripting for large-scale systems Comprehensive knowledge of modern testing frameworks (TDD, DDT, BDT) and their application in data-focused environments Familiarity with CI/CD practices, including implementation of pipelines using tools like Jenkins or GitHub Actions Highly developed problem-solving abilities and an analytical mindset capable of interpreting complex datasets into actionable business outcomes Exceptional verbal and written communication skills in English (B2 level or higher), paired with experience guiding discussions with stakeholders Nice to have Extensive hands-on experience with programming languages like Java, Scala, or advanced Bash scripting for production-level data solutions Advanced knowledge of XPath and its applications in data validation or transformation workflows Experience designing customized data generation tools and sophisticated synthetic data techniques for testing scenarios We offer International projects with top brands Work with global teams of highly skilled, diverse peers Healthcare benefits Employee financial programs Paid time off and sick leave Upskilling, reskilling and certification courses Unlimited access to the LinkedIn Learning library and 22,000+ courses Global career opportunities Volunteer and community involvement opportunities EPAM Employee Groups Award-winning culture recognized by Glassdoor, Newsweek and LinkedIn

Nivel de antigüedad Intermedio Tipo de empleo Jornada completa Función laboral Tecnología de la información, Ingeniería y Control de calidad Sectores Desarrollo de software, Servicios y consultoría de TI y Tecnología, información e internet

Data Quality StrategyData Quality InitiativesTesting MethodologiesData Quality GovernanceAutomated Validation PipelinesCross-functional CollaborationMentoringPythonBig Data PlatformsNoSQL DatabasesData VisualizationCloud EcosystemsSQLETL ProcessesMDM ToolsCI/CD Practices

Cookies & analytics consent

We serve candidates globally, so we only activate Google Tag Manager and other analytics after you opt in. This keeps us aligned with GDPR/UK DPA, ePrivacy, LGPD, and similar rules. Essential features still run without analytics cookies.

Read how we use data in our Privacy Policy and Terms of Service.