Post Job Free
Sign in

Senior Data Engineer / Data Analyst Control Automation & ETL Testing

Company:
MARKS IT Solutions
Location:
Wilmington, Devon, United Kingdom
Posted:
January 28, 2026
Apply

Description:

Job Title Options:

Senior Data Engineer / Data Analyst – Control Automation & ETL Testing

Location:

Wilmington, DE (Hybrid – 3 days onsite/week)

Engagement Type:

Long-Term Contract (W2)

Client:

Capital One (Former Capital One experience required)

Job Summary / Description:

We are seeking an experienced Data Analyst to join the Risk and Controls organization at Capital One. This role bridges data engineering and control execution, with a strong focus on ETL development, automated data validation, and scripted QA testing.

The ideal candidate will bring deep expertise in Python and SQL, along with hands-on experience building and monitoring production-grade ETL pipelines. This position plays a critical role in ensuring data integrity, lineage, and compliance across first-line risk controls in a highly collaborative, fast-paced environment.

Key Responsibilities:

• Design, develop, and maintain scalable ETL pipelines that transform raw data into reliable datasets supporting risk and control frameworks.

• Optimize complex SQL queries and Python-based data pipelines to improve performance and reliability across platforms such as Postgres and Snowflake.

• Integrate structured and unstructured data sources (including JSON) into a unified data layer for reporting and control execution.

• Build and maintain automated data quality and QA test suites using Python frameworks such as PyTest and Great Expectations.

• Implement data-as-code testing frameworks to proactively detect anomalies, schema drift, and data integrity issues.

• Perform unit and integration testing to validate ETL logic against business and system rules.

• Support data governance initiatives, including metadata management, technical lineage, and CI/CD deployment of data assets.

• Evaluate upstream and downstream integration points to ensure SQL logic accurately reflects system states and reporting requirements.

• Identify bottlenecks in data pipelines and implement automation solutions to eliminate manual processes.

• Partner closely with Engineering, Operations, and Risk teams to translate control requirements into technical ETL specifications.

• Communicate data risks, discrepancies, and remediation plans clearly to both technical and non-technical stakeholders.

Basic Qualifications:

• Master’s degree in a quantitative or technical discipline.

• Proven experience developing and supporting ETL pipelines in production environments.

• Expert-level proficiency in Python and SQL for data manipulation, transformation, and automated testing.

• Experience working with relational and non-relational databases (Postgres, MySQL, DynamoDB, Cassandra, or similar).

Preferred Qualifications:

• Experience building automated QA and data validation frameworks.

• Hands-on experience with AWS services such as S3, Glue, Lambda, and IAM.

• Familiarity with data orchestration tools (Airflow, Prefect) and version control systems (Git).

• Strong experience processing and transforming unstructured data (JSON) for structured analytics and reporting.

Apply