JustinBradley’s client, a leading source of mortgage financing, is seeking a highly skilled Senior Principal Data Engineer with deep expertise in data engineering and designing and implementing scalable, cloud-native data platforms using AWS and Azure. This individual will play a critical role in building robust, modern data architectures and driving real-time analytics, AI-driven insights, and data governance at scale.
Must be local to Reston, VA or Plano, TX for their on-site requirement.
Responsibilities:
Architect and develop modern, cloud-native data platforms on AWS and Azure, ensuring scalability, reliability, and performance.
Design and implement advanced data architectures including Data Lake, Delta Lake, Lakehouse, OneLake, and Data Mesh to support real-time analytics and AI initiatives.
Lead the development and optimization of complex ETL/ELT pipelines, including Change Data Capture (CDC) mechanisms for high-velocity data ingestion and transformation.
Integrate and manage enterprise data using tools like Databricks, Apache Spark, PySpark, Apache Airflow, and Apache Flink.
Leverage AWS services such as S3, Redshift, Glue, and EMR, as well as Azure Data Lake, to enable secure, high-throughput data processing and analytics.
Build and maintain self-service BI ecosystems using Power BI, promoting business agility and data democratization.
Ensure enterprise-level data governance, quality, and compliance across data platforms and pipelines.
Collaborate closely with data scientists, analysts, architects, and business stakeholders to ensure that data solutions meet strategic objectives.
Stay abreast of emerging trends in big data, cloud computing, and analytics to continuously improve architecture and tools.
Requirements:
15+ years of experience in data engineering, architecture, and platform development.
5+ years of hands-on experience designing and deploying scalable, cloud-native solutions on AWS and/or Azure.
Proven expertise in modern data architectures: Data Lake, Delta Lake, Lakehouse, OneLake, Data Mesh.
Strong experience with Databricks, Apache Spark, PySpark, Airflow, Flink, and similar data tools.
Deep knowledge of ETL/ELT, data ingestion, transformation, and Change Data Capture (CDC) strategies.
Hands-on experience with AWS S3, Redshift, Glue, EMR, Azure Data Lake, and Power BI.
Exceptional problem-solving skills and ability to work independently in a fast-paced environment.
Experience leading cross-functional data engineering teams or enterprise-scale migration projects.
Certifications in AWS, Azure, or Databricks are a strong plus.
Familiarity with data privacy regulations and compliance (e.g., GDPR, HIPAA).
JustinBradley is an EO employer - Veterans/Disabled and other protected employees.