Post Job Free
Sign in

Data Engineer Quality

Location:
Austin, TX
Posted:
September 26, 2025

Contact this candidate

Resume:

Sandya Poreddy

Austin, TX Open to Relocation & Remote Email: ***************@*****.*** Phone: 571-***-****

Professional Summary

Data Engineer with 6+ years of experience designing, building, and optimizing cloud-based data solutions. Proficient in Snowflake, Databricks, Apache Airflow, ADF, Spark, and Python for ETL/ELT pipelines in batch and real-time environments. Skilled in SQL, data modeling, CI/CD, Git, performance tuning, and implementing data quality frameworks (Deequ, Great Expectations). Experienced in data migration, streaming (Kafka), and cloud platforms including Azure and AWS. Proven record of optimizing data pipelines to reduce costs by up to 25% and improving processing times by 40%.

Technical Skills

Programming: Python, PySpark, SQL

Big Data & Processing: Apache Spark, Databricks, Delta Lake

Data Warehousing: Snowflake, Azure Synapse Analytics, SQL Server, DB2

Cloud Platforms: Azure, AWS

Workflow Orchestration: Apache Airflow, Azure Data Factory (ADF)

Streaming: Kafka

Version Control & CI/CD: Git, GitHub, Azure DevOps, Jenkins

Data Quality: Deequ, Great Expectations, custom Python validation

Other: Docker, Linux

Professional Experience

Disney – Data Migration (Snowflake + Unity Catalog)

- Migrated large-scale datasets across Snowflake environments using zero-copy cloning and incremental loads, ensuring 99.9% uptime.

- Built and scheduled Airflow DAGs with custom Python operators for workflow orchestration.

- Implemented automated reconciliation framework, reducing data validation time by 60%.

- Tuned Snowflake compute workloads, reducing warehouse costs by 20% while improving query runtime by 35%.

- Worked with Unity Catalog for centralized governance, data lineage, and fine-grained access control.

Hussmann – Data Migration (DB2 Azure Synapse Analytics)

- Designed and developed ADF pipelines to migrate legacy DB2 data into Azure Synapse Analytics.

- Automated schema mapping, data validation, and reconciliation checks to ensure 100% data accuracy.

- Optimized pipelines with partitioning and parallelism, reducing migration runtime by 45%.

- Integrated monitoring dashboards to track pipeline performance and data quality metrics.

- Used Git and CI/CD pipelines in Azure DevOps to deploy and maintain ETL workflows.

Doordash – ETL & Business Logic Automation (SQL + Airflow)

- Implemented SQL-based business transformations for reporting and analytics.

- Designed Airflow DAGs for automated workflows with retries, alerts, and dependencies.

- Integrated data quality checks (null, uniqueness, range validation) directly in ETL.

- Optimized SQL queries and workflows, reducing execution time by 35%.

- Leveraged Git and CI/CD for version control and automated deployment of data pipelines.

Assistant Professor – Computer Science

- Taught undergraduate and postgraduate Computer Science courses.

- Subjects included Database Systems, Programming, Data Structures, and Distributed Systems.

- Mentored students in academic projects, research, and career preparation.

- Published and presented research papers in academic conferences.

Education

Master of Computer Science

Bachelor of Computer Science



Contact this candidate