Sandya Poreddy
Austin, TX Open to Relocation & Remote Email: ***************@*****.*** Phone: 571-***-****
Professional Summary
Data Engineer with 6+ years of experience designing, building, and optimizing cloud-based data solutions. Proficient in Snowflake, Databricks, Apache Airflow, ADF, Spark, and Python for ETL/ELT pipelines in batch and real-time environments. Skilled in SQL, data modeling, CI/CD, Git, performance tuning, and implementing data quality frameworks (Deequ, Great Expectations). Experienced in data migration, streaming (Kafka), and cloud platforms including Azure and AWS. Proven record of optimizing data pipelines to reduce costs by up to 25% and improving processing times by 40%.
Technical Skills
Programming: Python, PySpark, SQL
Big Data & Processing: Apache Spark, Databricks, Delta Lake
Data Warehousing: Snowflake, Azure Synapse Analytics, SQL Server, DB2
Cloud Platforms: Azure, AWS
Workflow Orchestration: Apache Airflow, Azure Data Factory (ADF)
Streaming: Kafka
Version Control & CI/CD: Git, GitHub, Azure DevOps, Jenkins
Data Quality: Deequ, Great Expectations, custom Python validation
Other: Docker, Linux
Professional Experience
Disney – Data Migration (Snowflake + Unity Catalog)
- Migrated large-scale datasets across Snowflake environments using zero-copy cloning and incremental loads, ensuring 99.9% uptime.
- Built and scheduled Airflow DAGs with custom Python operators for workflow orchestration.
- Implemented automated reconciliation framework, reducing data validation time by 60%.
- Tuned Snowflake compute workloads, reducing warehouse costs by 20% while improving query runtime by 35%.
- Worked with Unity Catalog for centralized governance, data lineage, and fine-grained access control.
Hussmann – Data Migration (DB2 Azure Synapse Analytics)
- Designed and developed ADF pipelines to migrate legacy DB2 data into Azure Synapse Analytics.
- Automated schema mapping, data validation, and reconciliation checks to ensure 100% data accuracy.
- Optimized pipelines with partitioning and parallelism, reducing migration runtime by 45%.
- Integrated monitoring dashboards to track pipeline performance and data quality metrics.
- Used Git and CI/CD pipelines in Azure DevOps to deploy and maintain ETL workflows.
Doordash – ETL & Business Logic Automation (SQL + Airflow)
- Implemented SQL-based business transformations for reporting and analytics.
- Designed Airflow DAGs for automated workflows with retries, alerts, and dependencies.
- Integrated data quality checks (null, uniqueness, range validation) directly in ETL.
- Optimized SQL queries and workflows, reducing execution time by 35%.
- Leveraged Git and CI/CD for version control and automated deployment of data pipelines.
Assistant Professor – Computer Science
- Taught undergraduate and postgraduate Computer Science courses.
- Subjects included Database Systems, Programming, Data Structures, and Distributed Systems.
- Mentored students in academic projects, research, and career preparation.
- Published and presented research papers in academic conferences.
Education
Master of Computer Science
Bachelor of Computer Science