Post Job Free
Sign in

Data Engineer Supply Chain

Location:
New York City, NY
Salary:
70000
Posted:
October 15, 2025

Contact this candidate

Resume:

Prasannakumar Kasindala

Data Engineer *************@*****.*** Linkedin 201-***-****

SUMMARY

Data Engineer with 4+ years of experience in building scalable data pipelines, cloud data platforms, and real- time streaming solutions. Specialized in AWS services (Glue, Redshift, S3) with strong hands-on experience in Snowflake, Databricks, Airflow, and DBT. Skilled at designing ETL/ELT pipelines, optimizing data warehouses, and applying governance best practices. Proven ability to deliver secure, cost-efficient, and business-focused data solutions for enterprise clients in retail and fintech. EDUCATION

Masters in Management Data Analytics, SPEC from Indiana Weslyean University, Dec 2024 SKILLS

Programming & Scripting: Python (Pandas, PySpark), SQL, Scala

Cloud Platforms: AWS (Glue, Redshift, S3, RDS), Snowflake, Databricks

Data Integration & Orchestration: Apache Airflow, DBT, AWS Step Functions, Kinesis

Data Warehousing & Modeling: Redshift, Snowflake, Star/Snowflake Schema, Data Vault

Streaming & Big Data: Apache Kafka, AWS Kinesis, Delta Lake

Automation & DevOps: Terraform, GitHub Actions, AWS CodePipeline

Data Security & Governance: IAM, GDPR, Data Masking, Compliance Standards

Visualization & BI: Amazon QuickSight, Tableau

EXPERIENCE

Data Engineer Kroger Feb 2024 – Current

Environment: AWS (Glue, Redshift, S3, RDS), Databricks, Airflow, Kafka, QuickSight

Designed and automated ETL pipelines in AWS Glue and PySpark to ingest retail transactions and supply chain data into Redshift for analytics and reporting.

Implemented Databricks with Delta Lake for advanced transformations, supporting scalable machine learning feature sets and reducing processing time by 30%.

Orchestrated workflows using Apache Airflow DAGs, improving pipeline reliability and reducing manual dependencies by 60%.

Built Kafka-based real-time ingestion pipelines, enabling near real-time visibility into store sales and inventory levels.

Partnered with BI teams to develop QuickSight dashboards that delivered actionable insights on product demand and supply chain efficiency.

Data Engineer Intuit Dec 2019 – Sep 2022

Environment: AWS (Glue, Redshift, S3), Snowflake, DBT, Kafka, Tableau

Migrated petabyte-scale datasets from legacy systems into Snowflake and Redshift, applying DBT models for standardized transformations and reusable data marts.

Developed ETL pipelines with Glue and PySpark to process structured/semi-structured data from APIs, S3, and RDS, improving ingestion throughput by 40%.

Built CDC pipelines with Kafka to replicate transactional data into Snowflake, enabling finance teams to access near real-time revenue metrics.

Optimized Snowflake cost and performance by tuning virtual warehouses, implementing clustering, and applying role-based access controls (RBAC).

Enhanced Redshift query performance via schema redesign, distribution/sort keys, and automated vacuuming, cutting query runtime by 30%.

Delivered interactive Tableau dashboards with self-service KPIs for finance and product teams, improving reporting turnaround by 50%.

Ensured compliance with IAM policies and GDPR by implementing secure access and data masking strategies.



Contact this candidate