PRIYANKA DAS
Senior Data Engineer
Rancho Santa Margarita, CA
***********.*@*****.*** 714-***-****
Professional Summary
Senior Data Engineer with 13+ years of experience, seeking a challenging role in a global organization where technical expertise can drive scalable data solutions and continuous growth.
Expertise in Informatica, MDM, Snowflake, Datastage, Teradata, Netezza, Unix, Oracle, PL/SQL, Autosys, Cronacle, CA Workload Automation.
Expertise with AWS ecosystem – AWS Glue data pipelines, AWS Redshift and S3 Buckets, AWS Lambda, Dynamo DB. Skilled in Cloud Integration using AWS, Dell Boomi and ServiceNow for Incident Management.
Extensive Hands-on experience in Data Warehousing, Data Integration, Business Intelligence, Data Modeling using Erwin, Data Quality & Reliability Improvement, performance tuning, ETL and ELT workflows, and SQL optimization. Developed and maintained complex ETL pipelines in Python. Developed Automation frameworks using Python, Robot Framework, and Selenium (C# with NUNIT).
Proficient in Hadoop ecosystem frameworks HDFS, YARN, MapReduce, Hive, Sqoop, Kafka, NiFi, Apache Spark.
Extensive experience in Power BI, Azure Synapse, SQL Server, and database performance tuning. Experience in API orchestration using Postman and Ready API.
Strong domain knowledge in Customer Data Engineering, Insurance, Retail, Healthcare, and Banking.
Proficient in Agile methodologies, CI/CD implementations with JIRA, Jenkins, and GitHub. Experienced as SDET Lead in Big Data MDM projects, built Test Data Generation Framework integrated with Jenkins for one-click test file generation.
Skills
ETL & Data Warehousing: Informatica, Informatica MDM, Datastage, SQL Server, Teradata, Netezza.
Cloud Platforms & Services: Snowflake, AWS S3, AWS Lambda, AWS Glue, Amazon Redshift, Amazon RDS, Amazon DynamoDB, Azure Synapse Analytics.
Big Data & Streaming: Apache Hadoop, Apache Kafka, PySpark, Sqoop, Hive, Oozie, NiFi, Teradata, Netezza
Databases: SQL Server, Oracle, DB2, PL/SQL, PostgreSQL.
Data Formats & Integration: JSON, Parquet, Dell Boomi, Informatica
DevOps, CI/CD & Tools: Git, Jenkins, Agile, JIRA, Azure DevOps, ServiceNow, ALM.
Scripting & Automation: Python, Robot Framework, Shell Scripting, Selenium (C# with NUNIT), Teradata BTEQ.
Analytics & Visualization: Power BI, Azure Synapse.
Scheduling Tools: Autosys, Cronacle, CA Workload Automation
API Tools: Postman, Ready API
Incident Management: ServiceNow
Work Experience
Principal Systems Analyst – DWH & BI K&N Filters Riverside, CA Feb 2024 – Jan 2025
Accomplishments:
Developed and maintained Power BI dashboards for Finance, Operations, Sales, and Marketing teams.
Extract, process, and transforming source data for BI reporting. Developed and maintained complex ETL pipelines in Python.
Designed and developed SQL scripts, stored procedures, and AWS Glue data pipelines. Design and manage data workflows between AWS Redshift and S3 Buckets, APIs, and third-party applications.
Set up monitoring and logging with CloudWatch, CloudTrail, and AWS Config for performance tuning and compliance.
Configured IAM policies, roles, and MFA for secure access control across multiple AWS accounts.
Deployed serverless applications using AWS Lambda, API Gateway, and DynamoDB.
Exposure to AI/ML analytics, creating bots using NLP.
Facilitate Agile development processes including Sprint Planning and Retrospectives.
Environment: SQL Server Management Studio 19, Python, AWS Services, Power BI, ServiceNow, Jenkins, Azure DevOps, TFS, GitHub.
ETL Lead Data Engineer Vaco Enterprise LLC - Hyderabad, India Nov 2021 – Jul 2023 Client: CareSource Project: ESB Core Project
Accomplishments:
Led automation testing for ETL flows migrated from Tibco XML to Boomi JSON.
Created and managed ServiceNow incident tickets and change requests.
Validated HL7 interfaces and clinical files using Boomi and Tibco.
Automated regression testing using Selenium (C#) and NUNIT framework.
Extract, process, and transform source data for BI reporting. Develop and maintain complex ETL pipelines using Snowflake.
Design and develop SQL scripts, stored procedures, and Azure data pipelines. Design and manage data workflows between databases, APIs, and third-party applications.
Used Kafka, Confluent, and Splunk for data validation.
Developed SQL queries for backend data validation.
Provided project implementation support and knowledge transfer to production teams.
Environment: Snowflake, Dell Boomi, Tibco EMS BW, SQL Server 2018, Ready API, Azure DevOps, Python, Kafka, Splunk, GitHub, Jenkins.
ETL Lead SDET Cognizant Technology Solutions - Kolkata, India Jan 2019 – Nov 2021 Client: Manulife Financial Project: DMO & SMDM
Accomplishments:
Designed and optimized database structures for large-scale data warehouses.
Extract, process, and transform source data for BI reporting. Develop and maintain complex ETL pipelines using Snowflake.
Design and develop SQL scripts, stored procedures, and Azure data pipelines. Design and manage data workflows between databases, APIs, and third-party applications.
Automated test scripts using Python and Robot Framework.
Developed test automation frameworks integrated with Jenkins.
Conducted performance tuning of SQL queries, ETL jobs, and automation scripts.
Provided Agile leadership in Sprint Planning, Retrospectives, and Daily Stand-ups.
Led onsite-offshore coordination and task allocation.
Environment: Snowflake, SQL Server 2018, Python 3.7, Kafka, NiFi, JIRA, GitLab, Jenkins, Docker, Splunk
ETL Lead Developer Wells Fargo Home Mortgage - Fort Mill, SC Feb 2014 – Nov 2018
Accomplishments:
Developed Teradata Views, Macros, BTEQs, Unix Scripts, and Crontab scheduling.
Performed bulk data loads from Oracle and legacy systems using Teradata utilities (BTEQ, MLOAD, TPT, FastLoad).
Optimized SQL queries, performed performance tuning, and collected statistics on FACT tables.
Developed tracking tables, improved SLA reporting using JIRA, and built Data Marts.
Led migration efforts for Teradata14 and SASGRID, including DDL migration and configuration changes.
Designed and implemented cleanup processes for staging environments.
Environment: Teradata 15 Teradata SQL Assistant, Teradata Manager, Teradata view pointer, BTEQ, MLOAD, FLOAD, FASTEXPORT, Erwin Designer, UNIX, Korn Shell scripts.
ETL Lead Developer Tata Consultancy Services - Kolkata, West Bengal Sep 2010 – Oct 2011 Client: GE Healthcare Project: Life Sciences General Ledger Visibility Project
Accomplishments:
Developed Informatica mappings and BTEQ scripts for data transformation and integration.
Designed Cronacle job chains, calling scripts, event waits, and locks for job scheduling.
Led unit testing, system testing, regression testing, and data validation efforts.
Environment: Informatica 8.6.1, Teradata, UNIX, Redwood Cronacle, SVN.
Sr. ETL Developer & Module Lead Infosys – Pune, Bhubaneswar, India Oct 2007 - Sep 2010
Accomplishments:
Client: Bank of America and Northwestern Mutual Life Insurance
Developed ETL mappings, DataStage jobs, and transformation logic for Financial Systems.
Designed mapping workflows, created Unix scripts for test automation, and implemented performance tuning techniques.
Client: PetSmart Inc.
Developed and migrated ETL processes for Netezza-based Data Warehousing.
Conducted system testing, performance tuning, and production support.
Education:
M.S. in Software Engineering
International Technological University - San Jose, CA
B.Tech in Information Technology
West Bengal University of Technology - Kalyani, West Bengal
Certifications:
Informatica Certified Developer (PowerCenter 8)
PowerCenter Advanced Mapping Design
PowerCenter Mapping Design
Keywords: Agile, Amazon DynamoDB, Amazon Redshift, Amazon Web Services, Apache Hadoop, Apache Kafka, Automation, Azure Synapse Analytics, Big Data, Business Intelligence, Cloud Computing, Data Warehousing, Dell Boomi, ETL, Git, Informatica, Jenkins, JIRA, JSON, Parquet, Performance Tuning, Power BI, PySpark, Python, Snowflake, SQL, Unix Scripting.