Post Job Free
Sign in

Data Engineer Software Development

Location:
Wilmington, DE
Posted:
October 29, 2024

Contact this candidate

Resume:

Sheetal Gandhi

Phone: 703-***-****

********.******@*****.***

Work Experience Summary:

●17+ years of experience in various phases of software development life cycle, including designing, development and testing of business applications using technologies such as Snowflake, Oracle 8i/9i/10g/11g, Hadoop, Azure, DataStage 8.5/9.1, Informatica, ER/Studio Data Architect, shell and Python scripting, various scheduling tools.

●Snowflake Data Engineer with extensive experience in migrating on-premise data warehouses to cloud. Experience with Snowflake on Azure - SnowSQL, zero copying cloning, azcopy, JavaScript stored proc etc.

●Familiar with AWS data tools such as Glue, RedShift etc.

●Experienced in development for NoSQL databases such as HBase.

●Extensive experience in PL/SQL, writing stored procedures, packages, triggers and other database objects, Import/Export, SQL Loader, used standard packages like DBMS_SCHEDULER, DBMS_PROFILER, DBMS_OUTPUT, DBMS_SQL, DBMS_UTILS and UTL_FILE utilities.

●Experienced in complex stored procedure, triggers, views and other database objects, bulk load, queuing, export/import, XML based data load, partition exchange, indexes and granting privileges.

●Hadoop frameworks utilizing Hive, HBase, Sqoop, Hue etc. in Avro, text file formats.

●Understanding of Teradata BTEQ scripting, tpt utility

●Experience in SDLC(Waterfall) and Agile methodologies such as Kanban, Scrum in various capacities –Developer, Scrum Master, BSA. Experience of conducting scrum ceremonies and related reporting.

●Experience in interacting with users, analyzing client business processes, documenting business requirements, performing design analysis and developing design specifications.

●High Volume ETL data load using IBM DataStage/Informatica/Ab Initio

●Scheduling jobs using Autosys, TWS, Control-M, Tidal, Cron

●Experience in Performance Tuning & Optimization of SQL statements.

●Possesses excellent analytical, programming and communication skills.

●Working knowledge of PQL(Celonis)and KSQL(Kafka).

●Passionate about exploring new technologies and finding innovative solutions to complex business problems with a commitment to continuous learning and professional development.

Certifications:

Machine Learning A-Z™: Hands-On Python In Data Science: Udemy, Oct’20

Deep Learning and Computer Vision A-Z™: OpenCV, SSD & GANs: Udemy, currently pursuing.

SnowPro Core Certificate COF-CO2: Currently pursuing

Technical Skills:

Operating Systems : Windows NT/95/98/2000/XP/8/10, UNIX, Linux

Cloud Environments : Azure, AWS

RDBMS/DB : Snowflake, Oracle 8i/9i/10g/11g, Hive, HBase, MS-Access, Sql Server

Programming : SnowSQL, PL/SQL, SQL, TSQL, C, C++, HiveSQL, Python for Data Science

Data Modeling Tools : ER/StudioData Architect7.0, ERwin Data Modeler

ETL Tools : Informatica 9.5, IBM DataStage 8.5/9.1, AB Initio

Scheduling Tools : Autosys, TWS, Control-M, Tidal, Cron

Reporting Tools : BOXI, Crystal Reports, Apache Solr

Application Dev Tools : Toad, SQL Developer, SQLPlus, SQL Loader, Hue, Eclipse, Rational ClearCase, SVN, WinCVS, git

Education:

University of Hertfordshire, UK- Master of Science,2010

Dr B.A.M.University, India- Bachelor of Engineering -2002

Work Experience:

Sr Snowflake Developer/Data Analyst, Staples May 2020 - Present

Involved in Teradata migration to Snowflake for multiple large legacy warehouses.

Created Snowflake framework stored procs to copy data from Prod to QA on a weekly basis.

Converted 300+ BTEQ scripts to equivalent SnowSQL scripts.

Migration of historical data for 800+ tables using Python framework/shell scripts.

1000+ Tidal jobs re-created for Tidal. Worked with extended team on plan for parallel run and plan for switching off legacy jobs.

Teradata stored procedure conversion to Snowflake stored proc in JavaScript/SQL

Worked with multiple vendors and business teams across the company to validate 60+ outgoing feeds’ migration to Snowflake from Teradata warehouses.

Extensive experience with various Snowflake features such as zero copy cloning, COPY INTO, virtual warehouse sizing, time travel, creating dashboards, etc.

Transformation from Json to flattened data in Snowflake.

Sqoop results of HiveQL queries from Hadoop to Snowflake.

Created POC Power BI report to display images from Snowflake external stage.

Rewrote existing data load frameworks, Unix scripts to work with Snowflake instead of Teradata

Imported source definitions in Informatica for Snowflake tables using Snowflake connector.

Analysis and conversion of legacy simple Informatica Workflows to Snowflake only processes, eliminating use of additional ETL tool.

Daily working sessions with various business/analyst teams across the organization to help them migrate reports, macros, etc. to Snowflake.

Data analysis using SQL and Python

Data visualization using Snowflake Dashboards, Excel charts and Power BI

Analysis of coupon usage and sales data via email and digital channels

Comparison and root cause analysis of differences in reports from Teradata vs Snowflake using custom code

Dropship vendor report card generation for various violations such as late shipping/delivery, additional shipping charges etc. using data from UPS, USPS and FedEx.

Experience presenting analytical findings to audiences at all levels.

Environments: Snowflake, Teradata, Azure, Tidal, Informatica, JavaScript, Jira, Shell scripting, Python Scripting, Azure Storage Explorer, Sqoop, Power BI, MS-Access

Sr PL/SQL Developer/Cross-functional Agile Member, Barclaycard US Aug 2013 - Mar 2020

Helped drive projects in Tech lead/Sr Dev capacity for the client. Have been the go-to SME for applications owned by the team with an understanding of both the business and technology aspects of the applications. Also filled in for other roles in Agile team as and when needed.

Enhanced Enterprise Communication engines involving emails, Letter, SMS, iPush as needed using Oracle Pl/SQL, IBM DataStage, Hadoop Sqoop, Hive, Unix Shell scripts, Control-M.

Migrating data from Verint Click-to-chat, Aprimo Campaign, email/letter engine databases to Barclays Data Lake using Hadoop frameworks utilizing Hive, HBase, Sqoop, Hue etc. in Avro, text file formats.

Utilized DataStage components to generate XML reports for publishing on the customer website.

Integrated real time Java Web Services and ETL Batch systems using the DataStage for awarding points earned through social media activity.

Created process to gather customer data from all lines of business and consolidate it in central Hive on HBase table. Export data from these tables to Oracle DB using internal Sqoop export framework. Built process to load flat files into Hadoop using internal file ingest/scheduler framework

Assisted in creating data models for KYC, social media communities and other applications using data governance guidelines.

Performance tune DataStage jobs, Sequencers to handle huge amount of Bank data.

Production On-Call support to help resolve high priority issues in a timely manner. Experience with creating End of Sprint wrap-up Summary, managing story backlog, and conducting scrum ceremonies.

Conducted brown bag technology sessions for various topics.

Environments: Oracle 10/11g, PL/SQL, Toad, Hive, Impala, HBase, Sqoop, IBM DataStage, Control-M, Tivoli Workload Scheduler, MicroStrategy, Unix, Shell script, git, Snowflake, S3, WinCVS, SVN, Eclipse, Cucumber, Rally, Jira

PL/SQL Developer- Fannie Mae, VA Aug 2011 - Apr 2013

The U.S.Securities and Exchange Commission (SEC) adopted Dodd Frank rule 15-Ga-1 mandated the GSEs such as Fannie Mae to disclose fulfilled and unfulfilled repurchase requests concerning assets in asset-backed securities. Repurchase Disclosure System was a new application built from scratch in a limited time for the report to be filed quarterly to SEC.

Also helped the client develop the Servicing Capacity P&L Analytical System (SCAPLS) which was built for internal reporting to help Finance Organization and National Servicing Organization to keep track of performance of the rapidly growing loan population transferred to ’high touch’ Servicers against benchmark and for additional related analytics on the same.

Wrote PL/SQL and SQL procedures, functions, and packages for data quality checks, logging, and staging to core table moves.

Developed triggers and exception handling procedures for sending alert notifications for business rule violations.

Developed processes to load over 65 M+ records from various systems for The Dodd-Frank report.

Environments: Oracle 11g, PL/SQL, Autosys, SQL, TOAD, ClearCase, Unix, Shell script, Visio, ER-Studio, Ab Initio, BOXI, Microsoft Project

PL/SQL Developer- Gamestec Leisure Limited, UK Oct 2007 – Feb 2011

Designed and developed complex procedures to handle errors and exceptions at both application and database levels using PL/SQL and shell scripts.

Analyzed and discussed the business requirements with the business owners and business analysts.

Environments: Oracle 9i/10g, PL/SQL, HTML, SQL, TOAD, VSS, Unix, Shell script

Software Engineer- Nexa Systems, India Jan 2005-Sep 2005

Involved in database development by creating Oracle PL/SQL Functions, Procedures, Triggers, Packages, Records, and Collections.

Environments: Oracle 9i, PL/SQL, SQL, TOAD, VSS

Jr. Software Engineer-Shah Electronics, India Jul 2002 – Dec 2004

Gathered application requirements and provided design recommendations.

Assisted in the design of tables, databases, forms, and reports.

Environments: Oracle 8, PL/SQL, SQL, SQL*Plus



Contact this candidate