Sheetal Gandhi
Phone: 703-***-****
********.******@*****.***
Work Experience Summary:
●17+ years of experience in various phases of software development life cycle, including designing, development and testing of business applications using technologies such as Snowflake, Oracle 8i/9i/10g/11g, Hadoop, Azure, DataStage 8.5/9.1, Informatica, ER/Studio Data Architect, shell and Python scripting, various scheduling tools.
●Snowflake Data Engineer with extensive experience in migrating on-premise data warehouses to cloud. Experience with Snowflake on Azure - SnowSQL, zero copying cloning, azcopy, JavaScript stored proc etc.
●Familiar with AWS data tools such as Glue, RedShift etc.
●Experienced in development for NoSQL databases such as HBase.
●Extensive experience in PL/SQL, writing stored procedures, packages, triggers and other database objects, Import/Export, SQL Loader, used standard packages like DBMS_SCHEDULER, DBMS_PROFILER, DBMS_OUTPUT, DBMS_SQL, DBMS_UTILS and UTL_FILE utilities.
●Experienced in complex stored procedure, triggers, views and other database objects, bulk load, queuing, export/import, XML based data load, partition exchange, indexes and granting privileges.
●Hadoop frameworks utilizing Hive, HBase, Sqoop, Hue etc. in Avro, text file formats.
●Understanding of Teradata BTEQ scripting, tpt utility
●Experience in SDLC(Waterfall) and Agile methodologies such as Kanban, Scrum in various capacities –Developer, Scrum Master, BSA. Experience of conducting scrum ceremonies and related reporting.
●Experience in interacting with users, analyzing client business processes, documenting business requirements, performing design analysis and developing design specifications.
●High Volume ETL data load using IBM DataStage/Informatica/Ab Initio
●Scheduling jobs using Autosys, TWS, Control-M, Tidal, Cron
●Experience in Performance Tuning & Optimization of SQL statements.
●Possesses excellent analytical, programming and communication skills.
●Working knowledge of PQL(Celonis)and KSQL(Kafka).
●Passionate about exploring new technologies and finding innovative solutions to complex business problems with a commitment to continuous learning and professional development.
Certifications:
Machine Learning A-Z™: Hands-On Python In Data Science: Udemy, Oct’20
Deep Learning and Computer Vision A-Z™: OpenCV, SSD & GANs: Udemy, currently pursuing.
SnowPro Core Certificate COF-CO2: Currently pursuing
Technical Skills:
Operating Systems : Windows NT/95/98/2000/XP/8/10, UNIX, Linux
Cloud Environments : Azure, AWS
RDBMS/DB : Snowflake, Oracle 8i/9i/10g/11g, Hive, HBase, MS-Access, Sql Server
Programming : SnowSQL, PL/SQL, SQL, TSQL, C, C++, HiveSQL, Python for Data Science
Data Modeling Tools : ER/StudioData Architect7.0, ERwin Data Modeler
ETL Tools : Informatica 9.5, IBM DataStage 8.5/9.1, AB Initio
Scheduling Tools : Autosys, TWS, Control-M, Tidal, Cron
Reporting Tools : BOXI, Crystal Reports, Apache Solr
Application Dev Tools : Toad, SQL Developer, SQLPlus, SQL Loader, Hue, Eclipse, Rational ClearCase, SVN, WinCVS, git
Education:
University of Hertfordshire, UK- Master of Science,2010
Dr B.A.M.University, India- Bachelor of Engineering -2002
Work Experience:
Sr Snowflake Developer/Data Analyst, Staples May 2020 - Present
Involved in Teradata migration to Snowflake for multiple large legacy warehouses.
Created Snowflake framework stored procs to copy data from Prod to QA on a weekly basis.
Converted 300+ BTEQ scripts to equivalent SnowSQL scripts.
Migration of historical data for 800+ tables using Python framework/shell scripts.
1000+ Tidal jobs re-created for Tidal. Worked with extended team on plan for parallel run and plan for switching off legacy jobs.
Teradata stored procedure conversion to Snowflake stored proc in JavaScript/SQL
Worked with multiple vendors and business teams across the company to validate 60+ outgoing feeds’ migration to Snowflake from Teradata warehouses.
Extensive experience with various Snowflake features such as zero copy cloning, COPY INTO, virtual warehouse sizing, time travel, creating dashboards, etc.
Transformation from Json to flattened data in Snowflake.
Sqoop results of HiveQL queries from Hadoop to Snowflake.
Created POC Power BI report to display images from Snowflake external stage.
Rewrote existing data load frameworks, Unix scripts to work with Snowflake instead of Teradata
Imported source definitions in Informatica for Snowflake tables using Snowflake connector.
Analysis and conversion of legacy simple Informatica Workflows to Snowflake only processes, eliminating use of additional ETL tool.
Daily working sessions with various business/analyst teams across the organization to help them migrate reports, macros, etc. to Snowflake.
Data analysis using SQL and Python
Data visualization using Snowflake Dashboards, Excel charts and Power BI
Analysis of coupon usage and sales data via email and digital channels
Comparison and root cause analysis of differences in reports from Teradata vs Snowflake using custom code
Dropship vendor report card generation for various violations such as late shipping/delivery, additional shipping charges etc. using data from UPS, USPS and FedEx.
Experience presenting analytical findings to audiences at all levels.
Environments: Snowflake, Teradata, Azure, Tidal, Informatica, JavaScript, Jira, Shell scripting, Python Scripting, Azure Storage Explorer, Sqoop, Power BI, MS-Access
Sr PL/SQL Developer/Cross-functional Agile Member, Barclaycard US Aug 2013 - Mar 2020
Helped drive projects in Tech lead/Sr Dev capacity for the client. Have been the go-to SME for applications owned by the team with an understanding of both the business and technology aspects of the applications. Also filled in for other roles in Agile team as and when needed.
Enhanced Enterprise Communication engines involving emails, Letter, SMS, iPush as needed using Oracle Pl/SQL, IBM DataStage, Hadoop Sqoop, Hive, Unix Shell scripts, Control-M.
Migrating data from Verint Click-to-chat, Aprimo Campaign, email/letter engine databases to Barclays Data Lake using Hadoop frameworks utilizing Hive, HBase, Sqoop, Hue etc. in Avro, text file formats.
Utilized DataStage components to generate XML reports for publishing on the customer website.
Integrated real time Java Web Services and ETL Batch systems using the DataStage for awarding points earned through social media activity.
Created process to gather customer data from all lines of business and consolidate it in central Hive on HBase table. Export data from these tables to Oracle DB using internal Sqoop export framework. Built process to load flat files into Hadoop using internal file ingest/scheduler framework
Assisted in creating data models for KYC, social media communities and other applications using data governance guidelines.
Performance tune DataStage jobs, Sequencers to handle huge amount of Bank data.
Production On-Call support to help resolve high priority issues in a timely manner. Experience with creating End of Sprint wrap-up Summary, managing story backlog, and conducting scrum ceremonies.
Conducted brown bag technology sessions for various topics.
Environments: Oracle 10/11g, PL/SQL, Toad, Hive, Impala, HBase, Sqoop, IBM DataStage, Control-M, Tivoli Workload Scheduler, MicroStrategy, Unix, Shell script, git, Snowflake, S3, WinCVS, SVN, Eclipse, Cucumber, Rally, Jira
PL/SQL Developer- Fannie Mae, VA Aug 2011 - Apr 2013
The U.S.Securities and Exchange Commission (SEC) adopted Dodd Frank rule 15-Ga-1 mandated the GSEs such as Fannie Mae to disclose fulfilled and unfulfilled repurchase requests concerning assets in asset-backed securities. Repurchase Disclosure System was a new application built from scratch in a limited time for the report to be filed quarterly to SEC.
Also helped the client develop the Servicing Capacity P&L Analytical System (SCAPLS) which was built for internal reporting to help Finance Organization and National Servicing Organization to keep track of performance of the rapidly growing loan population transferred to ’high touch’ Servicers against benchmark and for additional related analytics on the same.
Wrote PL/SQL and SQL procedures, functions, and packages for data quality checks, logging, and staging to core table moves.
Developed triggers and exception handling procedures for sending alert notifications for business rule violations.
Developed processes to load over 65 M+ records from various systems for The Dodd-Frank report.
Environments: Oracle 11g, PL/SQL, Autosys, SQL, TOAD, ClearCase, Unix, Shell script, Visio, ER-Studio, Ab Initio, BOXI, Microsoft Project
PL/SQL Developer- Gamestec Leisure Limited, UK Oct 2007 – Feb 2011
Designed and developed complex procedures to handle errors and exceptions at both application and database levels using PL/SQL and shell scripts.
Analyzed and discussed the business requirements with the business owners and business analysts.
Environments: Oracle 9i/10g, PL/SQL, HTML, SQL, TOAD, VSS, Unix, Shell script
Software Engineer- Nexa Systems, India Jan 2005-Sep 2005
Involved in database development by creating Oracle PL/SQL Functions, Procedures, Triggers, Packages, Records, and Collections.
Environments: Oracle 9i, PL/SQL, SQL, TOAD, VSS
Jr. Software Engineer-Shah Electronics, India Jul 2002 – Dec 2004
Gathered application requirements and provided design recommendations.
Assisted in the design of tables, databases, forms, and reports.
Environments: Oracle 8, PL/SQL, SQL, SQL*Plus