Experience Summary
I have more than * years of IT experience in DataStage, Hadoop, Big Data technologies, Data warehousing (DWH) and Business intelligence (BI) tools and technologies. I have strong skills in ranging from ETL design and development, Data Analysis, Issue resolution and Support. I have good exposure to data warehousing concepts and has the capability to quickly adapt to new technology.
Extensive knowledge and hands-on experience in Teradata, MongoDB, Hadoop, Hive, Sqoop, TDCH, Pyspark, R Programming, PL-SQL, ETL tool DataStage, Informatica and SQL.
Involved in different phases of Software Development cycle like Coding, unit testing, System Testing, User acceptance testing support and worked on Waterfall as well as Agile model
Have strong business and technical expertise in designing and delivering enterprise level data using Hadoop and ETL substantiated with Retail domain
Good communication, team and leadership skills and align quickly with the project needs.
Work Experience
Organization
Designation
Duration
Cognizant Technology Solutions
Programmer Analyst
April’2014 to Nov’2015
Tata Consultancy Services (Contractor by Enterprise Solutions Inc.)
Hadoop/ETL Developer
Feb’2018 to Aug’2018
Tata Consultancy Services (Contractor by Nithyo Infotech Inc.)
Hadoop/ETL Developer
Oct’2018 to Feb’2019
UST GLOBAL Inc. (Contractor by Data Capital Inc.)
Software Developer (Big Data Developer)
Mar’2019 till Present
Summary
Had done development, unit testing, System Testing, User acceptance testing support and deployment work in Hadoop, Hive, MongoDB, Teradata, Sqoop, TDCH, Pyspark, PL-SQL, R Programming, DataStage 8.5,9.1 & 11.5, UNIX, Informatica and Oracle.
Had done installation, configuration, management and deployment of Big Data solutions and the underlying infrastructure of Hadoop Cluster with the knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts
Had done enhancement and maintenance work in PL-SQL & UNIX scripting.
a.Ability to analyze critical business applications
b.Had done enhancement and maintenance work in different technologies too like UNIX and Oracle.
Ability to maintain and build client relationships, able to comfortably converse with all facets in the Clients organization.
Proficient in providing Analytical Support to key business Applications /solutions
Good verbal and written skills, especially with regards to presenting findings to both technical and non-technical addressees
Thorough understanding of Software Development Life Cycle (SDLC) and Agile model involved in various phases like Requirements, Analysis/Design, Development and Testing.
Acted as liaison with all teams for Requirements, Development, QA and Deployment
Working on automation of project with assistance of senior resource.
Create and update SQL based queries and other Database related tasks in GRS.
Understand business pain points and drive business process.
Restoration of service in shortest possible time.
Good analytical, logical, programming and communication skills.
Technical Skills
Hardware
Unix, Windows, IBM S390
Operating System
Win95/98,Windows XP, UNIX,LINUX
Databases
MongoDB, Oracle, Teradata, DB2, Informix, HDFS
Programming Languages & Tools
Websphere DataStage 8.5,9.1,11.5, Hive, Sqoop, TDCH, Pyspark, R Programming, Teradata BTEQ Scripts, PL-SQL, Informatica & MySQL Workbench 6.0, Visual Studio 2010/2015
Other Utilities
CI-CD Pipeline using Concord, SVN, Team Forge, HPLM, Rapid Deploy, GITHUB, ServiceNow, Remedy
Education
Course
Institution
University/Board
Year of Passing
Percentage
B.E. (Electrical& Electronics Engineering)
SSGPURC
Punjab University
2013
68.49
12th Standard
SGGS Collegiate Public School
C.B.S. E
2009
79.70
10TH Standard
Little Flower Convent School
I.C.S.E
2007
79.40
Live Project Walkthrough
Project Name
Assortment Analytics Feb’19-Present (Bentonville- Arkansas)
Customer
Wal-Mart Stores
Description
The aim of the product is to develop measurable insights into customer decisions and how they stop, item loyalty and substitutability, identifying missing items in the assortment, optimizing space, clustering etc .The tool also helps to find demand transfer for the Product according to Customer visits to the Stores and trends of Stores separated by demographics.
Technology
MongoDB, Teradata BTEQ, Hadoop, Hive, Sqoop, TDCH, Pyspark, PL-SQL, Unix, R Programming
Tools
Team forge, SVN, GitHub, ServiceNow, Ca7 Scheduler, Automic, Machine learning Platform(element)
Role
Software Developer (Big Data Developer)
Responsibilities
Work on design, development and implementation of connectivity products that allow efficient exchange of data between Walmart core database engine and Hadoop ecosystem for Assortment Project for Walmart Stores Inc.
Understand and analyze business requirement, building the design documents such as requirement document, change request, unit test documents and implement into production by raising change controls and provide post implementation validation and support.
Perform requirement gathering, analysis, estimation, architecture designing, development and implementation of HQL/PYSPARK/DataStage/R and Hadoop Jobs.
Enhancements of Oracle Procedures/Packages to meet Client’s performance needs and changing demands.
Prepare Teradata to Mongo DB/Hadoop system mappings, Teradata BTEQ Scripts, workflows, Unix scripts, PLSQL scripts as part of the coding.
Design of HQL/PYSPARK/DataStage/R and Hadoop Jobs
Analyzing and resolving production issues within Service level Agreements (SLA), handling day to day support activities to ensure business continuity.
Responsible for bug fixing and providing Support for User Acceptance Testing.
Adds Validation Checkpoints in system to create alerts in the event of potential functional issues & also develop tools for monitoring system performance.
Evaluate user requests for new or modified requirements and consult with user to identify current operating procedures to clarify the objectives.
Prepare flowcharts and diagrams to illustrate sequence of steps, that must be followed and describe logical operations involved
Perform requirement gathering, business analysis, project planning, status reporting and handling escalations.
Create CA7/Automic/Oozie scheduling jobs and review with Mainframe Operation Support/Automic/Hadoop Admin team.
Schedule the Jobs in CA7/Automic/Oozie and test the scheduling with the batch dependencies with tester for scheduling approval.
Work on importing, exporting, building and coding Hadoop languages.
Interact with representatives of other industrial groups to understand and recommend data interchange formats, APIs, messaging strategies, application server software, as well as hardware and virtual machine requirements.
Provide estimation level of effort, evaluate new options of similar technology, offer suggestions to improve processes and provide comments on some design aspects.
Coordinate with business partners for functional requirements gathering, which involves review of existing systems configurations and operating methodologies in a collaborative manner.
Coordinate between client and offshore teams for any clarifications on requirements.
Project Name
DataStage Migration Project Oct’18-Feb’19 (Bentonville- Arkansas)
Customer
Wal-Mart Stores
Description
Migrating more than 15k Jobs for Walmart Stores from DataStage 9.1 to DataStage 11.5 for areas like Transportation, Logistics, Compliance, OMS etc. for US and Intl Market.
Technology
DataStage 9.1/11.5, Unix, Pl/Sql, Oracle, Teradata, Db2, Informix
Tools
GitHub, ServiceNow, Rapid Deploy
Role
DataStage Developer
Responsibilities
Worked on Migrating DataStage Jobs from Infosphere Information Server, Version 9.1 to Infosphere Information Server, Version 11.5.
Worked on gathering Requirements, designing topology.
Testing jobs in Dev, QA and Production Implementation/Support.
Environmental setup like SFTP, DB Connections.
Fixed bugs in various DataStage jobs which were incompatible with version 11.5.
Coordinating with offshore to fix the issues in DataStage 11.5.
QA Testing Walkthrough with Client with thorough functional Analysis on data.
Worked with DataStage Admin to fix various environmental issues and raised IBM PMR for certain issues.
Fixed DataStage jobs to remove Warnings.
Used Rapid Deploy Tool for Production Release.
Worked on Converting XML Input Stage to Hierarchal Data Stage and enhanced the Performance.
Parsed Large Schema XML through Hierarchal Data Stage by understanding the requirements from business.
Have a good understanding on Mappings in Hierarchal Data Stage
Project Name
CA COGS Feb’18-4th Aug’2018 (Bentonville- Arkansas)
Customer
Wal-Mart Stores
Description
CA COGS project objective is to build a CA DNP tool which will be used by business to analyze Profitability of business by calculating sales and other adhoc costs which includes vendor funding, payment terms with vendor. Previously tool was built with SQL Server as an ETL which was sequential process and takes 3 weeks to run. Current approach of design is to build everything on Hadoop Platform to do parallel processing which will take 2 days to run and reflect reports on UI.
Technology
Hadoop, Hive, Sqoop, TDCH, Pyspark, Scala, PL-SQL, Unix
Tools
Team forge, SVN, GitHub, ServiceNow
Role
Hadoop Developer
Responsibilities
Work on design, development and implementation of connectivity products that allow efficient exchange of data between Walmart core database engine and Hadoop ecosystem for CA COGS project for Walmart Stores Inc.
Understand and analyze business requirement, building the design documents such as requirement document, change request, unit test documents and implement into production by raising change controls and provide post implementation validation and support.
Develop SQL, Procedures/Functions, Triggers and Packages on database (RDBMS) packages for the application.
Perform import and export data using Sqoop from HDFS to Relational Database systems and vice-versa.
Work on installation, configuration, management and deployment of Big Data solutions and the underlying infrastructure of Hadoop Cluster with the knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts
Work on MapReduce Programming/HDFS Frame work including Hive, Sqoop, Spark and set up standards and processes for Hadoop based application design and implementation.
Work on importing, exporting, building and coding Apache Spark using Pyspark & Scala languages.
Interact with representatives of other industrial groups to understand and recommend data interchange formats, APIs, messaging strategies, application server software, as well as hardware and virtual machine requirements.
Provide estimation level of effort, evaluate new options of similar technology, offer suggestions to improve processes and provide comments on some design aspects.
Coordinate with business partners for functional requirements gathering, which involves review of existing systems configurations and operating methodologies in a collaborative manner.
Coordinate between client and offshore teams for any clarifications on requirements.
Developed the New ECOMM flow in Spark by working closely with end client
Project Name
GRS (Global Replenishment Solutions) Oct’14-Nov’15 (Chennai-India)
Customer
Wal-Mart Stores
Description
Legacy system used to maintain purchase order lifecycle is too complex with multiple ways of creating POs that requires excessive manual intervention steps and customer-maintained scripts to process data and perform mass updates. Main benefit of GRS to migrate from Legacy is to provide a single platform for performing forecasting and replenishment activities. This can be carried for all the product types and also it provides visibility to the forecast and carries replenishment activity plans at item/location level.
The Global Replenishment System (GRS) project was initiated by Walmart to design, develop and deploy a common global cross platform for replenishment planning and purchase order execution. The completed solution will replace existing portfolio of legacy applications that handle these functions currently. GRS consist of a core replenishment engine and an interface layer that handles the data exchange with dozens of Walmart legacy systems. The Replenishment engine consist of JDA's manuagistics packaged application software, tailored by JDA to meet Walmart's special requirements. The interface layer consists of more than 1500 DataStage, Mainframe, PL/SQL and java programs for different regions
Technology
Oracle, PL/SQL, Teradata, DB2, Hadoop, Hive, Sqoop
Tools
IBM Information Server 9.1 Data Stage, Team forge, SVN
Role
Production Support and Enhancement
Responsibilities
Performed major enhancements in Reporting Workbench, Item Complexity Interfaces
Root cause analysis and permanent fixes of major failures.
Project Management – creates and maintains project documents like Flow diagram of project, functionality of each interface.
Monitors the Flow of GRS UI, PO Cuts, DC Picks and Reporting jobs which is used by various customers on replenishment forecast as well as critical reports to achieve the SLA’s.
Performance Tuning on Long Running PL/SQL Scripts
Tuned the Performance of Long Running ETL jobs by implementing DataStage best Practices and recommendations.
Analyzed Data and Code issues in Production and implemented the necessary fix.
Analyzed Batch flow for various Outages on Teradata, DataStage etc.
Project Name
Nexus Integration April’14-Oct’14 (Chennai-India)
Customer
Academy Sports, USA
Description
Academy sports and outdoors is a sporting goods discount store chain based out of Texas. As part of the Nexus integration project, ETL batch process is being used to replace the existing Island pacific (IP) and legacy systems to the new Oracle Retail Merchandising System (ORMS). DataStage is used for extraction and transformation of data from the new ORMS system and also to integrate the downstream systems with ORMS.
Technology
IBM Information Server V8.5 Data Stage
Role & Responsibilities
The Project involved in understanding a wide variety of applications that the client uses in its Enterprise Landscape
Designed and developed DataStage sequences with Audit and Error Handling
Tuned the Performance of Long Running ETL jobs by implementing DataStage best Practices and recommendations.
Unit testing of data stage ETL jobs.
Create test cases in compliance with requirement.
Handle all the design related activities and initiatives.
Metadata management – coordinate with the source system application teams to track metadata level changes and take steps to implement the changes in the warehouse after analyzing the end – end impact.
Performance tuning data stage ETL jobs.
Daily and weekly status calls with onsite team