Post Job Free

Resume

Sign in

Big Data Developer

Location:
Bentonville, AR
Posted:
January 28, 2021

Contact this candidate

Resume:

Experience Summary

I have more than * years of IT experience in DataStage, Hadoop, Big Data technologies, Data warehousing (DWH) and Business intelligence (BI) tools and technologies. I have strong skills in ranging from ETL design and development, Data Analysis, Issue resolution and Support. I have good exposure to data warehousing concepts and has the capability to quickly adapt to new technology.

Extensive knowledge and hands-on experience in Teradata, MongoDB, Hadoop, Hive, Sqoop, TDCH, Pyspark, R Programming, PL-SQL, ETL tool DataStage, Informatica and SQL.

Involved in different phases of Software Development cycle like Coding, unit testing, System Testing, User acceptance testing support and worked on Waterfall as well as Agile model

Have strong business and technical expertise in designing and delivering enterprise level data using Hadoop and ETL substantiated with Retail domain

Good communication, team and leadership skills and align quickly with the project needs.

Work Experience

Organization

Designation

Duration

Cognizant Technology Solutions

Programmer Analyst

April’2014 to Nov’2015

Tata Consultancy Services (Contractor by Enterprise Solutions Inc.)

Hadoop/ETL Developer

Feb’2018 to Aug’2018

Tata Consultancy Services (Contractor by Nithyo Infotech Inc.)

Hadoop/ETL Developer

Oct’2018 to Feb’2019

UST GLOBAL Inc. (Contractor by Data Capital Inc.)

Software Developer (Big Data Developer)

Mar’2019 till Present

Summary

Had done development, unit testing, System Testing, User acceptance testing support and deployment work in Hadoop, Hive, MongoDB, Teradata, Sqoop, TDCH, Pyspark, PL-SQL, R Programming, DataStage 8.5,9.1 & 11.5, UNIX, Informatica and Oracle.

Had done installation, configuration, management and deployment of Big Data solutions and the underlying infrastructure of Hadoop Cluster with the knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts

Had done enhancement and maintenance work in PL-SQL & UNIX scripting.

a.Ability to analyze critical business applications

b.Had done enhancement and maintenance work in different technologies too like UNIX and Oracle.

Ability to maintain and build client relationships, able to comfortably converse with all facets in the Clients organization.

Proficient in providing Analytical Support to key business Applications /solutions

Good verbal and written skills, especially with regards to presenting findings to both technical and non-technical addressees

Thorough understanding of Software Development Life Cycle (SDLC) and Agile model involved in various phases like Requirements, Analysis/Design, Development and Testing.

Acted as liaison with all teams for Requirements, Development, QA and Deployment

Working on automation of project with assistance of senior resource.

Create and update SQL based queries and other Database related tasks in GRS.

Understand business pain points and drive business process.

Restoration of service in shortest possible time.

Good analytical, logical, programming and communication skills.

Technical Skills

Hardware

Unix, Windows, IBM S390

Operating System

Win95/98,Windows XP, UNIX,LINUX

Databases

MongoDB, Oracle, Teradata, DB2, Informix, HDFS

Programming Languages & Tools

Websphere DataStage 8.5,9.1,11.5, Hive, Sqoop, TDCH, Pyspark, R Programming, Teradata BTEQ Scripts, PL-SQL, Informatica & MySQL Workbench 6.0, Visual Studio 2010/2015

Other Utilities

CI-CD Pipeline using Concord, SVN, Team Forge, HPLM, Rapid Deploy, GITHUB, ServiceNow, Remedy

Education

Course

Institution

University/Board

Year of Passing

Percentage

B.E. (Electrical& Electronics Engineering)

SSGPURC

Punjab University

2013

68.49

12th Standard

SGGS Collegiate Public School

C.B.S. E

2009

79.70

10TH Standard

Little Flower Convent School

I.C.S.E

2007

79.40

Live Project Walkthrough

Project Name

Assortment Analytics Feb’19-Present (Bentonville- Arkansas)

Customer

Wal-Mart Stores

Description

The aim of the product is to develop measurable insights into customer decisions and how they stop, item loyalty and substitutability, identifying missing items in the assortment, optimizing space, clustering etc .The tool also helps to find demand transfer for the Product according to Customer visits to the Stores and trends of Stores separated by demographics.

Technology

MongoDB, Teradata BTEQ, Hadoop, Hive, Sqoop, TDCH, Pyspark, PL-SQL, Unix, R Programming

Tools

Team forge, SVN, GitHub, ServiceNow, Ca7 Scheduler, Automic, Machine learning Platform(element)

Role

Software Developer (Big Data Developer)

Responsibilities

Work on design, development and implementation of connectivity products that allow efficient exchange of data between Walmart core database engine and Hadoop ecosystem for Assortment Project for Walmart Stores Inc.

Understand and analyze business requirement, building the design documents such as requirement document, change request, unit test documents and implement into production by raising change controls and provide post implementation validation and support.

Perform requirement gathering, analysis, estimation, architecture designing, development and implementation of HQL/PYSPARK/DataStage/R and Hadoop Jobs.

Enhancements of Oracle Procedures/Packages to meet Client’s performance needs and changing demands.

Prepare Teradata to Mongo DB/Hadoop system mappings, Teradata BTEQ Scripts, workflows, Unix scripts, PLSQL scripts as part of the coding.

Design of HQL/PYSPARK/DataStage/R and Hadoop Jobs

Analyzing and resolving production issues within Service level Agreements (SLA), handling day to day support activities to ensure business continuity.

Responsible for bug fixing and providing Support for User Acceptance Testing.

Adds Validation Checkpoints in system to create alerts in the event of potential functional issues & also develop tools for monitoring system performance.

Evaluate user requests for new or modified requirements and consult with user to identify current operating procedures to clarify the objectives.

Prepare flowcharts and diagrams to illustrate sequence of steps, that must be followed and describe logical operations involved

Perform requirement gathering, business analysis, project planning, status reporting and handling escalations.

Create CA7/Automic/Oozie scheduling jobs and review with Mainframe Operation Support/Automic/Hadoop Admin team.

Schedule the Jobs in CA7/Automic/Oozie and test the scheduling with the batch dependencies with tester for scheduling approval.

Work on importing, exporting, building and coding Hadoop languages.

Interact with representatives of other industrial groups to understand and recommend data interchange formats, APIs, messaging strategies, application server software, as well as hardware and virtual machine requirements.

Provide estimation level of effort, evaluate new options of similar technology, offer suggestions to improve processes and provide comments on some design aspects.

Coordinate with business partners for functional requirements gathering, which involves review of existing systems configurations and operating methodologies in a collaborative manner.

Coordinate between client and offshore teams for any clarifications on requirements.

Project Name

DataStage Migration Project Oct’18-Feb’19 (Bentonville- Arkansas)

Customer

Wal-Mart Stores

Description

Migrating more than 15k Jobs for Walmart Stores from DataStage 9.1 to DataStage 11.5 for areas like Transportation, Logistics, Compliance, OMS etc. for US and Intl Market.

Technology

DataStage 9.1/11.5, Unix, Pl/Sql, Oracle, Teradata, Db2, Informix

Tools

GitHub, ServiceNow, Rapid Deploy

Role

DataStage Developer

Responsibilities

Worked on Migrating DataStage Jobs from Infosphere Information Server, Version 9.1 to Infosphere Information Server, Version 11.5.

Worked on gathering Requirements, designing topology.

Testing jobs in Dev, QA and Production Implementation/Support.

Environmental setup like SFTP, DB Connections.

Fixed bugs in various DataStage jobs which were incompatible with version 11.5.

Coordinating with offshore to fix the issues in DataStage 11.5.

QA Testing Walkthrough with Client with thorough functional Analysis on data.

Worked with DataStage Admin to fix various environmental issues and raised IBM PMR for certain issues.

Fixed DataStage jobs to remove Warnings.

Used Rapid Deploy Tool for Production Release.

Worked on Converting XML Input Stage to Hierarchal Data Stage and enhanced the Performance.

Parsed Large Schema XML through Hierarchal Data Stage by understanding the requirements from business.

Have a good understanding on Mappings in Hierarchal Data Stage

Project Name

CA COGS Feb’18-4th Aug’2018 (Bentonville- Arkansas)

Customer

Wal-Mart Stores

Description

CA COGS project objective is to build a CA DNP tool which will be used by business to analyze Profitability of business by calculating sales and other adhoc costs which includes vendor funding, payment terms with vendor. Previously tool was built with SQL Server as an ETL which was sequential process and takes 3 weeks to run. Current approach of design is to build everything on Hadoop Platform to do parallel processing which will take 2 days to run and reflect reports on UI.

Technology

Hadoop, Hive, Sqoop, TDCH, Pyspark, Scala, PL-SQL, Unix

Tools

Team forge, SVN, GitHub, ServiceNow

Role

Hadoop Developer

Responsibilities

Work on design, development and implementation of connectivity products that allow efficient exchange of data between Walmart core database engine and Hadoop ecosystem for CA COGS project for Walmart Stores Inc.

Understand and analyze business requirement, building the design documents such as requirement document, change request, unit test documents and implement into production by raising change controls and provide post implementation validation and support.

Develop SQL, Procedures/Functions, Triggers and Packages on database (RDBMS) packages for the application.

Perform import and export data using Sqoop from HDFS to Relational Database systems and vice-versa.

Work on installation, configuration, management and deployment of Big Data solutions and the underlying infrastructure of Hadoop Cluster with the knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts

Work on MapReduce Programming/HDFS Frame work including Hive, Sqoop, Spark and set up standards and processes for Hadoop based application design and implementation.

Work on importing, exporting, building and coding Apache Spark using Pyspark & Scala languages.

Interact with representatives of other industrial groups to understand and recommend data interchange formats, APIs, messaging strategies, application server software, as well as hardware and virtual machine requirements.

Provide estimation level of effort, evaluate new options of similar technology, offer suggestions to improve processes and provide comments on some design aspects.

Coordinate with business partners for functional requirements gathering, which involves review of existing systems configurations and operating methodologies in a collaborative manner.

Coordinate between client and offshore teams for any clarifications on requirements.

Developed the New ECOMM flow in Spark by working closely with end client

Project Name

GRS (Global Replenishment Solutions) Oct’14-Nov’15 (Chennai-India)

Customer

Wal-Mart Stores

Description

Legacy system used to maintain purchase order lifecycle is too complex with multiple ways of creating POs that requires excessive manual intervention steps and customer-maintained scripts to process data and perform mass updates. Main benefit of GRS to migrate from Legacy is to provide a single platform for performing forecasting and replenishment activities. This can be carried for all the product types and also it provides visibility to the forecast and carries replenishment activity plans at item/location level.

The Global Replenishment System (GRS) project was initiated by Walmart to design, develop and deploy a common global cross platform for replenishment planning and purchase order execution. The completed solution will replace existing portfolio of legacy applications that handle these functions currently. GRS consist of a core replenishment engine and an interface layer that handles the data exchange with dozens of Walmart legacy systems. The Replenishment engine consist of JDA's manuagistics packaged application software, tailored by JDA to meet Walmart's special requirements. The interface layer consists of more than 1500 DataStage, Mainframe, PL/SQL and java programs for different regions

Technology

Oracle, PL/SQL, Teradata, DB2, Hadoop, Hive, Sqoop

Tools

IBM Information Server 9.1 Data Stage, Team forge, SVN

Role

Production Support and Enhancement

Responsibilities

Performed major enhancements in Reporting Workbench, Item Complexity Interfaces

Root cause analysis and permanent fixes of major failures.

Project Management – creates and maintains project documents like Flow diagram of project, functionality of each interface.

Monitors the Flow of GRS UI, PO Cuts, DC Picks and Reporting jobs which is used by various customers on replenishment forecast as well as critical reports to achieve the SLA’s.

Performance Tuning on Long Running PL/SQL Scripts

Tuned the Performance of Long Running ETL jobs by implementing DataStage best Practices and recommendations.

Analyzed Data and Code issues in Production and implemented the necessary fix.

Analyzed Batch flow for various Outages on Teradata, DataStage etc.

Project Name

Nexus Integration April’14-Oct’14 (Chennai-India)

Customer

Academy Sports, USA

Description

Academy sports and outdoors is a sporting goods discount store chain based out of Texas. As part of the Nexus integration project, ETL batch process is being used to replace the existing Island pacific (IP) and legacy systems to the new Oracle Retail Merchandising System (ORMS). DataStage is used for extraction and transformation of data from the new ORMS system and also to integrate the downstream systems with ORMS.

Technology

IBM Information Server V8.5 Data Stage

Role & Responsibilities

The Project involved in understanding a wide variety of applications that the client uses in its Enterprise Landscape

Designed and developed DataStage sequences with Audit and Error Handling

Tuned the Performance of Long Running ETL jobs by implementing DataStage best Practices and recommendations.

Unit testing of data stage ETL jobs.

Create test cases in compliance with requirement.

Handle all the design related activities and initiatives.

Metadata management – coordinate with the source system application teams to track metadata level changes and take steps to implement the changes in the warehouse after analyzing the end – end impact.

Performance tuning data stage ETL jobs.

Daily and weekly status calls with onsite team



Contact this candidate