Post Job Free

Resume

Sign in

Data Etl

Location:
Sunnyvale, CA
Salary:
125000
Posted:
March 27, 2020

Contact this candidate

Resume:

Prasanna Lakshmi R adcgl4@r.postjobfree.com +1-479-***-****

PROFESSIONAL SUMMARY

ETL professional having 15+ Years IT experience and track record in Analyzing, Designing, Developing, Testing, Implementing and maintaining client/server business systems.

Having more than 9 years of ETL development experience and 6+ years of QA experience in cloud-based environments with Extraction Transformation and Loading DataStage (ETL) tool –Knowledge in Data warehousing involving data warehouse analysis and design using IBM INFORMATION SERVER (Datastage)/Talend/SSIS, Apache Airflow, Apache Kafka.

Excellent QA experience in all phases of Software Testing Life Cycle (STLC) and Software Development Life Cycle (SDLC) with good working knowledge of QA methodologies, disciplines, tasks, resources and documenting.

Experience in Integration, Functional, System, Regression, SIT, Stress, Black Box, White Box and coordinating User Acceptance Testing (UAT).

Strong experience on Bigdata Hadoop Testing & Extensive experience in planning and executing strategy for testing the Database, DWH, ETL, Data Migration projects from various legacy systems to target database in cloud-based environments.

Expert in writing SQL and PL/SQL queries and worked with Bigdata/NoSQL/Cloud technologies- HIVE, Big Query in Google Cloud Platform (GCP)

Immense experience in testing Cloud based data pipelines and building Google Big Query SQL for data validation.

Worked on different Scheduling tools like CA7, AutoSys and good experience in writing JIL scripts and designing Autosys jobs using Web GUI and proven track record in troubleshooting of DataStage jobs and addressing production issues.

Skilled in writing technical specification documents, translating user requirements to technical specifications and creating and reviewing mapping documents.

SKILLS

ETL Tools

IBM Information Server 8x, Data stage 7.5x (Server & PX), Informatica (Beginner), Talend Studio.

Testing Tools

Test Link, VersionOne, JIRA, HPALM, Qtest

Languages

Unix shell scripting, SQL, Python (Basics), PL/SQL

Databases

Oracle 8.1, SQL Server 6.5, DB2, Teradata, MySQL, MongoDB, DB2, Big Query, Apache Hive

Job Scheduling

CA7, Control-M

Operating System

UNIX, WINDOWS, LINUX

Web Service Tools

SOAP UI, Rest API - Postman

Domains

Retail, Finance, banking, Securities and Electronics

CERTIFICATIONS

Certified in Infosphere DataStage 8.0.

Big Query

Machine Learning with Apache Spark

PROFESSIONAL EXPERIENCE

GSPANN Technologies, Sr. ETL Consultant

Milpitas CA - Nov 2013 to Till Date

Kohls - EDW Migration:

EDW migration project is to migrated all the data from legacy (Teradata) Database into Big Query Database in Google Cloud Platform. The data from different Business areas like OMS (Order Management system), Products, Vendor, Inventory, sales, marketing data is migrated to Big query.

Responsibilities:

Review Business requirements, functional specifications, use cases, and test plans.

Based on the STM (mapping document) /requirements, we prepare the Test scripts in Teradata and Big query and do the comparison to validate all the business transformations.

Historical & Incremental load data validation is performed based on multiple sources like DB2, Oracle, Flat files, Kafka being migrated to Big Query.

Validate all the data loaded related to OMS (Order Management system), Products, Vendor, Inventory, sales, marketing data is migrated to Big query through Apache Airflow into Big Query.

Validated the tokenized customer data sent in Encrypted files and loaded into BigQuery.

Executed the testcases and updated the test results in Qtest and raised the defects in JIRA for tracking.

Analyze failures and document software defects using bug tracking systems like report the defects involving program functionality, output, online screen and content to software developers. Follow up with developers, support and other teams on the resolution.

Validation of the data loaded into Audit tables for each EDW data load.

Study and analysis of the mapping document indicating the source tables, columns, data types, transformations required, business rules to be applied, target tables, columns and data types

Developed various Test Scripts, Test Execution using TestLink based on the Functional Specifications.

Writing complex SQL and PL/SQL queries using Case Logic, Intersect, Minus, Sub Queries, Inline Views, and Union.

Developed SQL queries for Back-end testing/Database testing and participated in end to end testing.

Performed Functional Testing, Regression Testing, End to End testing.

Provided daily and weekly status reports.

Environment: Apache Airflow, Teradata, DB2, Oracle, Big Query, ESP scheduler, JIRA, Qtest

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

Kohls - ECMDM (Enterprise Customer Master Data Management) data Analytics project.

The main role of ECMDM is to serve as a system of reference for customer master data. ECMDM collects customer data from all available source systems to form a single customer record that combines all data available on a customer. In this project we bring various types of data comes from multiple data sources like First data, KCC (Kohls Credit card), 3CC, (Third Party Credit card), Loyalty, ATG (webstore), Sales Alert data in flat files, Kafka, which we will be loaded to HIVE staging DB and MYSQL Database on cloud.

Responsibilities:

Review Business requirements, functional specifications, use cases, and test plans.

Prepared unit testcases which included different functional and technical scenarios.

Validate all the data loaded related to 3CC, KCC, Loyalty, ATG which are migrated from legacy application to New system though Batch jobs into Dimension tables like CUSTOMER, EMAIL, PROFILE, ACCOUNT, CREDIT CARD (MYSQL and Hive DB).

We do connect to HIVE/MYSQL database and validate the source data is loaded accurately based on the requirements.

The data is related to all customer information, which is tokenized. We validate the tokenized data by detokenizing on the HIVE server.

All the customer sensitive data is sent in Encrypted files and loaded into MYSQL.

Validate the Real streaming data sent through Kafka MQ Topics into MYSQL.

Validation of RestAPI services through Postman to Get/Post customer data from MYSQL.

Executed the testcases and updated the test results and raised the defects in JIRA for tracking.

Analyze failures and document software defects using bug tracking systems like report the defects involving program functionality, output, online screen and content to software developers. Follow up with developers, support and other teams on the resolution.

Validation of the data loaded into Audit tables for each data load.

We perform Regression testing as well as part of each story.

Environment: Hive, MYSQL, Batch jobs, Flat files, XML, Kafka (MQ), RestAPI postman, Protegrity tool.

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

Rodan & Fields (R+F) – ATLAS Project:

ATLAS project scope is for migrating heirloom application to Hybris and changing all the data sources to new system for maintaining product/Account/order related info for different types of customers like Preferred Customer (PC), Retail Customer (RC) and Consultants in MSSQL Database. On top of the migration implementation of promotions data load to DWH system using ETL (SSIS). Migration of SSRS reports to MSTR (Micro Strategy) reports. Integration of Amazon S3 with MicroStrategy Reporting to enable reporting for Account/Order/Product Audit Reports

Responsibilities:

Review Business requirements, functional specifications, use cases, and test plans.

Work closely with the Business, Solution Architect, Data Modeler, Data Mapper, Business Analysts to understand business requirements, providing expert knowledge and solutions on Data Warehousing, ensuring delivery of business needs in a timely cost-effective manner

Prepared unit testcases which included different functional and technical scenarios.

Validate all the data loaded related to Products, Customers, Orders, Shipments & Payments which are migrated from legacy application to New system though ETL SSIS package into EDW (SQL Server).

Validation of MicroStrategy reports which are built on top of the EDW loaded Tables, Stored procedures & Views etc.

Validated the migrated reports functionality from SSRS to MSTR based on the data comparison.

Executed the testcases and updated the test results and raised the defects in JIRA for tracking.

Analyze failures and document software defects using bug tracking systems like report the defects involving program functionality, output, online screen and content to software developers. Follow up with developers, support and other teams on the resolution.

Validation of the data loaded from AWS (S3) Athena Audit tables to EDW tables.

Environment: SSIS, SQL Server, MSTR Reports, AWS S3 services.

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

Lam Research – OPT Project

Online Planning Tool (OPT): The project is to create OPT dashboard to replace exiting OPT application and allow the business users and the IT to query their near real time, ad-hoc, standard reports, operational and analytical reports from SAP Hana.

Responsibilities:

Serve as technical lead and work with business partners to define product roadmap and requirements for SAP data migration program.

Interact closely with key business stakeholders to understand user demands; product architects, development team to understand technical application dependencies and help build finer and detailed product specifications packaging features/functionality into phased product releases.

Communicate functional and non-functional requirements via user stories to dev teams.

Specify data mapping & data provisioning methodologies along with application interface detailing.

Handled migration of legacy applications from SQL server to HANA models, enabling faster performance, better memory utilization and easier access to simpler interface modeling for future.

Evaluate and design data collection, data staging, data movement, analytics delivery, data quality and archiving strategies

Review the jobs for OPT ETL applications created using DataStage/BODS tool, which loads data into CSV’s, and TAB’s. These files are created by extracting the data from Oracle DB using PL/SQL and DataStage Technologies and the files are sent to down Stream applications.

Identify opportunities to optimize the ETL environment, implement monitoring, quality and validation processes to ensure data accuracy and integrity.

Extensive ETL tool experience using IBM Infosphere /WebSphere DataStage, Ascential DataStage.

Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.

Environment: IBM Information Server 8.0.1, BODS, SAP HANA, SQL Server, JIRA

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

Macys – STELLA Project

Stella is a Product Information Management System (PIM). All Product information and product images on the Macys.com website and Bloomingdales.com are managed in Stella. Stella manages Product assets including module like PRODUCT, ATTRIBUTES, PRICING, PROMOTIONS, UPC’s etc.

Responsibilities:

Work closely with the Business, Solution Architect, Data Modeler, Data Mapper, Business Analysts to understand business requirements, providing expert knowledge and solutions on Data Warehousing, ensuring delivery of business needs in a timely cost-effective manner

Evaluate and design data collection, data staging, data movement, analytics delivery, data quality and archiving strategies

Developed the jobs for Stella ETL applications using DataStage tool, which loads data into CSV’s, and TAB’s. Extracting the data from Oracle DB using PL/SQL creates these files and DataStage Technologies and the files are sent to down Stream applications.

Involved in unit testing of ETL DataStage jobs.

Identify opportunities to optimize the ETL environment, implement monitoring, quality and validation processes to ensure data accuracy and integrity.

Extensive ETL tool experience using IBM Infosphere/WebSphere DataStage, Ascential DataStage.

Worked on DataStage tools like DataStage Designer, DataStage Director.

Strong understanding of the principles of Data Warehousing using fact tables, dimension tables and star/snowflake schema modeling.

Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.

Developed parallel jobs using different processing stages like Transformer, Aggregator, Lookup, Join, Sort, Copy, Merge, Funnel, CDC, Change Apply and Filter.

Environment: IBM Information Server 8.0.1, Oracle, DB2.

IBM, Sr. ETL Developer

Bentonville AR - Sep 2009 to Nov 2013

Project Description:

Global Replenishment System (GRS) - a common global platform for replenishment planning and purchase order execution across all countries and formats. GRS is based on JDA forecast algorithms, a supply chain planning software, to perform the replenishment planning of Wal-Mart, considering item, supplier, sales history, events and other data from various source systems through ETL based batch interfaces. IBM is to deliver this data integration to ensure the functions as required by Wal-Mart business are met and implement the changes needed in the source systems, interfaces and the target system peripheral adapters.

Responsibilities:

Actively worked in all phases of the SDLC starting from requirements, design (LLD), development and unit testing, integration testing of all DataStage applications.

Interaction with client for analysis of Organization data, ETL requirement gathering & data Consolidation.

Take higher responsibilities to lead the team and deliver the good quality of code.

Worked in Data stage 8.0 PX job enhancements based on the new Requirements from Business.

Extensively worked in Performance tuning of DataStage jobs that runs long while monitoring the flow.

Used DataStage Designer for developing various jobs to extract, cleansing, transforming, integrating and loading data into Data Warehouse

Created and Managed the Defects that occurred in GRS project in HPQC tool.

Analysis of functional and Technical defects and provided the relevant data stage code fix or appropriate solution within the given ETA.

Maintained the different versions of code in POWER while deploying the code fixes to different phases like SIT, UAT.

Raised CRQ in Remedy7 tool for deploying the code in Production environment.

Environment: IBM Information Server 8.0.1, Oracle, DB2, Teradata, SQL server and CA7.

Wipro Infotech, ETL Developer

Bangalore, INDIA – Sep 2006 to Sep 2009

Project Description:

State Street is the world's leading provider of financial services to institutional investors. Our broad and integrated range of services spans the entire investment spectrum, including research, investment management, trading services and investment servicing.

Responsibilities:

Analysis and Understanding the process of the Requirement.

Involved in Creation of Design documents like TAD (Technical Analysis document) etc.

Designed the Jobs in Data stage 7.5.2 PX.

Developed mappings using the stages like Sequential file, Lookup Stage, Aggregator Stage, Transformer stage and etc.Created sequences which include the jobs in data stage project.

Running and monitoring of Jobs using DataStage Director and checking logs.

Involved in creating UNIX Shell scripts for Header and Trailer validation and other functional validation of source files/Flat files.

Created UTC (Unit Test Cases) and DIT (Development Integration Test Castes)

Involved in creation of Job Information Language (JIL) files in Autosys for scheduling ETL jobs.

Environment: Datastage 7.5, Oracle

Valtech India Pvt. Ltd, Associate Software Eng.

Bangalore, INDIA – Jul 2004 to Sep 2006

Client: Louis Vuitton

Project Responsibilities

Analysis and Understanding the process of the WIBI.

Interaction with the Source data owner for the data (SAP Team).

Writing Specification & Writing Unit Test conditions.

Design the Job in Datastage 7.5.2 PX. Developed mappings using the Datastage jobs using DataStage tool to load data warehouse and Data Mart.

Created sequences which include the jobs in data stage project.

Design and Develop ETL Performance tuning of Datastage jobs.

Environment: Datastage 7.5.2, DB2 and AIX, Control-M



Contact this candidate