Prasanna Lakshmi R adcgl4@r.postjobfree.com +1-479-***-****
PROFESSIONAL SUMMARY
ETL professional having 15+ Years IT experience and track record in Analyzing, Designing, Developing, Testing, Implementing and maintaining client/server business systems.
Having more than 9 years of ETL development experience and 6+ years of QA experience in cloud-based environments with Extraction Transformation and Loading DataStage (ETL) tool –Knowledge in Data warehousing involving data warehouse analysis and design using IBM INFORMATION SERVER (Datastage)/Talend/SSIS, Apache Airflow, Apache Kafka.
Excellent QA experience in all phases of Software Testing Life Cycle (STLC) and Software Development Life Cycle (SDLC) with good working knowledge of QA methodologies, disciplines, tasks, resources and documenting.
Experience in Integration, Functional, System, Regression, SIT, Stress, Black Box, White Box and coordinating User Acceptance Testing (UAT).
Strong experience on Bigdata Hadoop Testing & Extensive experience in planning and executing strategy for testing the Database, DWH, ETL, Data Migration projects from various legacy systems to target database in cloud-based environments.
Expert in writing SQL and PL/SQL queries and worked with Bigdata/NoSQL/Cloud technologies- HIVE, Big Query in Google Cloud Platform (GCP)
Immense experience in testing Cloud based data pipelines and building Google Big Query SQL for data validation.
Worked on different Scheduling tools like CA7, AutoSys and good experience in writing JIL scripts and designing Autosys jobs using Web GUI and proven track record in troubleshooting of DataStage jobs and addressing production issues.
Skilled in writing technical specification documents, translating user requirements to technical specifications and creating and reviewing mapping documents.
SKILLS
ETL Tools
IBM Information Server 8x, Data stage 7.5x (Server & PX), Informatica (Beginner), Talend Studio.
Testing Tools
Test Link, VersionOne, JIRA, HPALM, Qtest
Languages
Unix shell scripting, SQL, Python (Basics), PL/SQL
Databases
Oracle 8.1, SQL Server 6.5, DB2, Teradata, MySQL, MongoDB, DB2, Big Query, Apache Hive
Job Scheduling
CA7, Control-M
Operating System
UNIX, WINDOWS, LINUX
Web Service Tools
SOAP UI, Rest API - Postman
Domains
Retail, Finance, banking, Securities and Electronics
CERTIFICATIONS
Certified in Infosphere DataStage 8.0.
Big Query
Machine Learning with Apache Spark
PROFESSIONAL EXPERIENCE
GSPANN Technologies, Sr. ETL Consultant
Milpitas CA - Nov 2013 to Till Date
Kohls - EDW Migration:
EDW migration project is to migrated all the data from legacy (Teradata) Database into Big Query Database in Google Cloud Platform. The data from different Business areas like OMS (Order Management system), Products, Vendor, Inventory, sales, marketing data is migrated to Big query.
Responsibilities:
Review Business requirements, functional specifications, use cases, and test plans.
Based on the STM (mapping document) /requirements, we prepare the Test scripts in Teradata and Big query and do the comparison to validate all the business transformations.
Historical & Incremental load data validation is performed based on multiple sources like DB2, Oracle, Flat files, Kafka being migrated to Big Query.
Validate all the data loaded related to OMS (Order Management system), Products, Vendor, Inventory, sales, marketing data is migrated to Big query through Apache Airflow into Big Query.
Validated the tokenized customer data sent in Encrypted files and loaded into BigQuery.
Executed the testcases and updated the test results in Qtest and raised the defects in JIRA for tracking.
Analyze failures and document software defects using bug tracking systems like report the defects involving program functionality, output, online screen and content to software developers. Follow up with developers, support and other teams on the resolution.
Validation of the data loaded into Audit tables for each EDW data load.
Study and analysis of the mapping document indicating the source tables, columns, data types, transformations required, business rules to be applied, target tables, columns and data types
Developed various Test Scripts, Test Execution using TestLink based on the Functional Specifications.
Writing complex SQL and PL/SQL queries using Case Logic, Intersect, Minus, Sub Queries, Inline Views, and Union.
Developed SQL queries for Back-end testing/Database testing and participated in end to end testing.
Performed Functional Testing, Regression Testing, End to End testing.
Provided daily and weekly status reports.
Environment: Apache Airflow, Teradata, DB2, Oracle, Big Query, ESP scheduler, JIRA, Qtest
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Kohls - ECMDM (Enterprise Customer Master Data Management) data Analytics project.
The main role of ECMDM is to serve as a system of reference for customer master data. ECMDM collects customer data from all available source systems to form a single customer record that combines all data available on a customer. In this project we bring various types of data comes from multiple data sources like First data, KCC (Kohls Credit card), 3CC, (Third Party Credit card), Loyalty, ATG (webstore), Sales Alert data in flat files, Kafka, which we will be loaded to HIVE staging DB and MYSQL Database on cloud.
Responsibilities:
Review Business requirements, functional specifications, use cases, and test plans.
Prepared unit testcases which included different functional and technical scenarios.
Validate all the data loaded related to 3CC, KCC, Loyalty, ATG which are migrated from legacy application to New system though Batch jobs into Dimension tables like CUSTOMER, EMAIL, PROFILE, ACCOUNT, CREDIT CARD (MYSQL and Hive DB).
We do connect to HIVE/MYSQL database and validate the source data is loaded accurately based on the requirements.
The data is related to all customer information, which is tokenized. We validate the tokenized data by detokenizing on the HIVE server.
All the customer sensitive data is sent in Encrypted files and loaded into MYSQL.
Validate the Real streaming data sent through Kafka MQ Topics into MYSQL.
Validation of RestAPI services through Postman to Get/Post customer data from MYSQL.
Executed the testcases and updated the test results and raised the defects in JIRA for tracking.
Analyze failures and document software defects using bug tracking systems like report the defects involving program functionality, output, online screen and content to software developers. Follow up with developers, support and other teams on the resolution.
Validation of the data loaded into Audit tables for each data load.
We perform Regression testing as well as part of each story.
Environment: Hive, MYSQL, Batch jobs, Flat files, XML, Kafka (MQ), RestAPI postman, Protegrity tool.
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Rodan & Fields (R+F) – ATLAS Project:
ATLAS project scope is for migrating heirloom application to Hybris and changing all the data sources to new system for maintaining product/Account/order related info for different types of customers like Preferred Customer (PC), Retail Customer (RC) and Consultants in MSSQL Database. On top of the migration implementation of promotions data load to DWH system using ETL (SSIS). Migration of SSRS reports to MSTR (Micro Strategy) reports. Integration of Amazon S3 with MicroStrategy Reporting to enable reporting for Account/Order/Product Audit Reports
Responsibilities:
Review Business requirements, functional specifications, use cases, and test plans.
Work closely with the Business, Solution Architect, Data Modeler, Data Mapper, Business Analysts to understand business requirements, providing expert knowledge and solutions on Data Warehousing, ensuring delivery of business needs in a timely cost-effective manner
Prepared unit testcases which included different functional and technical scenarios.
Validate all the data loaded related to Products, Customers, Orders, Shipments & Payments which are migrated from legacy application to New system though ETL SSIS package into EDW (SQL Server).
Validation of MicroStrategy reports which are built on top of the EDW loaded Tables, Stored procedures & Views etc.
Validated the migrated reports functionality from SSRS to MSTR based on the data comparison.
Executed the testcases and updated the test results and raised the defects in JIRA for tracking.
Analyze failures and document software defects using bug tracking systems like report the defects involving program functionality, output, online screen and content to software developers. Follow up with developers, support and other teams on the resolution.
Validation of the data loaded from AWS (S3) Athena Audit tables to EDW tables.
Environment: SSIS, SQL Server, MSTR Reports, AWS S3 services.
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Lam Research – OPT Project
Online Planning Tool (OPT): The project is to create OPT dashboard to replace exiting OPT application and allow the business users and the IT to query their near real time, ad-hoc, standard reports, operational and analytical reports from SAP Hana.
Responsibilities:
Serve as technical lead and work with business partners to define product roadmap and requirements for SAP data migration program.
Interact closely with key business stakeholders to understand user demands; product architects, development team to understand technical application dependencies and help build finer and detailed product specifications packaging features/functionality into phased product releases.
Communicate functional and non-functional requirements via user stories to dev teams.
Specify data mapping & data provisioning methodologies along with application interface detailing.
Handled migration of legacy applications from SQL server to HANA models, enabling faster performance, better memory utilization and easier access to simpler interface modeling for future.
Evaluate and design data collection, data staging, data movement, analytics delivery, data quality and archiving strategies
Review the jobs for OPT ETL applications created using DataStage/BODS tool, which loads data into CSV’s, and TAB’s. These files are created by extracting the data from Oracle DB using PL/SQL and DataStage Technologies and the files are sent to down Stream applications.
Identify opportunities to optimize the ETL environment, implement monitoring, quality and validation processes to ensure data accuracy and integrity.
Extensive ETL tool experience using IBM Infosphere /WebSphere DataStage, Ascential DataStage.
Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.
Environment: IBM Information Server 8.0.1, BODS, SAP HANA, SQL Server, JIRA
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
Macys – STELLA Project
Stella is a Product Information Management System (PIM). All Product information and product images on the Macys.com website and Bloomingdales.com are managed in Stella. Stella manages Product assets including module like PRODUCT, ATTRIBUTES, PRICING, PROMOTIONS, UPC’s etc.
Responsibilities:
Work closely with the Business, Solution Architect, Data Modeler, Data Mapper, Business Analysts to understand business requirements, providing expert knowledge and solutions on Data Warehousing, ensuring delivery of business needs in a timely cost-effective manner
Evaluate and design data collection, data staging, data movement, analytics delivery, data quality and archiving strategies
Developed the jobs for Stella ETL applications using DataStage tool, which loads data into CSV’s, and TAB’s. Extracting the data from Oracle DB using PL/SQL creates these files and DataStage Technologies and the files are sent to down Stream applications.
Involved in unit testing of ETL DataStage jobs.
Identify opportunities to optimize the ETL environment, implement monitoring, quality and validation processes to ensure data accuracy and integrity.
Extensive ETL tool experience using IBM Infosphere/WebSphere DataStage, Ascential DataStage.
Worked on DataStage tools like DataStage Designer, DataStage Director.
Strong understanding of the principles of Data Warehousing using fact tables, dimension tables and star/snowflake schema modeling.
Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.
Developed parallel jobs using different processing stages like Transformer, Aggregator, Lookup, Join, Sort, Copy, Merge, Funnel, CDC, Change Apply and Filter.
Environment: IBM Information Server 8.0.1, Oracle, DB2.
IBM, Sr. ETL Developer
Bentonville AR - Sep 2009 to Nov 2013
Project Description:
Global Replenishment System (GRS) - a common global platform for replenishment planning and purchase order execution across all countries and formats. GRS is based on JDA forecast algorithms, a supply chain planning software, to perform the replenishment planning of Wal-Mart, considering item, supplier, sales history, events and other data from various source systems through ETL based batch interfaces. IBM is to deliver this data integration to ensure the functions as required by Wal-Mart business are met and implement the changes needed in the source systems, interfaces and the target system peripheral adapters.
Responsibilities:
Actively worked in all phases of the SDLC starting from requirements, design (LLD), development and unit testing, integration testing of all DataStage applications.
Interaction with client for analysis of Organization data, ETL requirement gathering & data Consolidation.
Take higher responsibilities to lead the team and deliver the good quality of code.
Worked in Data stage 8.0 PX job enhancements based on the new Requirements from Business.
Extensively worked in Performance tuning of DataStage jobs that runs long while monitoring the flow.
Used DataStage Designer for developing various jobs to extract, cleansing, transforming, integrating and loading data into Data Warehouse
Created and Managed the Defects that occurred in GRS project in HPQC tool.
Analysis of functional and Technical defects and provided the relevant data stage code fix or appropriate solution within the given ETA.
Maintained the different versions of code in POWER while deploying the code fixes to different phases like SIT, UAT.
Raised CRQ in Remedy7 tool for deploying the code in Production environment.
Environment: IBM Information Server 8.0.1, Oracle, DB2, Teradata, SQL server and CA7.
Wipro Infotech, ETL Developer
Bangalore, INDIA – Sep 2006 to Sep 2009
Project Description:
State Street is the world's leading provider of financial services to institutional investors. Our broad and integrated range of services spans the entire investment spectrum, including research, investment management, trading services and investment servicing.
Responsibilities:
Analysis and Understanding the process of the Requirement.
Involved in Creation of Design documents like TAD (Technical Analysis document) etc.
Designed the Jobs in Data stage 7.5.2 PX.
Developed mappings using the stages like Sequential file, Lookup Stage, Aggregator Stage, Transformer stage and etc.Created sequences which include the jobs in data stage project.
Running and monitoring of Jobs using DataStage Director and checking logs.
Involved in creating UNIX Shell scripts for Header and Trailer validation and other functional validation of source files/Flat files.
Created UTC (Unit Test Cases) and DIT (Development Integration Test Castes)
Involved in creation of Job Information Language (JIL) files in Autosys for scheduling ETL jobs.
Environment: Datastage 7.5, Oracle
Valtech India Pvt. Ltd, Associate Software Eng.
Bangalore, INDIA – Jul 2004 to Sep 2006
Client: Louis Vuitton
Project Responsibilities
Analysis and Understanding the process of the WIBI.
Interaction with the Source data owner for the data (SAP Team).
Writing Specification & Writing Unit Test conditions.
Design the Job in Datastage 7.5.2 PX. Developed mappings using the Datastage jobs using DataStage tool to load data warehouse and Data Mart.
Created sequences which include the jobs in data stage project.
Design and Develop ETL Performance tuning of Datastage jobs.
Environment: Datastage 7.5.2, DB2 and AIX, Control-M