RAMYA KOTARU
Mobile: 408-***-****; E-Mail: ******.*@*****.***
PROFESSIONAL SUMMARY
Over 7+ years of experience on ETL tools like IBM InfoSphere DataStage with strong business understanding and knowledge of Financial, Insurance and Banking projects. Hands on experience in all aspects of Software Development Life Cycle (SDLC) and Agile/scrum methodologies.
Expertise in IBM InfoSphere DataStage v7.5/8.5/8.7/9.1/11.5 tools like DataStage Designer, DataStage Director,DataStage Administrator and Expertise in Informatica PowerCenter v9.5 Designer tools like Source Analyzer, Warehouse Designer, Mapping Designer, Mapplet Designer, Transformation Developer, Workflow Manager and Workflow Monitor.
Have an overview to translate business rules/requirements into logical and physical models using Erwin tool.
Practical understanding of the Data modeling (Dimensional & Relational) data warehouse concepts like OLTP, OLAP, Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
Experience with Data Extraction, Transformation, and Loading (ETL) from disparate Data sources.
Excellent Experience in Designing, Developing, Documenting, Testing of ETL jobs and mappings in Server and Parallel jobs using IBM InfoSphere Data Stage to populate tables in Data Warehouse and Data marts.
Experience in building ETL jobs like EXTRACT, LOADS, BATCHES and SEQUENCERS.
Expert in designing Parallel jobs using various stages like Join, Merge, Lookup, Remove duplicates, Filter, Dataset, Complex flat file, Modify, Aggregator.
Extensively worked with Informatica Power Center Transformations such as Source Qualifier, Lookup, Filter, Expression, Router, Normalizer, Joiner, Update Strategy, Rank, Aggregator, Stored Procedure, Sorter, Union,Sequence Generator .
Experience in integration of various data sources (DB2-UDB, SQL Server, PL/SQL, Oracle and Teradata) into data staging area.
Implemented Slowly Changing Dimensions - Type I & II in Dimension tables as per the requirements.
Proven track record in troubleshooting of Data Stage jobs and addressing production issues like performance tuning and enhancement.
Proficient in writing, implementation and testing of triggers, procedures and functions in PL/SQL and Oracle.
Expertise in various types of ETL testing such as Integration, Volume, Performance, Structure Validation, Count Validation and Data Validation.
Experience in UNIX Shell scripting. Knowledge in Perl Scripting.
Strong skills in coding and debugging Teradata utilities like Fast Load, Fast Export, MultiLoad and Tpump for Teradata ETL processing huge volumes of data throughput.
Experience in Performance Tuning and production.
Experience on Autosys and Control-M (third party scheduling tool) for scheduling the jobs.
Excellent communication skills, good organizational skills, self-motivated and extremely hardworking with ability to implement business concepts quickly and efficiently.
Excellent communicator with exceptional team-building skills.
TECHNICAL SKILLS
RDBMS : Oracle 10g,Teradata,SQL Server,DB2
ETL Tools : Datastage 7.5Px/ 8.1/8.5/9.1/11.5, Informatica 9.1/9.5
Reporting Tools : Tableau
OS : Windows,UNIX
Scheduling Tools : Autosys, Cronacle,Control-M
Programming Language : Oracle SQL,PL/SQL,Perl
Data Modeling Tools : Erwin data modeler 7.3.8
Version control Tools : Subversion
EDUCATIONAL QUALIFICATION
MCA from J.K.C College, Guntur Affiliated to A.N.U
CERTIFICATIONS
IBM Certified Solution Developer for IBM InfoSphere Datastage v8.5
PROFESSIONAL EXPERIENCE
Project#1
Macys’s Inc, Johns Creek, Atlanta, GA Apr’2018-till date
DataStage Developer.
Description: INFOR –INFOR is a third party web application which is used as a financial planning tool within macy’s merchandising domain and its end to end systems for predicting the future plans depending on the current plan data taking into consideration the subject areas like Location,Period,Vendor .
Responsibilities:
Developing the jobs as per the Stories mentioned in the Iterations with regards to the sprints and showcasing as per the deadlines provided .
Wrote complex and medium queries in Teradata for creating the views.
Have written fast export script scripts to pull the data from Teradata tables to files.
Extracted data from source files, Teradata and Loaded into data warehouse.
Developed application views in Teradata.
Creation of datastage jobs as per the stories by using the stages like Sequential files, Lookup, Join, Sort, Aggregator, Transfomer, Copy etc
Created Sequencer jobs for controlling the job flow.
Involved in error handling, debugging and troubleshooting while testing.
Used Pre-SQL and Post-SQL scripts for loading the data into targets according to the requirement.
Provided the unit test documents .
Generated UNIX shell scripts for triggering/automating the execution of the Datastage jobs, encryption and decryption of data files; secure FTP to vendor sites, pre and post data processing/validations, and automated email notifications. Used Curl and wget functions to get the files from the cloud INFOR portal.
Worked on performance tuning of DataStage jobs.
Involved in writing test plan, Unit, Integration and Regression testing of the DataStage jobs and UNIX shell scripts.
Worked on deployment and production checkout planning of data and schema changes in the database.
Deployed the code to other environments by Jenkins and IBM Information Server Manager.
Coordinated with different infrastructure teams to execute deployments.
Environment: IBM InfoSphere DataStage v11.5/V11.7,Teradata, Unix, Control –M,Jenkins,JIRA
Project#2
Macys’s Inc, Johns Creek, Atlanta, GA Nov’2017-Apr’2018
DataStage Developer.
Description: ZMT -This project is to develop a new Zone Markdown Tool to assist the end users in analyzing the business . Current database setup has been moved to Teradata from SQL server. So as per the new database setup the database objects are created to support this new Mark down tool. This new mark down recommendation tool is all about marking down the prices to the aged products as per the recommended budget. In order justify this, a new Zone Markdown Recommendation report has been designed to analyze the business for making the decision.
Responsibilities:
Developing the jobs as per the Stories mentioned in the Iterations with regards to the sprints and showcasing as per the deadlines provided .
Wrote complex and medium queries in Teradata for creating the views.
Extracted data from Excel files, Teradata and Loaded into data warehouse.
Developed application views in Teradata.
Creation of datastage jobs as per the stories by using the stages like Unstructured file stage, Lookup, Join, Sort, Aggregator, Transfomer, Copy etc
Involved in error handling, debugging and troubleshooting while testing.
Used Pre-SQL and Post-SQL scripts for loading the data into targets according to the requirement.
Provided the unit test documents .
Generated UNIX shell scripts for triggering/automating the execution of the Datastage jobs, encryption and decryption of data files; secure FTP to vendor sites, pre and post data processing/validations, and automated email notifications. Used Curl and wget functions to get the files from the portal.
Worked on performance tuning of DataStage jobs.
Involved in writing test plan, Unit, Integration and Regression testing of the DataStage jobs and UNIX shell scripts.
Worked on deployment and production checkout planning of data and schema changes in the database.
Coordinated with different infrastructure teams to execute deployments.
Environment: IBM InfoSphere DataStage v11.5,Teradata,Excel files, Unix, Control -M.
Project#3
HCSC, Cognizant, Richardson,TX Apr’2017- Oct'2017
Data Warehousing Analyst.
Description: Health Care Service Corporation (HCSC) is the licensee of the Blue Cross and Blue Shield Association for 5 states. It concentrates its operations in Illinois, Montana, New Mexico, Oklahoma, and Texas.As the 4th largest health insurer in the US overall, it employs more than 22,000 people and serves nearly 15 million members. It offers group life, disability, and dental solutions, and a range of other individual solutions. The company also provides various care management and wellness resources.
Responsibilities:
Provided timely resolution of production issues, development and implementation of enhancements and change request from the management team.
Wrote complex and medium queries in DB2 for analyzing the data.
Extracted data from Flat files, DB2 and Loaded into data warehouse.
Developed application views in DB2 .
Involved in error handling, debugging and troubleshooting the jobs using log files.
Used Pre-SQL and Post-SQL scripts for loading the data into targets according to the requirement.
Provided the template to the Zena jobs to schedule the run of the mappings.
Implemented Slowly Changing Dimensions (SCD Type 2) to update the dimensional schema and used IBM CDC tool to get the recent updates.
Generated UNIX shell scripts for triggering/automating the execution of the Datastage jobs, encryption and decryption of data files; secure FTP to vendor sites, pre and post data processing/validations, and automated email notifications.
Worked on performance tuning of SQL queries in DataStage jobs.
Involved in writing test plan, Unit, Integration and Regression testing of the DataStage jobs and UNIX shell scripts.
Worked on deployment and production checkout planning of data and schema changes in the database.
Coordinated with different infrastructure teams to execute deployments.
Coordinated with users in incident management in processing the claims
Environment: IBM InfoSphere DataStage v8.5, DB2, Flat files, Unix, Zena, IBM CDC tool.
Project#4
MDSS, Wells Fargo, Hyderabad, India June'2015- March’2016
DataStage Developer.
Description: Wells Fargo is an American multinational banking and financial services holding company operates cross 35 countries and has over 70 million customers globally .MDSS is a data mart comprised of all accounts from all lines of business within wells fargo. MDSS presents data from both the Household and the customer perspective. Users of MDSS perform analysis, trending, Ad hoc queries, generate market campaigns and more.
Responsibilities:
Analyzed, designed, developed, implemented and maintained Parallel jobs using IBM info sphere Data stage.
Adept knowledge and experience in mapping source to target data using IBM Data Stage 8.x.
Experience in PX file stages that include Complex Flat File stage, DataSet stage, LookUp File Stage, Sequential file stage.
Implemented Shared container for multiple jobs and Local containers for same jobs as per requirements.
Implemented multi-node declaration using configuration files (APT_Config_file) for performance enhancement.
Successfully implemented pipeline and partitioning parallelism techniques and ensured load balancing of data.
Worked on different partitioning methods like Hash by column, Round Robin, Entire, Modulus, and Range for bulk data loading and for performance boost.
Used ETL Data Stage Director to schedule and run the jobs, test and debug its components & monitor performance statistics.
Created Batches (DS job controls) and Sequences to control set of jobs. Executed jobs through sequencers and created batch jobs for better performance and easy maintenance.
Debugged, tested and fixed the transformation logic applied in the parallel jobs
Involved in creating UNIX shell scripts for database connectivity and executing queries in parallel job execution. Knowledge on NDM file transfers.
Participated in build/review of the BTEQ Scripts, FastExports, Multiloads and Fast Load scripts.
Performed unit testing and provided the test cases.
Scheduled the jobs using AutoSys and Corntab.
Responsible for daily verification that all scripts, downloads, and file copies were executed as planned, troubleshooting any steps that failed, and providing both immediate and long-term problem resolution.
Provided technical assistance and support to IT analysts and business community.
Provided maintenance support for MDSS month end loads.
Environment: IBM InfoSphere Datastage v8.5, Cronacle, Teradata, Oracle 10g, subversion, UNIX,
SQL*Plus.
Project#5
AIG Chartis Insurance, IBM, Parsipanny, NJ Apr'2013- May’2015
DataStage Developer.
Description: AIG is the world’s largest insurance company. AIG had assets of approximately $1 trillion, $110 billion in annual revenues, 74 million customers and 116,000 employees in 130 countries. Chartis is a world leading property-casualty and general insurance organization serving more than 40 million clients in over 160 countries.
Responsibilities:
Provided Technical support to the team as the ETL developer. Addressed best practices and productivity enhancing issues.
Loaded data into load, staging and lookup tables. Staging area was implemented using flat files.
Created jobs in DataStage to import data from heterogeneous data sources like Oracle 9i, Text files.
Extensively worked on Job Sequences to Control the Execution of the job flow using various Activities & Triggers (Conditional and Unconditional) like Job Activity, Wait for file, Email Notification, Sequencer, Exception handler activity and Execute Command.
Assisted Mapping team to transform the business requirements into ETL specific mapping rules.
Enhanced various complex jobs for performance tuning.
Responsible for version controlling and promoting code to higher environments.
Performed Unit Testing, System Integration Testing and User acceptance testing
Involved in ongoing production support and process improvements. Ran the DataStage jobs through third party schedulers Autosys.
Involved with the batch team in preparing the JILs.
Utilized Parallelism through different partition methods to optimize performance.
Developed DS jobs to populate the data into staging and Data Mart.
Performed Unit testing for jobs developed to ensure that it meets the requirements.
Developed UNIX shell scripts to automate file manipulation and data loading procedures.
Scheduled the jobs using AutoSys.
Involved in some of the admin tasks for deleting the datasets. Restarting the servers and maintaining the sftp connections.
Creating of New DataStage projects and maintenance.
Autosys setup for new interfaces.
Rebooting the servers as part of the regular maintenance activities and performing health checks.
Code migration to different datastage projects .
Sftp setup between servers.
Monitoring error and warning less data transmission from source to target.
Develop and implement strategies for performance tuning.
Performed unit testing and provided the test cases.
Environment: IBM InfoSphere Datastage v8.5 & v9.1, Autosys,Oracle 10g, Erwin, UNIX,PVCS Serena Dimensions.
Project#6
Stonesoup, IBM, Foster City, CA Jun’2011- March'2013
DataStage Developer.
Description: Stonesoup project is an Industry-based data integration model solution set. The purpose is to develop and package a set of components and accelerators that focus on the Data Integration space, providing IBM the ability to demonstrate a broader view of Business Intelligence. Solution sets include Banking, Healthcare, Insurance and Life science data warehouse models and the associated Subject areas.
Responsibilities:
Worked on DataStage tools like DataStage Designer, DataStage Director and DataStage Administrator.
Strong understanding of the principles of Data Warehousing using fact tables, dimension tables Extensive ETL tool experience using IBM Infosphere/Websphere DataStage, Ascential DataStage.
Experience in Data modeling strategies star/snowflake schema modeling.
Knowledge in using Erwin as leading Data modeling tool for logical (LDM) and physical data model (PDM).
Developed parallel jobs using different processing stages like Transformer, Aggregator, Lookup, Join, Sort, Copy, Merge, Funnel, CDC, Change Apply and Filter.
Experience in clear understanding of design goals of ER modeling for OLTP and dimension modeling for OLAP.
Assisted in development efforts for Data marts and Reporting.
Worked SCDs to populate Type I and Type II slowly changing dimension tables from several operational source files
Generation of Surrogate IDs for the dimensions in the fact table for indexed and faster access of data in server jobs.
Extensively worked on Job Sequences to Control the Execution of the job flow using various Activities & Triggers (Conditional and Unconditional) like Job Activity, Wait for file, Email Notification, Sequencer, Exception handler activity and Execute Command.
Extensive experience in Unit Testing, Functional Testing, System Testing, Integration Testing, Regression Testing, User Acceptance Testing (UAT) and Performance Testing.
Prepared the test documents.
Fixing of code and data related bugs in the existing application.
Internal process review and Peer to Peer review.
Process documentation and revision as and when required.
Environment: IBM InfoSphere Datastage v8.5, Erwin DataModeler, Autosys, DB2, UNIX, SQL*Plus.
Project #7
CITRUS, IBM, Hyderabad, India Sep’ 2010- Jun’2011
DataStage Developer.
Description: The Contractor Invoicing, Tracking and Reporting System (CITRuS) is serves as an interface to contractor information for use worldwide in managing costs, vendor relationships, and use strategies, while optimizing labor resources.
Business Objective:
Provide instant interactive visibility to consistent, complete contractor information for use in:
Optimizing labor resources (e.g., headcount, skills, rates)
Managing contractor spending (e.g. visibility to rates, spending; automated invoice reconciliation; support for managing PO liabilities)
Managing vendor relationships and vendor use strategies
Enforce common process to manage contractor invoices tracking and reconciliation
Capture and maintain contractor information with minimal manual intervention
Responsibilities:
Based on the Business Requirements developed Data stage Jobs to Extract, Transform and Load the data from Source to Target and Target to Data Distribution Area.
Expert in designing Server jobs using hashed file stages, Link partitioner, Link collector stages and Sequencer jobs using Job activity, wait-for-file activity, Terminator etc.
Used Data Stage Designer to develop process for extracting, cleansing, transforming, integrating and loading the data into IBM DB2 Data warehouse.
Used extensively Reject Link, Job Parameters, and Stage Variables in developing jobs.
Involved in job level performance tuning.
Involved in promoting the jobs by version controlling from development to integration.
Assists in Production support & Fixing the production issues as back up.
Used Transformer, Remove Duplicates, Copy, Funnel, Lookup, and Change Capture Stages in designing jobs.
Involved in Unit Testing, Integration testing.
Import and Export of jobs using Data Stage Manager.
Involved in Different reviews like Internal and external code review, weekly status calls, issue resolution meetings and onsite code acceptance meetings.
Environment: IBM InfoSphere Datastage 7.5PX,DB2,UNIX.