ac4mg9@r.postjobfree.com
LinkedIn URL : www.linkedin.com/in/venkatesh-gilla-ramamoorthy-75255a17
SUMMARY
12+ years of experience in Requirement gathering, Gap analysis, Designing, Coding, testing, implementation, Production support, Resource Management in Data warehousing with business knowledge of Insurance, Pharmaceutical, Telecom, Banking and Travel domains
Good working experience in SDLC methodologies like waterfall, Agile
Strong Datawarehouse ETL experience of using Informatica Power Center 10.1/9.6/9.1/8.6.1/7.1 PowerCenter Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor
Experienced with Data Extraction, Transformation and Loading(ETL) from disparate Data Source databases like Oracle, SQL Server, flat files, CSV files, XSD and XML files and loaded into target warehouse by using various transformations in Informatica
Worked with different Informatica transformations like Aggregator, Lookup, Joiner, Filter, Router, Update strategy, Transaction Control, Union, Normalizer, SQL in ETL development
Experienced in Debugging and Performance tuning of targets, sources, mappings and sessions and experience in optimizing the Mappings and implementing the complex business rules by creating re-usable transformations and Mapplet.
Worked on scalability and performance tuning batch processing to handle large volumes.
Extensive experience in writing UNIX shell scripts and automation of the ETL processes using UNIX shell scripting
Experience in resolving on-going maintenance issues and bug fixes; monitoring Informatica sessions as well as performance tuning of mappings and sessions
Having good Experience on Scheduler Tool like Tidal, AutoSys, Control M, ActiveBatch
Working knowledge on Informatica Master Data Management (MDM)
Excellent interpersonal and communication skills, and experienced in working with senior level managers, business users, and developers across multiple disciplines
Effectively managed globally dispersed teams of up-to 20 members
Technical Skills:
ETL Tools
Informatica Power Center 10.1/9.6/9.1/8.6.1/ 7.1
Data Modeling
Dimensional Data Modeling, using Star Join Schema Modeling, Snowflake Modeling, FACT and Dimensions Tables, Physical and Logical Data Modeling
Databases
Oracle 11g/10g/9i, SQL Server
Scripting
Unix Shell scripting, SQL stored procedures
Scheduler Tools
Control -M, Autosys, Tidal, ActiveBatch
MDM Tool
Informatica/Siperian MDM, IDD, SIF
DB tools
SQL / PL SQL, TOAD, Oracle Developer
Operating System
UNIX (Putty, WinSCP), Windows
Versioning tools
TFS, SVN
Release Management tools
SWIM, IBM Urban Code (uDeploy)
Testing Tools
HP Quality Center, QTP
Academic proliferation
Master of Financial Management (M.F.M) from Pondicherry University, India, 2010
Bachelor of Commerce (B.Com) from AVS College of arts & science, Periyar University, India 2001
Employment Summary
From
To
Duration
Company Name
Designation
Role
Dec '16
Present
Horizon International
ETL Lead
ETL Lead
Sep '15
Nov '16
1 year 3 months
Syntel Inc
Project Manager
ETL Lead
Dec ‘14
Sep ‘15
9 months
Formac Inc
Senior System Analyst
Project Lead
Mar‘08
Dec ‘14
7 years 9 months
Cognizant Technology Solutions
Senior Associate
Project Lead
Apr ‘06
Feb ‘08
1 year 11 months
Verizon Data services India P Ltd
Analyst
Developer
PROJECT DETAILS
GCC, Boston, MA Feb 2017 – Jan 2018
Position: Informatica lead
GCC is a global enterprise comprised of a family of travel companies. Few of its major divisions are River cruise, Small ship cruise and Overseas adventure travel. The Data Service team (Data warehouse), plays an centric role in maintaining data quality and data governance, which sourced from different systems like web, call center, etc).
Responsibilities:
Lead design, development and implementation of the ETL projects end to end
Involved in the analysis of source to target mapping provided by data analysts and prepared function and technical design documents
Responsible for ETL technical design discussions and prepared ETL high level technical design document
Extracted data from flat files, XML, Oracle,Sql server using Informatica ETL mappings and loaded to Data Mart
Created complex Informatica mappings using transformations Unconnected Lookup, joiner, Rank, Source Qualifier, Sorter, Aggregator, Lookup and Router transformations to extract, transform and loaded data mart area
Worked extensively on shell scripting for file management
Created re-usable transformations/mapplets and used across various mappings
Developed Informatica workflows/worklets/sessions associated with the mappings using Workflow Manager
Created pre-session, post session,pre-sql,post sql commands in Informatica
Created and maintained Active Batch jobs to schedule Informatica Workflows
Expert in performance tuning of Informatica code using standard informatica tuning steps
Work closely with DBAs, application, database and ETL developers, change control management for migrating developed mappings to PROD
Responsible for Production support on rotation basis among team members
Attend release meeting, clarification meeting with BA/end users & sprint retrospective meeting
Provide weekly status report to Project Manager
Skills Used: Informatica 10.1, Oracle 12c, MS SQL Server 2008, Toad, UNIX, Putty, ActiveBatch Scheduler, Team management, Agile
Aegon Asset Management, Cedar Rapids, IA
Feb 2015 – Dec 2016
Position: Informatica lead
Aegon Asset Management (AAM) is a leading global investment manager with $375 billion in assets managed/advised by about 250 investment professionals in North America, Europe and Asia.
Investment Data Warehouse (IDW) acts as a is AAM’s centralized reporting repository. IDW extracts and holds financial information received from multiple investment source systems, as well as market data. Data is cleansed, transformed and validated for accuracy before storing the data into one central storage repository. The system provides our business units a consolidated view of data for query and reporting across all asset classes
Responsibilities:
Involved in all phases of SDLC from requirement, design, coding, testing, deployment and production support
Worked extensively in Informatica Power Center Repository Manager, Designer, Workflow Manager/Monitor
Used the techniques like Incremental aggregation, Incremental load and Constraint based loading for better performance.
Created Reusable Transformations, Mapplets, Sessions and Worklets and made use of the Shared Folder concept using shortcuts wherever possible to avoid redundancy.
Expertise in designing and implementing SCD - slowly changing dimensions types (1, 2 and 3) and CDC – change data capture
Involve in tuning the mappings in the transformations by tracking the reader, writer, transformation threads in the session logs and used tracing level to verbose during development & only with very small data sets
Involved in DB Performance tuning of different tables which has huge data by creating partitions and using stored procedures.
Used Oracle Index techniques, partitioning concepts on DB to improve query performance
Used various debugging techniques and informatica debugger tool to debug the mappings
Create/modify Control M jobs to schedule workflows
Provide estimations on ETL deliverables and oversee the quality of ETL Deliverables
Involved in Unit testing, Iterative testing to check whether the data loads into target are accurate, which was extracted from different source systems as per the user requirements
Support for code deployment in QA and PROD environment and responsible for code validation
Skills Used: Informatica 9.6, Oracle 11g, SQL Server 2008, SQL Developer, Toad, UNIX, Putty, WinSCP, Control-M Scheduler, Team Foundation Server (TFS), BizTalk, Team management, Agile
Sanofi - NA AMS, Bridgewater, NJ
May 2014 – Dec 2014
Position: Transition Lead
Sanofi is the diversified global healthcare leader, focused on patients’ needs. Sanofi operates in three primary business areas: Pharmaceuticals, Human Vaccines and Animal Health.
The objective of Sanofi - North America Application Maintenance Services (NA AMS) is to maintain all the Sanofi applications under one roof. The supported applications are CCT (Common Customer Table) which centralizes customer information across the system; SRT (Spend Review Tool) which centralizes all the expenses related to clinical trials, meetings, product promotions.
Responsibilities:
Manage and distribute work to offshore team and lead for timely and defect free deliverables
Handle 8 offshore resources and individual task allocation
Responsible to lead Knowledge transition from Prime vendor to CTS for all the client application. Need to ensure that all KT have documented and capture video of all KT sessions
Involved in requirement gathering on new requirements /enhancements. And responsible for preparing and getting sign off from client on requirement understanding document, design document and Unit test case scenarios
Worked extensively in Informatica Power Center Repository Manager, Designer, Workflow Manager and Workflow Monitor
Worked with flat file sources, CSV sources and Oracle sources.
Extensively used shell scripting for FTP of files, renaming the files, etc. Wrote scripts to run workflows using pmcmd command from the UNIX server.
Extensively worked with Aggregator, Sorter, Lookup, Router, Expression, Filter, Update Strategy, and Stored Procedure transformations.
Created reusable transformations and mapplets for better re-usability.
Created reusable sessions and workflows and extensively used parameter files and variables.
After development, performed extensive unit testing to check the proper and timely flow of data. Performed tuning of sessions by monitoring the cache size of the transformations
Support for code deployment in QA and PROD environment and also responsible for code validation
Responsible for Production support activities like Jobs monitoring, log the load statistics, analysis and resolve incase of any production issues, coordinate with business to fix any source file issues, coordinate with DBA to resolve any table space or file system related issues. And also responsible for deliverables out of daily/weekly/monthly jobs
Responsible for resolving the tickets raised by end users and adhoc requests from client. This includes the data analysis, manipulating the production data to fix the issues (with client approval) and Close the ticket with resolution
Analysis existing code/process and apply performance techniques
Coordinate with other application teams, in order to avoid production job conflicts due to dependency
Responsible for status report to management on weekly basis
Skills Used: Informatica 9.1, Oracle, UNIX, Tidal Scheduler, Team management
ING Annuity - Risk Management,
(West Chester, PA / Chennai, India)
Jul 2009 – Apr 2014
Position: Informatica Lead
ING offers a comprehensive array of financial services to retail and institutional clients. ING Group offers banking, investments, life insurance and retirement services to over 85 million private, corporate and institutional clients in more than 40 countries. The objective of this project is to implement Pegasus Datawarehouse system. The ETL process can be categorized in three phases like, source to Staging, Staging to Pegasus data warehouse and Algo process to the Pegasus data warehouse. The Staging area represents the assimilation of data from legacy systems like: Market, Reference, Garwin, Hartford and Bloomberg. All these integrated data are loaded to Pegasus datawarehouse. The Pegasus data sourced and processed through Algo process (Apply calculations using C++ Algorithm). The output of Algo process acts as a source for Pegasus system. The files will be loaded in to tables of Pegasus database.
Responsibilities:
Analyze the requirements and arrange for clarification sessions with Business to complete the requirement analysis
Identify the requirement gap to analysis and get a clear understanding before the start of the coding design and also an exercise to minimize the requirement defects
Responsible for Requirement Traceability Matrix
Review & finalizing the coding and deliverables
Design and create Autosys jobs to perform various tasks like file watcher jobs, executing informatica workflow etc. (creation of JIL)
Involved in implementation of ETL performance tuning techniques like Session Partitioning, Incremental Aggregation, Indexing lookup condition columns, optimizing SQL query, etc
Point of contact for project audits. The audit includes the study of the issues and risk to the project. Based on the monthly project audit, the project health score will be calculated. Also as part of audit, the codes will be verified, in order to check, whether all the industry coding standards has been adhered.
Prepare QA/PROD implementation plan and support on QA/PROD code deployment
Responsible for code validation after post deployment.
Study existing Informatica, Oracle codes and suggest the best practice / enhancements to team
Responsible for Level 2 production support, which includes analysis and resolution within specified SLA
Responsible to update production job run books and inform Level 1 production support, in case of any job impact due to the new requirement / enhancements
Responsible for task allocation and Project management
Carryout the project quality assurance demonstration to project stakeholders
Involve in project core team meeting to streamline the requirement and process
Skills Used: Informatica 9.1, Oracle 11g, Toad, UNIX, Putty, WinSCP, SVN, Autosys Scheduler
Bank of New York Mellon, Chennai, TN
Apr 2009 – Jun 2009
Position: Informatica Developer/ Lead
BNYM is global financial services company operating in 34 countries. BNY Mellon is a leading provider of financial services for institutions, corporations and high-net-worth individuals. BNYM led by its own management team and focuses on fulfilling the product and service needs of the independent agents in its respective region. BNYM provides superior services in the following areas like, Asset Servicing, Asset and Wealth Management, Issuer Services and Treasury Services.
The BNY Mellon Hub is the information backbone of the Bank of New York Mellon account. It serves as the focal point for the day-to-day operations which include Maintaining Associate information, Work order management, Knowledge sharing, reporting and much more. The sub-project of the BNYM is the ‘Fund system’, acts as the leaf of the Asset and Wealth Management service. This project uses a two tiered stage approach; the data will be loaded into the EDW in three basic steps. Step one is extracting atomic data from the application systems into an acquisition stage database, Step two will involve performing all the EDW logic to move the data into another stage database with the same overall structure as the EDW base tables and Step three is the move from the base stage database to the EDW base tables. The main objective of this project is to reduce the manual interventions and automate the existing process and to tune the existing ETL Informatica objects and Oracle objects to improve the code execution performance.
Responsibilities:
Developed mappings, sessions, and workflows in Informatica PowerCenter
Developed standard and reusable mappings and mapplets using various transformations like Expression, Aggregator, Joiner, Router, Lookup (Connected and Unconnected) and Filter
Identified performance issues in existing sources, targets and mappings by analyzing the data flow, evaluating transformations and tuned accordingly for better performance
Modified the shell scripts as per the business requirements
Involved in Unit testing, Iterative testing to check whether the data loads into target are accurate, which was extracted from different source systems as per the user requirements
Prepared and used test data/cases to verify accuracy and completeness of ETL process
Skills Used: Informatica 7.1, Oracle SQL/ PL/SQL, TOAD, MicroStrategy, UNIX, Team management
ING,( Tokyo, Japan / Chennai, India)
Mar 2008 – Feb 2009
Position: Development Lead
ING Life is a NETEHRLANDS based Life Insurance Company currently holds two individual systems for Life-A and Life-J. This project is to implement a Data Integration solution to load the data from the existing two source systems to the three layers: Staging, Information HUB and ODS.
Data analysis and data profiling are done on source systems for data integration and need to implement the BATCH and REAL TIME process for all the layers. Staging Area represents the assimilation of data from legacy Life/J and Life/A sources. Data quality analysis and enrichment occurs against this data prior to migrating it into the strategic enterprise information environment. The staging area is utilized for loading source data only. The goal is to open and close communication with the source system as efficiently as possible. The Information Hub helps facilitate the Data movement into the ODS/DW from the Stage database.
Responsibilities:
Involved in gathering and analyzing the requirements and preparing business rules.
Designed and developed complex mappings by using Lookup, Expression, Update, Sequence generator, Aggregator, Router, Stored Procedure, etc., transformations to implement complex logics while coding a mapping.
Worked with Informatica power center Designer, Workflow Manager, Workflow Monitor and Repository Manager.
Developed and maintained ETL mappings to extract the data from multiple source systems like Oracle, SQL server and Flat files and loaded into Oracle.
Developed Informatica Workflows and sessions associated with the mappings using Workflow Manager
Involved in creating new table structures and modifying existing tables to fit into Data Model
Extracted data from different databases like Oracle and external source systems like flat files using ETL tool
Involved in debugging Informatica mappings, testing of Stored Procedures and Functions, Performance and Unit testing of Informatica Sessions, Batches and Target Data.
Developed Mappings, Mapplets, Reusable Transformations, Source and Target definitions
Generated queries using SQL to check for consistency of the data in the tables and to update the tables as per the Business requirements.
Involved in Performance Tuning of mappings in Informatica.
Good understanding of source to target data mapping and Business rules associated with the ETL processes
Skills Used: Informatica 8.6, Informatica Versioning, Oracle (SQL, PL/SQL), UNIX, TOAD, Putty, Tortoise SVN, Team management
Verizon Data Services Pvt Ltd, Chennai, India
May 2007 – Feb 2008
Network Metric Protocol (NMP)
Position: Analyst
NMP was a complete solution for its Landline Telecom service. The project was to identify the operation process and to automate/performance tune, which would gain time/cost/manual efforts. My role was to lead the team to automate the manual process and performance saving action which would reduce the operation cost and steps.
Responsibilities:
Involved in analyzing and development of the Data Warehouse.
Worked on Informatica power center tool - Source Analyzer, Warehouse Designer, Mapping Designer, Mapplets and Transformation Developer
Created various mappings using Aggregate, Filter, Join, Expression, Lookup, Update Strategy and Router.
Extensively used ETL to load data from different databases and flat files to Oracle.
Involved in the development of Informatica mappings and also tuned them for better performance.
Created and scheduled Sessions and Batches through the Informatica Server Manager.
Worked with sessions and batches using Server Manager to load data into the target database.
Testing for Data Integrity and Consistency.
Skills Used: Informatica 7.2, Oracle (SQL, PL/SQL), Unix, Business Objects, Team management
Verizon Data Services Pvt Ltd, Chennai, India
Dec 2006 – Apr 2007
DBViews
Position: Analyst
As a part of the security features, NMP’s sensitive data, like clients master information, were masked from the view of offshore users for preventive action of malpractice and to ensure only the onsite users to have all the privilege on the data.
Responsibilities:
Involved in development, testing, and operations as well as performance tuning
Reduced the manual operational steps by identifying the areas that could be automated and then did automation hence cutting the manual steps by 50%
Designed and created different MicroStrategy objects such as advanced Metrics, Filters, and Prompts
Assisted with the creation of schema objects, facts, attributes and set relationship
Implemented the change control requests (CCR) on a need basis from client side
Delivered the reports to the business owners on weekly/ monthly basis
Skills Used: Informatica 7.2, Oracle, Unix, Business Objects, Team management
Verizon Data Services Pvt Ltd, Chennai, India
Apr 2006 – Nov 2006
NMP Upgrade to 10g
Position: Analyst
As a part of Technology Competitive and entering Advance features of Oracle 10g, Verizon decided to move its NMP Project which was developed in Oracle 8i version into Oracle 10g. My role was to ensure that the Oracle 8i code behaved the same in Oracle 10g.
Responsibilities:
Involved in development, testing, and operations as well as performance tuning
Reduced the manual operational steps by identifying the areas that could be automated and then did automation hence cutting the manual steps by 50%
Assisted with the creation of schema objects, facts, attributes and set relationship
Prepared Software Engineering Quality Process related documents
Implemented the change control requests (CCR) on a need basis from client side
Delivered the reports to the business owners on weekly/ monthly basis
Skills Used: Informatica, Oracle, Unix