Post Job Free
Sign in

Data Etl

Location:
Atlanta, GA
Posted:
July 24, 2020

Contact this candidate

Resume:

Summary

Over **+ years of experience in analysis,design and development of client/server,web based and n-tier application along with extensive experience in Data warehousing, Data modeling, Data integration, Data Migration, ETL process and Business Intelligence as a ETL Developer using Pentaho KETTLE and Informatica. Excel under pressure in deadline driven environment with superior interpersonal communication and technical documentation skills.

Professional Expertise

Drawing on Experience in all aspects of analytics/data warehousing solutions (Database issues, Data modeling, Data mapping, ETL Development, metadata management, data migration and reporting solutions).

Strong understanding of Data Modeling (Relational, dimensional, Star and Snowflake Schema), Data analysis, implementations of Data warehousing using Windows and UNIX.

Developed mappings in Informatica to load the data from various sources into the Data warehouse different transformations like Source Qualifier, JAVA, Expression, Lookup, Aggregate, Update Strategy and Joiner.

Extensive experience in developing stored procedures, functions, Views and Triggers, Complex queries using SQL Server, TSQL and Oracle PL/SQL.

Developed ETL process using Pentaho PDI to extract the data from HIS and Vista, and populated it in to BI Data mart.

Experience in resolving on-going maintenance issues and bug fixes and monitoring Informatica sessions as well as performance tuning of mappings and sessions.

Efficient in implementing QA Process, established standards, procedures, and methodologies and also have a solid expertise in HP Quality Center.

Good understanding of database client/server configuration on Oracle 8i/9i/10g, SQL Server 2005/2008 and DB2.

Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType,, tflowtoIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tWarn, tLogCatcher, tMysqlScd, tFilter, tGlobalmap, tDie.

Good Understanding of Agile approaches like user stories, Programming, Test-Driven Development, Daily Standups, Retrospectives, Pair Programming and Scrum meetings.

Worked in close co-ordination with the testing team in developing test plan and test cases from functional and business requirements.

Prepared Test plans, Test cases reviews, test cases execution, designing, executing, verifying and analyzing test programs.

Experience in working on projects developed using JAVA, VB, ASP,.NET, Web services -SOAP, UDDI, WSDL, HTML, Oracle, MS SQL Server, VBScript and JavaScript.

Formulated and tested queries with SQL on an Object/Traditional Relational Database.

Exposure to all phases of Software Development Life Cycle (SDLC). Also worked in SDLC models like Waterfall, RUP and Agile.

Power Center 8.5/7.1.3/6.2.1 (Designer, Workflow Manager, Workflow Monitor), and Power Connect.

Extensive work in ETL process consisting of data transformation, data sourcing, mapping, conversion and loading.

Knowledge on Big Data using Spark and Scala.

Extensive experience in implementation of Data Cleanup procedures, transformations, Scripts, Stored Procedures and execution of test plans for loading the data successfully into the targets.

Emphasized on Performance of Mapping Optimizations in Informatica.

Experienced on working with Big Data and Hadoop File System (HDFS).

Experienced on working with NoSQL databases like Cassandra and MongoDB.

Expertise in developing SQL and PL/SQL codes through various stored procedures, functions, triggers to implement the business logics of database in Oracle and SQL Server 2000/2008.

Excellent analytical, communication and Interpersonal skills. Also have led small teams in implementation of projects. Good at mentoring newbies and have excellent leadership skills.

Good team player with the ability to manage and work independently in a time sensitive environment.

Technical Skills

Testing Tools

QTP 9.0/9.5/10.0/11.0, Quality Center 8.2/9.0/9.5/10.0, Load Runner 9.50, Team Foundation Server 2010, Test Manager 2010

Database Testing

Transaction Testing (front end to back end testing), Data Mapping / Data Conversion Testing and Data & File Validation

Programming Languages

C, C++, ASP.NET, VB.NET, ADO.NET, Java, Python

ETL Tools

Informatica Power Center 8.6.1/8.1/7.1/7.0/6.2/5.1

, SSIS, DTS, Kettle 4.2,5.2,6.07.0,7.1, Talend

Business Intelligence Reporting Tools

Oracle Business Intelligence Suite (OBIEE), SSRS

Operating system

Windows 95/98/NT/2000/XP, Unix, Linux, MAC

Databases (SQL & NoSQL)

Oracle 8i/9i/10g/11g, Teradata, PL/SQL, SQL Server 2000/ 2005/2008, MySQL, MS Access, MongoDB, Cassandra

Web Technologies

Java Script, VB Script, HTML

Assembly Languages

8085/8086 Microprocessors

Version Control Tools

VSS, TortoiseSVN, CVS, ClearCase

Utility Tools

MS Word, MS Excel, MS PowerPoint, Eclipse, MS

Visual Studio, Toad, SQL Developer

BigData Tools

Spark & Scala

Professional Experience

EFX, Atlanta, GA From Aug ’18 – Till date

Senior Pentaho Consultant

Equifax is a global information solutions company that uses unique data, innovative analytics, technology and industry expertise to power organizations and individuals around the world by transforming knowledge into insights that help make more informed business and personal decisions.

Headquartered in Atlanta, Ga., Equifax operates or has investments in 24 countries in North America, Central and South America, Europe and the Asia Pacific region. It is a member of Standard & Poor's (S&P) 500® Index, and its common stock is traded on the New York Stock Exchange (NYSE) under the symbol EFX. Equifax employs 10,400 employees worldwide.

Responsibilities:

Design, review, implement and optimize data transformation processes using Pentaho suite.

Extensively worked with Business Users in gathering requirements and actively cataloging and supporting various issues and providing their solution.

Used bunch of transformations in Pentaho transformations including Row Normalizer, Row Demoralizer, Database Lookup, Database Join, Calculator, Add Sequence, Add Constants and various types of inputs and outputs for various data sources including Tables, Access, Text File, Excel and CSV file.

Participated in design of Staging Databases and Data Warehouse/Data mart database using Star Schema/Snowflakes schema in data modeling.

Gathered business requirement by understanding business Processes and needs.

Installed and configured Pentaho Suite 8.1 and tested the transformations using the same.

Plan data integration process by defining specifications and data integration processes best fit for data ingestion into CDA system based on rigorous analysis of legacy applications and emergent systems.

Used Pentaho Import Export utility to Migrate Pentaho Transformations and Job from one environment to others.

Prepared ETL (Extract, Transform and Load)standards, Naming conventions and wrote ETL flow documentation for Stage, ODS and Mart.

Configured Pentaho Server for ETL deployment by creating database connections in Pentaho enterprise console for central usage by the ETL scripts deployed in the repository.

Used Pentaho Design Studio for creating custom parameters as well as generating report.

Automated file transfer processes and mail notifications by using FTP Task and Send Mail task in Transformations

Integrating Kettle (ETL) with Hadoop, Pig, Hive, Spark, Storm, HBase, Kafka and other Big Data component for various functionalities and other various NoSQL data stores can be found in the Pentaho Big Data Plugin.

Created and saved Pentaho jobs in enterprise repository and scheduled them to run in production on weekly basis.

Created and used various re-usable tasks, workflows, worklets, mapplets, and reusable transformations.

Experience with data integration using hive Scripts, processing the data over HDFS.

Responsible for creating database objects like table, views, Store Procedure, Triggers, Functions etc. using T-SQL to provide structure to store data and to maintain database efficiently.

Extensively used joins and sub queries to simplify complex queries involving multiple tables.

Optimized the performance of queries with modification in TSQL queries, removed unnecessary columns, eliminated redundant and inconsistent data, normalized table, established joins and created Clustered, Non-Clustered indexes whenever necessary.

Environment: Hortonworks 2.3.5, Sqoop, Hive, Oracle 12c, Pentaho Data Integration Spoon 8.1, PL/SQL, Pentaho Enterprise console, Hadoop, Linux, Windows

MidAmerican Energy Services, Atlanta, GA From Oct ’13 – Jul ’18

Senior ETL/Enterprise Developer & Team Lead

MidAmerican Energy Services has developed and operates a powerful cloud based set of managed service solutions that delivers its customers the economic benefit of Virtual Peaking Power Plants, Virtual Power Plants and Virtual Microgrids through Capacity Management, Peak Load Management, Energy Efficiency Optimization, Asset Management and Distributed Energy Resources Management.

Real Time Energy Management Systems – It is designed to match today’s real-time business environment to manage energy in real-time. It expands the traditional approaches through real-time energy measurement and business empowerment that transitions the control and optimization of energy cost to a real-time model

Responsibilities:

Design, review, implement and optimize data transformation processes in Big Data ecosystems.

Design, develop, document, and test ETL solutions using industry standard tools.

Plan data integration process by defining specifications and data integration processes best fit for data ingestion into IROM system based on rigorous analysis of legacy applications and emergent systems.

Participate in requirements gathering sessions with business and technical staff to distill technical requirements from business requests.

Partner with clients to fully understand business philosophy and IT Strategy and recommend process improvements to increase efficiency and reliability in ETL development.

Using the data Integration tool Pentaho KETTLE for designing ETL jobs in the process of building Data warehouses and Data Marts.

Present ETL documentation and designs to team members and convey complex information in a clear and concise manner.

Enhance ETL team accomplishments and competence by planning delivery of solutions and mentoring team members.

Good experience in using various PDI steps in cleansing and load the data as per the business needs.

Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.

Experienced in Oracle Applications Development, Customization, Testing, Debugging and Support of Oracle Applications.

Involved in ETL Development using native Oracle tools (SQL*LOADER, Oracle PL/SQL).

Good Experience in creating cubes by using Pentaho Schema Workbench.

Developed complex Talend ETL jobs to migrate the data from flat files to database.

Review and analyze data sources to define normalize and standardize them to facilitate automated data and content integration into IROM Customer Portal.

Extract data from multiple sources, integrate disparate data into a common data model, and integrate data into a target database, application, or file.

Possesses both Technical and Functional knowledge of Oracle E-Business Suite.

Validate data integration by developing, and executing test plans and scenarios including data design, tool design, data extract/transform, networks, and hardware.

Ensure timely implementation of data integration projects based on the determined milestones and quality standards.

Ensure transparency and knowledge sharing through effective usage of documentation and cross training.

Assist in the ongoing development of technical best practices for data movement, data quality, data cleansing and other ETL-related activities.

Environment: Pentaho ETL (KETTLE) tool, Informatica PowerCenter 8.6/8.1, Talend Open studio V (6.1.1, 6.2.1), Tableau, Java, Windows/Web Applications, BI Reporting, NoSQL: HBase, HDFS, MYSQL, Oracle, Cassandra, MongoDB, Oracle Applications/Oracle E-Business Suite R12/11i(11.x)/10.x, Agile-Scrum Methodologies, VB Script, Linux.

InterCall, Atlanta, GA From Jun ’12 - Sep ‘13

Senior ETL Consultant

InterCall, the world’s largest conferencing provider, offers more audio, web and video conferencing solutions for effective business communications. Our integrated web, audio and video conferencing makes web meetings, webinars, audio conferencing, online training seminars and multi-point video conferencing easier and more productive.

UMS – Usage Management System is the project which he works on. This project deals with the migration of conference processing system from IBM Informix to Oracle 11g.

Responsibilities:

Involved in the analysis of the user requirements and identifying the sources.

Created technical specification documents based on the requirements by using S2T Documents.

Involved in the preparation of High level design documents and Low level design documents.

Involved in Design, analysis, Implementation, Testing and support of ETL processes for Stage, ODS and Mart.

Prepared ETL standards, Naming conventions and wrote ETL flow documentation for Stage, ODS and Mart.

Administered the repository by creating folders and logins for the group members and assigning necessary privileges.

Collect and link metadata from diverse sources, including relational databases and flat files.

Designed and developed Informatica’s Mappings and Sessions based on business user requirements and business rules to load data from source flat files and oracle tables to target tables.

Developed ETL process using Pentaho PDI KETTLE to extract the data from legacy system.

Worked on various kinds of transformations like Expression, Aggregator, Stored Procedure, Java, Lookup, Filter, Joiner, Rank, Router and Update Strategy.

Worked on Joblets (reusable code) & Java routines in Talend.

Developed reusable Mapplets and Transformations.

Used debugger to debug mappings to gain troubleshooting information about data and error conditions.

Followed the organization defined Naming conventions for naming the Flat file structure, Talend Jobs and daily batches for executing the Talend Jobs.

Involved in monitoring the workflows and in optimizing the load times.

Used Change Data Capture (CDC) to simplify ETL in data warehouse applications.

Involved in writing procedures, functions in PL/SQL.

Involved in extensive performance tuning by determining bottlenecks at various points like targets, sources, mappings, sessions or system. This led to better session performance.

Worked with SQL*Loader tool to load the bulk data into Database.

Prepared test Scenarios and Test cases in HP Quality Center and involved in unit testing of mappings, system testing and user acceptance testing.

Defect Tracking and reports are done by Rational Clear Quest.

Environment: Informatica PowerCenter 8.6/8.1 (PowerCenter Designer, workflow manager, workflow monitor, PowerConnect), Pentaho ETL (KETTLE) tool, Oracle Data integrator, MS Visio, ERWIN,Talend 5.x, SQL*Loader, HP Quality Center, Oracle9i/10g, Teradata,PL/ SQL, Sybase, AUTOSYS, Rational Clear case, Rational Clear Quest, Windows XP, TOAD, UNIX.

TMX Finance LLC, Alpharetta, GA From Sep ’11 - Jun ‘12

Senior ETL/Automation Consultant

TMX Finance, LLC is one of the largest, fastest growing and most profitable consumer specialty finance companies in the United States. One of the US's largest Insurance Company, Title Max offers a variety of insurance products for individuals and businesses with its products ranging from automotive, immovable assets, mechanical coverage, personal lines, and commercial lines including reinsurance. It involves maintenance of its customer information, policy details, agent’s information, premium tariffs, vehicle information, history of their road accident, personal injury and claims settlement including self and third party liability. TMX Finance has various sites like Bonus Plan, Corporate Functions, Project Initiatives, Project Center, Virtual Office and many more which are managed in SharePoint portal.

Responsibilities:

Developed Logical and Physical data models that capture current state/future state data elements and data flows using Erwin 4.5.

Responsible for design and build data mart as per the requirements.

Extracted Data from various sources like Data Files, different customized tools like Meridian and Oracle.

Extensively worked on Views, Stored Procedures, Triggers and SQL queries and for loading the data (staging) to enhance and maintain the existing functionality.

Done analysis of Source, Requirements, existing OLTP system and identification of required dimensions and facts from the Database.

Created Data acquisition and Interface System Design Document.

Developed jobs, components and Joblets in Talend. Designed ETL Jobs/Packages using Talend Integration Suite (TIS).

Designed the Dimensional Model of the Data Warehouse Confirmation of source data layouts and needs.

Extensively used Oracle ETL process for address data cleansing.

Developed and tuned all the Affiliations received from data sources using Oracle and Informatica and tested with high volume of data.

Created complex mappings in Talend using tHash, tDenormalize, tMap, tUniqueRow, tPivotToColumnsDelimited as well as custom component such as tUnpivotRow.

Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Oracle and Informatica PowerCenter.

Created common reusable objects for the ETL team and overlook coding standards.

Reviewed high-level design specification, ETL coding and mapping standards.

Designed new database tables to meet business information needs. Designed Mapping document, which is a guideline to ETL Coding.

Used ETL to extract files for the external vendors and coordinated that effort.

Migrated mappings from Development to Testing and from Testing to Production.

Performed Unit Testing and tuned for better performance.

Created various Documents such as Source-to-Target Data mapping Document, and Unit Test Cases Document.

Environment: Informatica Power Center 8.1/7.1.2, Erwin 4.5, Oracle 10g/9i, Talend 5.x, XML, PL/SQL, SQL Server 2005/2000 (Enterprise Manager, Query Analyzer), Sybase, SQL* Loader, SQL * Plus, Autosys, OLAP, Windows XP/NT/2000, Sun Solaris UNIX, MS Office 2003, Visio Project, Shell scripts, Mainframe.

Orion Systems Integrators Inc., Alpharetta, GA From Sep ’10 - Sep ’11

QA/ETL Consultant

Client : TMX Finance, LLC

ATLAS is a web based internal application helping in the title mortgage insurance business. It is used to create account information of the customers and also maintaining the transaction details of the customers.It helps the user track the details of the loans given to the customer.This application can be accessed by corporate, division, region, district and store managers.Currently it is deployed in the states of Virginia, Texas and Georgia as a pilot and expanded to other states of business.

Responsibilities:

Agile development - Participated in daily scrum meetings to report daily progress of Automation activities and Sprint planning for test planning activities.

Involved in requirement gathering and performed ETL Process.

Created Dimension Tables and Fact Tables based on the warehouse design.

Wrote Triggers and Stored Procedures using PL/SQL for Incremental updates

Integrated various sources in to the Staging area in Data warehouse to Integrate and for Cleansing data.

Involved in building of the Data Warehouse which includes the Design of Data mart using Star Schema

Created the Source and Target Definitions in Informatica Power Center Designer

Created Reusable Transformations and Mapplets to use in Multiple Mappings

Developed and updated documentation of processes and system.

Environment: Power Center 6.2/7.1, Power Connect, Business Objects 6.5/6.0, Erwin 4.0, Oracle 9i, Teradata V2R5, Autosys, Toad, SQL, PL/SQL, Sun Solaris 8.0 and Windows 2000, UNIX shell Scripts, Mainframe.

Transworks Information Services Ltd, India Jan ’05 – Feb ’08

Software Engineer

TransWorks services are designed to provide end-to-end "outsourcing solutions" to our clients. Our service delivery platforms are organized along 2 main areas - CRM and BPO. A dedicated operations team supports each delivery platform. Through a mixture of scale, technology, and best practices, TransWorks helps clients define how to improve efficiency and lower costs for their various functions by offshore outsourcing.

Responsibilities:

Working as a team lead managing 6 resources from offshore and handling 4 modules Order Entry, Shipping and Sales and their reports.

Analyzed usecase & design specification documents and created Test Plan, Test Strategy documents

Understanding the Landauer business, Usecases and transer the Knowledge to team using visio diagrams.

Identifying the areas of automation and creating QTP scripts to create data for down stream modules and reporting

Creating and maintaining QTP scripts using modular framework with QC.

Create QTP scripts for integration scenarios and to create data for reporting.

Creating library functions using VBscript in QTP.

Creating XML’s as per the schema file using VB script and hitting Web Service with large volume of data (Volume Testing)

Creating Load Runner scripts to test customer interface, WSS – Web Self Service application

Providing estimates for CR’s and the work for modules being handled.

Creating and generate daily reports in QC to manage the work for team (Summary reports for Test case creation and Execution)

Participanting in daily defect review meetings to discuss and assign defects

Reviewing testing artifacts and getting approvals from client/SME’s.

Resolving functional issues from offshore team by discussing with SME’s at Landauer and Development team in Canada(Agresso).

Involving in offshore meetings, Build release meetings and QA status meetings.

Written and modified required simple SQL validation scripts to validate the outputs.

Prepare Status Reports in QC every week with QA status to help in monitoring the amount of work done for current week and plan for next week.

Resource work plan to determine the completion date by using Microsoft project plan (MPP).

Involving in Project management meetings (PMO) and escalate issues with reasons for showstoppers.

Environment: .Net 2003 and SQL Server2000.

Education

M.S. in Computer Technology, Eastern Illinois University, Charleston, Illinois

B.Tech in Electronics and Communication Engineering, Anna University, Chennai, India



Contact this candidate