Post Job Free
Sign in

Data Developer

Location:
Hartford, CT, 06101
Posted:
March 30, 2018

Contact this candidate

Resume:

Anil Reddy

*************@*****.***

860-***-****

PROFESSIONAL SUMMARY:

Over all7+ years of experience in handling various aspects of Business Intelligence Viz. Data Warehouse Design, Development,ETL Integration, ETL Administration and Production Support.

Expertise in development of Data warehousing solutions using Informatica PowerCenter / Data Quality/ Cloud/ Big Data Edition (BDE), Talend Data Integration,Oracle PL/SQL.

Expertise in Data Modeling and in building Data Warehouse, Data Marts using Star Schema and Snow Flake Schema. Knowledge in Data Modeling (Logical, Physical, Star Schemas) with Erwin.

Worked extensively in building Dimensions, Bridges, Facts, and Star Schemas, Snow Flake (Extended Star) Schemas and Galaxy Schemas.

Very strong in Data Warehousing Concepts like Dimensions Type I, II and III, Facts, Surrogate keys, ODS, Staging area.

Experienced in creating PL/SQL Procedures using Netezza / Oracle and writing SQL's using Analytical functions in Oracle, Netezza.

Expertise in using Teradata SQL Assistant and data load/export utilities like BTEQ, Fastload, MultiLoad, Fast export and TPump.

Have used BTEQ, FEXP, FLOAD, MLOAD Teradata utilities to export and load data to/from Flat files.

Experience in writing Salesforce SOQL queries.

Proficient in the Integration of various data sources with multiple relational databases like Oracle11g /Oracle10g/9i, MS SQL Server, DB2, Teradata, Flat Files into the staging area, ODS, Data Warehouse and Data Mart.

Strong Knowledge in Relational Database Concepts, Entity Relation Diagrams, Normalization and De Normalization Concepts.

Built Data Integration Solutions using Extraction-Transformation-Load (ETL) / Extraction-Load-Transformation (ELT) models and tools Informatica Power Center 8.6.1 / 9.1.0 / 9.5.1/ 10.1.0., Talend Open Studio, Oracle PL/SQL, Teradata SQL and Netezza SQL.

Experience in working with structured data viz. Oracle / Teradata / Netezza RDBMS / Salesforce objects, semi structured data Viz. XML files and unstructured data viz. Text files, flat files.

Experience working with various connectors using Informatica PowerExchange / Talend Viz. Salesforce and Netezza Bulk reader and writer.

Strong experience in designing and developing complex mappings from varied transformation logic like Unconnected and Connected lookups, Normalizer, Source Qualifier, Router, Filter, Expression, Aggregator, Joiner, Sequence Generator, Sorter, mapplets and Update Strategy.

Good working knowledge of various Informatica Workflow manager components like Sessions, Worklets, Command Tasks, Email Tasks, File Watcher, Timer, Assignment Task, Decision Task, Connection Objects viz. SFTP, Salesforce, Relational etc.

Good working knowledge of Informatica Advanced Concepts like Salesforce Merge transformation, Concurrent workflow Execution, Variable Ports etc.

Experienced in designing / builfng reusable ETL common loading frameworks using Informatica mapping/ workflow Variables and Parameters.

Extracted data from multiple operational sources for loading staging area, Data warehouse, Data Marts using SCDs (Type 1/Type 2/ Type 3) loads.

Experience working with Informatica Cloud and creating Jobs to update, Insert, delete and dump data from Salesforce.

Experience integrating with the Salesforce.com using Informatica Cloud Connectors.

Built Data Quality / Data Profiling solutions using Informatica Data Quality 9.1.0,9.5.1 & 10.1.0 / Talend Open Studio / PL/SQL and experience in Data Cleansing, Data Analysis, Data Standardization, Data Consolidation, Data Matching and Rejection Automation.

Built Data Quality Profiles and Score cards using Informatica developer and Informatica Analyst for validating the source system files.

Experience converting profile findings into Data Quality rules and exporting them into PowerCenter as mapplets to be automated into the ETL solution.

Excellent knowledge on Informatica Data Quality Transformations such as: Labeler, Parser, Address Validator, Standardizer, etc., along with general transformations like Lookup, Expression, Filter, Router, Normalizer Etc.

Experienced in designing / building reusable ETL solutions for doing data quality checks across projects using Audit tables.

Good knowledge of Hadoop Ecosystem and various components such as HDFS, Map Reduce, Hive, HBase, Sqoop, Zookeeper.

Extensively worked on loading the Data in to HDFS by using Informatica Developer client (Big Data Edition).

Create mappings and workflows to load the data from different sources like Oracle, Flat File, HDFC, Netezza and SFDC (Salesforce) to Hive target in Hadoop through Informatica Developer (BDE) & Sqoop.

Experience in managing and reviewing Hadoop log files.

Good knowledge on Oozie scheduler to automate data loads into Hadoop Distributed File Systems to pre-process data.

Experience in importing and exporting data using Sqoop and writing custom shell scripts from HDFS to Relational Database Systems and vice-versa.

Involved in import /export of existing mappings into different environments including Development, QA and Production.

Intensively worked for client ETL code migration from Informatica to Talend Studio.

Worked extensively on Error Handling, Performance Analysis and Performance Tuning of Informatica ETL Components, Teradata Utilities, UNIX Scripts, SQL Scripts etc.

Proven track record in troubleshooting Informatica Sessions and addressing production issues like performance tuning and enhancement.

Good experience in working with UNIX Shell Scripts for automatically running sessions, aborting sessions and creating parameter files using control-m scheduler.

Worked on UNIX shell scripting for file processing to third party vendor through SFTP, encryption and decryption process.

Worked extensively on NZLOAD, NZSQL scripts to read and write data with Netezza Database.

Designed and created data migration system by use nzsql/UNIX shell script to migrate data from Oracle to Netezza database in batch.

Experience in monitoring and scheduling using Control M& Job Conductor (Talend Admin Console) and using UNIX (Korn& Bourn Shell) Scripting.

Good experience in writing number of Shell scripts to run various batch jobs.

Experience on post-session and pre-session Shell Scripts for tasks like merging flat files after Creating, deleting temporary files, changing the file name to reflect the file generated date etc.

Working experience with the Tableau &Microstrategy to create customized Scorecard reports & enhancement of existing reports to meet user requirements.

Detail oriented with good problem solving, organizational, analysis and requirement gathering skills and Proficient in Database analysis, Performance tuning, Requirement analysis, Data Migration, ETL development, Forms/Reports Development, Repository (RPD) and Reports development, Dashboard Creation and Unit Testing.

Performed unit testing and Integration test of the ETL code with test cases.

Involved in Unit testing, Iterative testing to check whether the data loads into target are accurate, which was extracted from different source systems according to the user requirements.

Performed Unit and Integration Testing and validates results with Business Analyst and end users.

Created detailed Unit Test Document with all possible Test cases/Scripts.

Excellent knowledge on Business Analysis, Data Analysis, requirement gathering and creating Business Requirement Documents and Technical Specification Documents

Experience in all aspects of Software development life cycle (SDLC) viz. design, analysis, development, implementation, testing and support and project development methodologies Agile Scrum & water fall.

Experience in Establishing best practice and standard for ETL process documents.

Experience in picking new technologies very fast and excelling in a dynamic environment.

Articulate with excellent communication and interpersonal skills with the ability to work in a team as well as individually.

TECHNICAL SKILLS:

ETL Tools: Informatica Power Center, IDQ, Informatica Developer (BDE), Talend Open Studio, Informatica Cloud.

Databases:Oracle 11g/10g/9i, MS SQL Server, Netezza, Teradata.

Environments:Unix, Windows, Linux

Tools:DB Visualizer9.1, Toad 9.1, SQL developer, ERWIN, MS Visio, BMC Control-M 9.0

Languages: UNIX shell Script, SQL, PL/SQL, VB.Net, HTML,CSS, Python.

Others: SalesForce.Com(SFDC), Hadoop Ecosystem (HDFS, Map Reduce, Hive, HBase, Pig, Sqoop, Zookeeper), Tableau

Operating Systems: Windows 95/98/NT/XP/2000/2003, Win 7 Sun Solaris, Unix/Linux

PROFESSIONAL EXPERIENCE:

Evariant, Farmington, CT Feb ‘2016 –Till Date

Role: Sr.ETL Developer

Description: This project aimed at building a ETL code for loading PBFS Customer Relationship Management (CRM) / Call Center (CC) and Physician Directory data into the Evariant Data warehouse and providing information to customers and the third-party vendors for benefit analysis. Project also involved enhancements to the ETL code for the existing Evariant sources.

Responsibilities:

Involved in analysis of source systems, business requirements and identification of business rule and responsible for developing, support and maintenance for the ETL process using Informatica.

Extracted data from heterogeneous sources like oracle, xml, flat file and perform the data validation and cleansing in staging area then loaded in to data warehouse in oracle 11g.

Created Informatica transformations/mapplets/mappings/tasks/worklets/workflows using PowerCenter to load the data from Source to Oracle Stage, Oracle Stage to Salesforce, Oracle Stage to Netezza Stage, Netezza Stage to Netezza Warehouse.

Made use of various Informatica source/target definitions viz. Flat files, Relational sources Viz. Oracle and Netezza, Salesforce Objects.

Utilized Informatica PowerExchange for Salesforce to load data into Salesforce Objects.

Utilized Informatica Cloud to load / Pull data from Salesforce.

Develop Data Replication/ Synchronization Tasks in Informatica Cloud (Cloud Application Integrations).

Created Filewatcher jobs to setup the dependency between Cloud and PowerCenter jobs.

Made use of various PowerCenter Designer transformations like Source Qualifier, Connected and Unconnected Lookups, Expression, Filter, Router, Sorter, Aggregator, Joiner, Rank, Router, Sequence generator, Union and Update Strategy transformations while creating mapplets/mappings.

Made use of reusable Informatica transformations, shared sources and targets.

Created different parameter files and changed Session parameters, mapping parameters, and variables at run time.

Implemented various loads like daily loads, weekly loads, and quarterly loads and on demand load using Incremental loading strategy and concepts of changes Data Capture (CDC).

Identified performance issues in existing sources, targets and mappings by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.

Created mappings for Type1, Type2 slowly changing dimensions (SCD) / complete refresh mappings.

Extensively used various Data Cleansing and Data Conversion functions like LTRIM, RTRIM, TO_DATE, Decode, and IIF functions in Expression Transformation.

Extensively used the Workflow Manager tasks like Session, Event-Wait, Timer, Command, Decision, Control and E-mail while creating worklets/workflows.

Deep understanding of core data quality design patterns and the associated challenges involved with data analysis, certification, modelling, quality improvement.

Integrated Informatica Data Quality IDQ with Informatica PowerCenter and Created various data quality mappings in Informatica Data Quality tool and imported them into Informatica powercenter as mappings mapplets.

Excellent knowledge on Informatica Data Quality Transformations such as: Labeler, Parser, Address Validator, Standardizer, Pair Generator Transformation, Match Transformation, Comparison transformation, Weight based analyzer, Cluster Transformation, Key generator transformation, Association Transformation, decision transformation, Consolidation transformation and built Match-Merge logic for Patient master data management.

Experience in Match & Merge setup and solid hands on experience for setting up fuzzy and exact match rules.

Utilized Salesforce merge transformation for Patient de-duping.

Good understanding of MDM architecture and Informatica MDM Hub console.

Create mappings and workflows to load the data from different sources like Oracle, Flat File, HDFC, Netezza and SFDC (Salesforce) to Hive target in Hadoop through Informatica Developer (BDE), Sqoop.

Created hive tables for the moved files in HDFS and Loaded the dataset into Hive for ETL Operations using Sqoop.

Ingested the Hadoop flat file to the Netezza database using HDFS connection.

Imported Hive table using PowerExchange and PowerExchange for Hadoop accesses Hadoop to extract data from HDFS or load data to HDFS/Hive.

Created Sqoop scripts to ingest data from HDFS to Netezza and from Oracle to HDFS and to PostgreSQL.

Worked with “pmcmd” command line program to communicate with the Informatica server, to start, stop and schedule workflows.

Created and edited custom objects and custom fields in Salesforce and checked the field level Securities.

Extracted various data from SalesForce.com using Informatica with Sales Force Adapter.

Modified SOQL (Sales force object Query Language) for Sales Force target at session level.

Worked with various salesforce.com objects like Accounts, Contacts, Leads, Opportunities, Reports, and Dashboards.

Created data visualization reports and dashboards as per the requirements using Tableau desktop.

Closely worked with the reporting team to ensure correct data is presented in the reports.

Developed various dashboards for clinical trial data visualization using Tableau Desktop.

Extensively used Data Blending to create Tableau workbooks from multiple data sources.

Created page layouts, search layouts to organize fields, custom links, related lists, and other components on a record detail and edit pages.

Extensively used UNIX Shell Scripts to automate the jobs.

Worked extensively with Netezza scripts to load the data from flatfiles to Netezza database.

Performed unit testing and Integration test of the ETL code with test cases.

Involved in Unit testing, Iterative testing to check whether the data loads into target are accurate, which was extracted from different source systems according to the user requirements.

Performed Unit and Integration Testing and validates results with Business Analyst and end users.

Effectively create standard review document and performed review of code to check if it maintains the coding standards and meets the business requirements.

Involved in the code review for optimization and the defect prevention.

Conducted code reviews on the code developed by my team mates before moving it into QA.

Involved in Performance tuning and peer review of code and using IDQ to test and compare the data in different tables.

Extensive experience in Relation Data modelling for creation logical and physical design for databases and proficient in creating E-R diagram using Erwin and Visio.

Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.

Designed tables required for the execution of the ETL processes using ERwin.

During the project, participated in multiple meetings with the client and data architect / ETL architect to propose better strategies for performance improvement and gather new requirements.

Environment: Informatica Power Center 10.1.0,Informatica Developer 10.1.0,Informatica Developer (BDE), Hive, Pig, HDFC, Oracle 11g, Flat Files, Putty, XML, UNIX,Netezza,SalesForce.Com(SFDC),Tableau.

Athene USA, West Des Moines, IA Jan ‘2015 – FEB’2016

Role: ETL Developer

Description:Athene USA provides products in the retirement savings market, including retail and fixed indexed annuity products as well as institutional products, such as funding agreements. The project consisted to retrieve some sensitive data, perform many queries manipulation, create reports and perform auditing services before inputting the data in the data warehouse for further analysis.

Responsibilities:

Involved in analysis of source systems, business requirements and identification of business rule and responsible for developing, support and maintenance for the ETL process using Informatica.

Created / updated ETL design documents for all the Informatica components changed.

Extracted data from heterogeneous sources like oracle, xml, Postgre, flat file and perform the data validation and cleansing in staging area then loaded in to data warehouse in oracle 11g.

Made use of various Informatica source definitions viz. Flat files and Relational sources.

Made use of various Informatica target definitions viz. relational data base targets.

Created Informatica transformations/mapplets/mappings/tasks/worklets/workflows using PowerCenter to load the data from source to stage, stage to persistent, stage to reject and stage to core.

Involved in migration projects to migrate data from data warehouses on Oracle/DB2 and migrated those to Teradata.

Fast Load jobs to load data from various data sources and legacy systems to Teradata Staging.

Worked on exporting data to flat files using Teradata Fast Export.

Did the performance tuning for Teradata SQL statements using TeradataExplain command.

Transformed bulk amount of data from various sources to Teradata database by using BTEQ scripts.

Made use of various PowerCenter Designer transformations like Source Qualifier, Connected and Unconnected Lookups, Expression, Filter, Router, Sorter, Aggregator, Joiner, Rank, Router, Sequence generator, Union and Update Strategy transformations while creating mapplets/mappings.

Made use of reusable Informatica transformations, shared sources and targets.

Created different parameter files and changed Session parameters, mapping parameters, and variables at run time.

Implemented various loads like daily loads, weekly loads, and quarterly loads and on demand load using Incremental loading strategy and concepts of changes Data Capture (CDC).

Identified performance issues in existing sources, targets and mappings by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.

Created mappings for Type1, Type2 slowly changing dimensions (SCD) / complete refresh mappings.

Extensively used various Data Cleansing and Data Conversion functions like LTRIM, RTRIM, TO_DATE, Decode, and IIF functions in Expression Transformation.

Extensively used the Workflow Manager tasks like Session, Event-Wait, Timer, Command, Decision, Control and E-mail while creating worklets/workflows.

Worked with “pmcmd” command line program to communicate with the Informatica server, to start, stop and schedule workflows.

Created Job Stream and added job definitions in Control-M and executed.

During the course of the project, participated in multiple meetings with the client and data architect / ETL architect to propose better strategies for performance improvement and gather new requirements.

Environment: Informatica Power Center 9.5, Oracle 11g,Db2, Teradata, XML, Flat Files, Win7, DbVisualizer, Control-M, Toad and Putty

Appstree IT Services, India July ’2010 - Nov ‘2013

Role: SQL Developer

Description: I was part of the team consisting of 4 members working on a project AVP Discounts (Online Retailer). We were supposed to create the mappings in Informatica and load data to data warehouse. I have worked on creating both TEST/PROD/DEV environments as per project requirements.

Responsibilities:

Developed ETL mappings, Transformations and Loading using Informatica Power Center 8.6.1.

Extensively used ETL to load data from Flat file, MS Excel, which involved both fixed width as well as Delimited files and from the relational database, which was Oracle 10g.

Developed and tested all the Informatica mappings, sessions and workflows - involving several Tasks.

Worked on Dimension as well as Fact tables, developed mappings and loaded data on to the relational database.

Worked extensively on different types of transformations like source qualifier, expression, filter, aggregator, update strategy, Lookup, sequence generator, joiner, Stored Procedure.

Analyzed the session, event and error logs for troubleshooting mappings and sessions.

Provided support for the applications after production deployment to take care of any post-deployment issues.

Environment: Informatica 8.6.1, UNIX shell scripting, Oracle 10g, SQL Programming, MS Excel, SQL *Plus.

EDUCATION:

Master of Science (Computer &Information Science)

Bachelors of Technology (InformationTechnology)



Contact this candidate