Post Job Free
Sign in

Data Etl

Location:
Aurora, IL
Posted:
March 27, 2020

Contact this candidate

Resume:

PRAVEEN YARRAGUNTA

Email: adcg4j@r.postjobfree.com

M: 989-***-****

PROFESSIONAL SUMMARY:

5+ years of experience in Analysis, Design, Development, Testing, Implementation, Enhancement and Support of ETL applications which includes strong experience in OLTP & OLAP environments as a Data Warehouse/Business Intelligence Consultant.

3+ years of experience in Talend Open Studio (6.x/5.x) for Data Integration and Big Data.

Experience in working with Data Warehousing Concepts like OLAP, OLTP, Star Schema, Snow Flake Schema, Logical Data Modeling, Physical Modeling and Dimension Data Modeling.

2+ years of experience in Expertise on Talend Data Integration suite and Bigdata Integration Suite for Design and development of ETL/Bigdata code and Mappings for Enterprise DWH ETL Talend Projects.

Widespread experience in using Talend features such as context variables, triggers, connectors for Database and flat files.

Hands on Involvement on many components which are there in the palette to design Jobs & used Context Variables to Parameterize Talend Jobs.

Experienced in ETL Talend Data Fabric components and used features of Context Variables, MySQL, Oracle, Hive Database components.

Tracking Daily Data load, Monthly Data extracts and send to client for their verification.

Strong experience in designing and developing Business Intelligence solutions in Data Warehousing using ETL Tools.

Excellent understanding and best practice of Data Warehousing Concepts involved in Full Development life cycle of Data Warehousing.

Experienced in analyzing, designing and developing ETL strategies and processes, writing ETL specifications.

Involved in extracting user's Data from various Data sources into Hadoop Distributed File Systems (HDFS)

Experience with Programming Model, Installation and Configuration of Hadoop, Hive, Sqoop and using Linux commands.

Experienced in using Talend Data Fabric tools (Talend DI, Talend DQ, Talend Data Preparation, ESB, TAC)

Experienced in working with different data sources like Flat files, Spreadsheet files, log files and Databases.

Knowledge in Data Flow Diagrams, Process Models, E-R diagrams with modeling tools like Erwin & RStudio.

Strong Understanding of Data Modeling (Relational, dimensional, star and snowflake schema) Data analysis implementation of Data Warehouse using Widows and Unix.

Extensive Experience in, Functions, Developing Stored Producers Views and Triggers, complex queries using SQL Server.

Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, systems integration and user acceptance testing.

Worked in all phases of BW/BI full life cycles including Analysis, Design, Development, Testing, Deployment, Post-Production Support/Maintenance, Documentation and End-User Training.

Highly Proficient in Agile, Test Driven, Iterative, Scrum and Waterfall software development life cycle.

Highly motivated with the ability to work effectively in teams as well as independently.

Excellent interpersonal and communication skills, and is experienced in working with senior level managers, business people and developers across multiple disciplines. Ability to grasp and apply new concepts quickly and effectively.

TECHNICAL SKILLS NAL EXPERIENCE

ETL Tools: Talend Studio Data Integration& Big Data 7.0.1/6.4/6.3/6.2.1/5.6.2/5.5.2/5.3.1,

Talend Administrator Console, Talend Management Console.

Operating Systems: UNIX, Linux, Windows 98/2000, 2003, Windows NT 4.0

Databases: Oracle 11g/10g/9i/8i, DB2 10.0.5, Teradata V 14.0, Oracle, SQL Server

2008/2005/2000, Hive 0.13,Hadoop.

Languages: XML, SQL, Unix/LINUX shell scripting

Database Tools: SQL developer, SQL Server Management Studio,

DB2, SQL*Loader

Other Tools: SQL Navigator, WinSCP, Putty, MS-Office& Excel, VMWare Workstation

Educational Details:

Master’s in Information Technology from Stratford University, VA - Oct2017

PROFESSIONAL EXPERIENCES

Talend Developer

SIRVA Worldwide Moving & Relocation Services, Chicago, IL October 2018 to Present

Responsibilities:

Collaborating with Data Integration Team to perform data and application integration with a goal of moving more data more effectively, efficiently and with high performance to assist in business- critical projects coming up with huge data extraction.

Perform technical analysis, ETL design, development, testing, and deployment of IT solutions as needed by business or IT.

Participate in designing the overall logical & physical Data warehouse data model and data architectures to support business requirements

Explore prebuilt ETL metadata, mappings, Develop and maintain SQL code as needed for SQL Server database

Create reusable Jobletes and routines in Talend.

Worked on web services using talend components like tSOAP, tREST, tWebService, tWebService Input etc.

Troubleshoot data integration issues and bugs, analyze reasons for failure, implement optimal solutions, and revise procedures and documentation as needed.

Worked on Tool Migration projects to migrate code from SSIS to Talend.

Platform for profiling and comparison of data, which will be used to make decisions regarding how to measure business rules and quality of the data.

Responsible to tune ETL mappings, Workflows and underlying data model to optimize load and query Performance.

Implemented Performance tuning in Mappings and Sessions by identifying the bottlenecks and Implemented effective transformation logic.

Developing ETL framework for Data Masking, Audit, Balance, Control, Validation architecture etc.

Worked on end-to-end development of software products from requirement analysis to system study, designing, coding, testing (Unit & Performance), documentation and implementation.

Working on Talend Management Console to assign users, Scheduling jobs, Server Monitoring, Task Creation, Plan Creation, Job Deployment etc.

Participate and contribute to code reviews, shared modules, reusable

Use of Visual Studio Team Services (VSTS) for Source Code Control, project related document sharing and team collaboration.

Environment: Talend Cloud Bigdata Integration 7.0.1, Talend Management Console, SQL Integration Services (SSIS), MS SQL Server 2012/2008, Microsoft Azure.

Talend Developer

Seacoast National bank, Stuart, Florida June 2018 to October 2018

Responsibilities:

Worked in the Data Integration Team to perform data and application integration with a goal of moving high volume data more effectively, efficiently and with high performance to assist in business-critical projects.

Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.

Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.

Created and managed Source to Target mapping documents for all Facts and Dimension tables.

Analyzing the source data to know the quality of data by using Talend Data Quality.

Involved in writing SQL Queries and used Joins to access Data from Oracle, and MySQL.

Assisted in migrating the existing data center into the AWS environment.

Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.

Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.

Implemented Change Data Capture (CDC)and SCD using Talend mapping parameter.

Used Informatica's features to implement Type 1, 2, 3changes in slowly changing dimension Change Data Capture (CDC).

Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)

Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.

Experienced in using debug mode of Talend to debug a job to fix errors.

Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.

Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.

Developed Talend jobs to populate the claims data to data warehouse - star schema.

Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.

Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tfilelist, tDie etc.

Worked Extensively on Talend Admin Console and Schedule Jobs in sky bolt.

Environment: Talend Enterprise Edition 6.4.1, Talend Administrator Console, Hive, My Sql, HDFS, Sqoop

Talend BigData Developer

T Rowe Price, Owings Mills, Maryland Feb 2016 to May 2018

Responsibilities:

Worked in the Data Integration Team to perform data and application integration with a goal of moving high volume data more effectively, efficiently and with high performance to assist in business-critical projects.

Has developed custom components and multi-threaded configurations with a flat file by writing JAVA code in Talend.

Interacted with Solution Architects and Business Analysts to gather requirements and update Solution Architect Document Created mappings and sessions to implement technical enhancements.

Deployed and scheduled Talend jobs in Administration console and monitoring the execution

Created separate branches with in the Talend repository for Development, Production and Deployment.

Excellent knowledge with Talend Administration console, Talend installation, using Context and global map variables in Talend.

Review requirements to help build valid and appropriate DQ rules and implement DQ Rules using Talend DI jobs.

Create cross-platform Talend DI jobs to read data from multiple sources like Hive, Hana, Teradata, DB2, Oracle, ActiveMQ.

Work with other CDC Supervisors to create and implement new processes to improve productivity and accuracy.

Create Talend Jobs for data comparison between tables across different databases, identify and report discrepancies to the respective teams.

Talend Administrative tasks like - Upgrades, create and manage user profiles and projects, manage access, monitoring, setup TAC notification.

Observed statistics of Talend jobs in AMC to improve the performance and in what scenarios errors are causing.

Created Generic and Repository schemas.

Performed Data Manipulations using various Talend Components like tMap. tjavarow, tjava, tOracleRow, tOracleInput, tOracleOutput, tMSSQLInput and many more.

Implementing complex business rules by creating re-usable transformations and robust mappings using Talend transformations like tConvertType, tSortRow, tReplace, tAggregateRow, tUnite etc.

Created standard and best practices for Talend ETL components and jobs.

Extraction, transformation and loading of data from various file formats like .csv, .xls, .txt and various delimited formats using Talend Open Studio.

Worked on HIVE QL to get the data from hive database.

Responsible for developing data pipeline with Amazon AWS to extract the data from weblogs and store in HDFS

Executed Hive queries on Parquet tables stored in Hive to perform data analysis to meet the business requirements.

Troubleshoot data integration issues and bugs, analyze reasons for failure, implement optimal solutions, and revise procedures and documentation as needed.

Responsible to tune ETL mappings, Workflows and underlying data model to optimize load and query performance.

Configure Talend Administration Center (TAC) for scheduling and deployment.

Create and schedule Execution Plans - to create Job Flows

Worked with production support in finalizing scheduling of workflows and database scripts using AutoSys.

Environment: Talend 6.2.1/6.0.1, Talend Open Studio Big Data/DQ/DI, Teradata V 14.0, Hive, HANA, PL/SQL, DB2, XML, JAVA. Erwin 7, UNIX Shell Scripting.

ETL Developer

Agility E Services, Hyderabad February 2014 to November 2015

Responsibilities:

Assisted gathering business requirements and worked closely with various Application and Business teams to develop Data Model, ETL procedures to design Data Warehouse.

Designed and developed star schema model for target database using ERWIN Data modeling.

Extensively used ETL Informatica tool to extract data stored in MS SQL 2000, Excel, and Flat files and finally loaded into a single Data Warehouse.

Used various active and passive transformations such as Aggregator, Expression, Sorter, Router, Joiner, connected/unconnected Lookup, and Update Strategy transformations for data control, cleansing, and data movement.

Designed and developed Mapplets for faster development, standardization and reusability purposes.

Implemented Slowly Changing Dimension Type 1 and Type 2 for inserting and updating Target tables for maintaining the history.

Used Debugger to validate transformations by creating break points to analyze and monitor Data flow.

Tuned performance of Informatica Session by increasing block size, data cache size, sequence buffer length and Target based commit interval, and mappings by dropping and recreation of indexes.

Worked along with the QA Team and provided production support by monitoring the processes running daily.

Involved in pre-and post-session migration planning for optimizing data load performance.

Interfaced with the Portfolio Management and Global Asset Management Groups to define reporting requirements and project plan for intranet applications for Fixed Income and Equities.

Performed Unit testing during the mapping phase to ensure proper and efficient implementation of the transformations.

Wrote UNIX Shell Scripts and pmcmd command line utility to interact with Informatica Server from command mode.

Environment: Informatica Power Center 8.x, Informatica Repository Manager, Oracle10g/9i, DB2, Erwin, TOAD, UNIX - AIX, PL/SQL, SQL Developer.



Contact this candidate