Post Job Free

Resume

Sign in

Data Etl

Location:
Broomfield, CO
Posted:
January 04, 2021

Contact this candidate

Resume:

SHASHANK

+1-720-***-****

adi523@r.postjobfree.com

SUMMARY

Around 5 Years of IT experience in all aspects of Analysis, Design, Testing, Development, Implementation, and Support of Relational Database, Data Warehousing Systems, and Data Marts in various domains.

Expertise in Data Warehouse/Data mart, ODS, OLTP, and OLAP implementations teamed with project scope, Analysis, requirements gathering, data modeling, Effort Estimation, ETL Design, development, System testing, Implementation, and production support.

Around 5 years of experience with Talend Open Studio & Talend Enterprise platform.

Experienced in working with Talend for Big Data, Talend for Data Integration,

Experience in using cloud components and connectors to make API calls for accessing data from cloud storage Amazon S3 in Talend Studio.

Implementing the Talend jobs to Migrate data from Salesforce Classic to Salesforce lightning.

Implementing jobs to load the data to Salesforce from different source systems.

Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType,, tflowtoIterate, tAggregate, tSortRow, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tWarn, tLogCatcher, tMysqlScd, tFilter, tGlobalmap, tDie etc.

Expertise in Data modeling techniques like Data Modeling- Dimensional/ Star Schema and Snowflake modeling, Slowly Changing Dimensions (SCD Type 2).

Experienced in working with the Snowflake database.

Experienced in Waterfall, Agile methodologies.

Experience in Big Data technologies like Hadoop/Map Reduce Hive, Dynamodb, Spark SQL.

Experience of Hadoop Architecture and various components such as HDFS, Name Node, Data Node, Job Tracker, Task Tracker, YARN, and Map Reduce.

Experienced with Talend Big Data, Hadoop, Hive and used Talend Big data components like tHDFSInput, tHDFSOutput,, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput. Experience in using spark SQL

Experience in the development and design of ETL (Extract, Transform, and Loading data) methodology for supporting data transformations and processing, in a corporate-wide ETL Solution using Informatica Power Center.

Created mappings using Lookup, Aggregator, Joiner, Expression, Filter, Router, Update Strategy, and Normalizer Transformations. Developed reusable Transformation and Mapplets.

Strong Experience with shell scripting, understanding of approaches for business intelligence, data warehouse.

Hands-on experience across all stages of the Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, system integration, and user acceptance testing.

Skills:

Big Data

HDFS, Hadoop MapReduce, KAFKA, Zookeeper, Hive

Databases

Teradata & utilities (FASTLOAD, MULTI LOAD), Oracle12c/11x/10g, Microsoft SQL Server, Hive, Aws s3, snowflake

Programming

T-SQL, PL/SQL

Scripting

UNIX shell scripting

Languages

SQL, Python.

Other Tools

SQL Navigator, Putty, MS-Office, AWS Workspace

Environment

Windows, UNIX & Linux

Professional Experience

BALL CORP, CO Feb 2020 to Present

Talend Developer

Responsibilities:

Worked on Snowflake cloud as a target, SAP, and JDE a database as source with talend as Data integration tool for ETL.

Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.

Developed jobs using various components in Talend.

Designed ETL Jobs/Packages using Talend Integration Suite (TIS).

Created complex mappings in Talend using Talend components like tHash, tDenormalize, tMap, tUniqueRow.

Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats into a Database table to record job history.

Created Talend Mappings to populate the data into dimensions and fact tables.

Frequently used Talend Administrative Console (TAC)

Developed complex Talend ETL jobs to migrate the data from flat files to the database.

Implemented custom error handling in Talend jobs and worked on different methods of logging.

Created ETL/Talend jobs both design and code to process data to target databases.

Created Talend jobs to load data into various Oracle tables. Utilized Oracle stored procedures and wrote a few Java code to capture global map variables and use them in the job.

Successfully Loaded Data into different targets from various source systems like Oracle Database, DB2, Flat files, XML files, etc. into the Staging table and then to the target database.

Prepared ETL mapping Documents for every mapping and Data Migration document for a smooth transfer of project from development to testing environment and then to the production environment.

Performed Unit testing and System testing to validate data loads in the target.

Environment: Talend 7.1, UNIX, Shell script, SQL Server, Oracle, Business Objects, SQL Scripts, Nexus, snowflake console, autosys.

FPL/NextEra energy, FL Aug 2019 – Jan2020

Lead Talend Developer

Responsibilities:

Developed complex ETL jobs from various sources such as SQL Server, PostgreSQL, and other files and loaded into target databases using the Talend OS ETL tool.

Created Talend jobs using the dynamic schema feature.

Interact with the business community and gathered requirements based on changing needs. Incorporated identified factors into Talend jobs to build the Data Mart.

Performance tuning - Using the map-cache properties, Multi-threading, and Parallelize components for better performance in case of huge source data. Tuning the SQL source queries to restrict unwanted data in the ETL process.

Involved in Preparing Detailed design and technical documents from the functional specifications.

Prepared low-level design documentation for implementing new data elements to EDW.

Provided Knowledge Transfer to the end-users and created extensive documentation on the design, development, implementation, daily loads, and process flow of the mappings.

Have used AWS components (Amazon Web Services) - Downloading and uploading data files (with ETL) to the AWS system using S3 Talend components.

Managing messages on Kafka topics using Talend Jobs.

Used more components in Talend and Few to be mentioned: tMap, tFilterRow, tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogback, tKafkaInput, tKafkaOutput components, etc. in many of my Jobs Design

Worked on Joblets (reusable code) & Java routines in Talend.

Design, Develop and Test ETL processes to meet project requirements

Used to be On-call Support if the Project is deployed to further Phases.

Environment: Talend Open studio V (6.1.1, 6.2.1), UNIX, AWS-S3, Aurora, DMS, Redshift, Postgres SQL, WINDOWS XP

AIG Houston, TX March 2019 – Aug 2019

Talend Developer

Responsibilities:

Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirement.

Setup ETL Framework, best practices around Talend for data integration implementation.

Closely worked with Data Architects in designing tables and even involved in modifying Technical Specifications.

Involved in Extraction, Transformation, and Loading of data from multiple source systems to SalesForce.

Extensively used t system component to push large sets of data to SalesForce Lightning.

Utilized Talend components like tSalesforceInput, tSalesforceGetUpdate, tSalesforceGetUpdate, tSalesForceOutput, tSalesForceOutputBulk, tSalesForceOutputBulkExec..

Experienced in loading parquet format files to AWS using Big Data Batch Jobs.

Extensively used copy command in snowflake to copy data to Aws using customized file formats like CSV and PARQUET.

Involved in Integrating IAM Roles in Talend Components.

Experienced in creating standard jobs. Involved in the development of Big data batch Jobs.

Extensively used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tfileList etc, ts3put, ts3get,treplicate,tsortrow,tdenormalize,tnormalize,tredshiftrow

Utilized Big Data components like tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHiveOutput, tHiveRow, tHiveConnection .

Experienced in executing the jobs in parallel using tparallelize component.

Worked on POC for TDM (Talend Data Mapper) to handle complex XML files.

Used debugger and breakpoints to view transformations output and debug mappings.

Load and transform data into HDFS from a large set of structured data /Oracle/SQL server using Talend Big data studio.

Worked on Global variables, Context variables, and extensively used tcontextload in most of the jobs.

Experienced in creating reusable jobs for error handling.

Experienced in tuning the ETL Jobs for better performance.

Extensively worked with TAC (Talend Administrator Console) for scheduling jobs using the execution plan.

Worked on Data Catalog such as data sampling, categorizing, and creating auto profiling.

Automatic discovery of data sources using data catalog features.

Creating data relationships and certifications.

Environment: Talend Data Integration 6.1/5.5.1, Talend Enterprise Big Data Edition 5.5.1, Talend Administrator Console, Oracle 11g, Hive, HDFS, SQL Navigator, Toad, Putty, Winscp

Zions Bank Corporation, Utah Jan 2018 – Feb 2019

Talend Developer

Responsibilities:

Participated in all phases of the development lifecycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.

Created Talend jobs to copy the files from one server to another and utilized Talend FTP components

Created and managed Source to Target mapping documents for all Facts and Dimension tables

Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.

Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.

Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.

Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport

Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components, etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.

Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)

Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.

Created Implicit, local, and global Context variables in the job. Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.

Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.

Developed stored procedure to automate the testing process to ease QA efforts and also reduced the test timelines for data comparison on tables.

Automated SFTP process by exchanging SSH keys between UNIX servers. Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.

Involved in production n deployment activities, creation of the deployment guide for migration of the code to production, also prepared production run books.

Environment: Talend Data Integration 6.4, Talend Enterprise Big Data Edition 6.4 Talend Administrator Console, Oracle 11g, Hive, HDFS, Netezza, SQL Navigator, Toad, Control M, Putty, Winscp.

Pieces Technologies, Dallas May2016 - Dec 2017

Talend / ETL Developer

Responsibilities:

Participated in JAD sessions with business users and SME's for a better understanding of the reporting requirements.

Design and developed end-to-end ETL processes from various source systems to the Staging area, from staging to Data Marts.

Analyzing the source data to know the quality of data by using Talend Data Quality.

Broad design, development, and testing experience with Talend Integration Suite and knowledge in Performance Tuning of mappings.

Developed jobs in Talend Enterprise edition from stage to source, intermediate, conversion, and target.

Involved in writing SQL Queries and used Joins to access data from Oracle, and My SQL.

Used tStatsCatcher, tDie, tLogRow to create a generic job let to store processing stats.

Using Talend transformations like tConvert Type, tSort Row, tReplace, tAggregate Row, tUnite etc.

Developed Talend jobs to populate the claims data to data warehouse - star schema.

Loading and unit tested the mappings.

Experienced in using debug mode of talend to debug a job to fix errors. Created complex mappings using tHashOutput, tHashInput, tNormalize, tDenormalize, tMap, tUniqueRow, tPivot To Columns Delimited etc. Used tRunJob component to run child job from a parent job and to pass parameters from parent to child job.

Created Context Variables and Groups to run Talend jobs against different environments

Server using components like tFile Copy, TFileAcrchive, tFileDelete, tCreateTemporaryFile, tFTPDelete, tFTPCopy.

tFTPRename, tFTPut, tFTPGet etc. Experienced in Building a Talend job outside of a Talend studio as well as on the TAC server.

Experienced in writing expressions within tmap as per the business need.

Handled insert and update Strategy using tmap. Used ETL methodologies and best practices to create Talend ETL jobs.

Extracted data from flat files/ databases applied business logic to load them in the staging database as well as flat files.

Environment: Talend 5.5/5.0, Oracle 11g, Teradata SQL Assistant, Redshift HDFS, MS SQL Server 2012/2008, PL/SQL, Agile Methodology, TOAD, ERwin, AIX, Shell Scripts, AutoSys, SVN.



Contact this candidate