Post Job Free

Resume

Sign in

Data Etl

Location:
Tampa, FL
Salary:
55/hr
Posted:
July 03, 2020

Contact this candidate

Resume:

Swethashri Manoharan

adea96@r.postjobfree.com

636-***-****

Professional Summary

Proven knowledge building data warehouses using Teradata, Oracle 11g/10g, MS SQL Server, MySQL.

Strong experience in Extraction, Transformation, Loading ( ETL) data from both structured and unstructured sources into Data Warehouses and Data Marts using Informatica PowerCenter.

Good experience working on various Data Warehousing/DataMarts, Data Integration, Data Migration, Data Consolidation, Data Conversion, Cube/Analytics projects.

Knowledge in OLTP/OLAP System study, developing Database schemas like star schema, snow flake schema Dimensional Data Modeling used in relational, dimensional modeling and slowly changing dimensions (SCD).

Experience with ODS (Operational Data Store) to integrate data from different heterogeneous data sources in order to facilitate operational reporting in real-time or near real-time.

Hands on experience in Performance Tuning of sources, targets, transformations and sessions in Informatica.

Experience in debugging Informatica PowerCenter mappings. Identified bugs in existing mappings by analyzing the data flow and evaluating transformations using debugger.

Experience in UNIX/LINUX working environment, writing UNIX/LINUX shell scripts and bash scripting for windows environment.

Experience in Integration of various data sources like Teradata, Oracle, SQL Server, XML files, Flat Files into Staging Area and CDM.

Experience in design and development of ETL methodology for supporting Data Migration, data transformations & processing in a corporate wide ETL Solution using Teradata TD 16.x & earlier.

Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.

Involvement in all phases of SDLC (Systems Development Life Cycle) from analysis and planning to development and deployment.

Implemented and followed a Scrum Agile development methodology within the cross functional team and acted as a liaison between the business user group and the technical team.

Experience in scheduling Sequence and parallel jobs using UNIX scripts and scheduling tools.

Expert in ETL tools using Informatica power center 10.x/9.x/8.x/7.x and Informatica power exchange 8.x/7.x components for Data Migration & Integrating Data from various sources Oracle, DB2, Sybase, SQL Server, Salesforce, EMC Green Plum and Semi-Structured sources like XML.

Experience in working with indexes, complex queries, stored procedures, views, triggers, user defined functions, complex joins, loops T-SQL, DTS/SSIS using MS SQL Server.

Experience working with dimensional Data Modeling using ERwin, Physical & Logical data modeling, Ralph Kimball Approach, Star Modeling, Data marts, OLAP, FACT & Dimensions tables

Strong working knowledge in RDBMS, ER Diagrams, Normalization and De Normalization Concepts.

Experience working with configuring SSIS packages using package logging, breakpoints, Checkpoints and Event handler to fix the errors.

Technical Skills

Methodologies

Agile (Scrum) and Waterfall

Software Testing

Functional, Desktop Applications and Web based UI Testing, Windows applications, SQA Testing & Methodologies, Test Plans, Cases & Processes, Functional Requirements, Scripting & Documentation, Regression & Negative Testing, UI & Compatibility Testing, Testing Automation, Defect/Bug Tracking, Test Strategies & Coverages, QA & QC Standards

Technical Leadership

Estimations, Tracking and Status Reporting, Risk tracking and Management, Team Coordination and working with various Cross Functional Teams, Impact Analysis, Change Management, Agile, Waterfall and Prototype Models

Testing Tools

QTP/UFT, HP ALM, JIRA, Quality Center

Programming Languages

Visual Basic

Databases

MYSQL, Oracle, SQL Server

Professional Experience

ETL Informatica Developer Sept 2018 – Present

Toyota - Torrance, CA

Creating software applications using Informatica, Teradata, hive, spark, Hadoop and different technologies to load the enterprise data into the database models.

Develop, Test, maintain and execute code units for data Ingestion from Oracle, SQL Server, Netezza, Salesforce to Apache Hadoop/Teradata systems using DIFA, Python, Hive scripts and Sqoop jobs.

Build Automation scripts to reduce manual regressive efforts and optimize data warehouse performance.

Develop Teradata SQL and hive scripts to complex problems that require a high degree of ingenuity, creativity and innovation within the boundaries of industry compliance regulations and maintain the coding standards.

Implementation and Maintenance of Security Module for sensitive data.

Created ETL transformations and jobs using Pentaho Data Integration Designer (Kettle-Spoon) and scheduling them using Cron Job

Used JavaScript, Regex and Java to build custom filters and steps which would satisfy the business requirements of the end user

Identify and analyze data discrepancies and data quality issues and works to ensure data consistency and integrity

Created ETL transformations and jobs using Pentaho Data Integration Designer (Kettle-Spoon) and scheduling them using Cron Job

Created Pentaho jobs to all environments

Develop ETL informatica mappings and workflows, check data integrity and accuracy, and perform data cleansing procedures as needed.

Involved in writing Teradata batch processing scripts for data transformation, aggregation and load using BTEQ.

Maintain up-to-date documentation of implementation, troubleshooting, and ETL processes related to Informatica systems. Prepare documents pertaining to specific issues and how they were resolved, including coding information and extraction and transformation processes.

Developed procedures to populate the customer data warehouse with transaction data, cycle and monthly summary data, and historical data.

Worked on Exception handling process with best practices and standards for exception handling routines.

Loading data to the Salesforce, validate data and fixing issues in reject records.

Testing data in Salesforce Lightning pages by logging as a banker with different profile and permission sets.

Support in performance tuning of Informatica session for large data files by increasing block size, data cache size, sequence buffer length and target-based commit interval.

Created Autosys JIL scripts to schedule jobs.

Develop full SDLC project plans to implement ETL solution and identify resource requirements.

ETL Pentaho Developer May 2016 – Aug 2018

Comcast - West Chester, PA

Understanding the Business Requirements. Discuss with Client and Business team for any clarifications. Work closely with project Business Analyst, Data Modeler and BI Lead to ensure that the end to end designs meet the business and data requirements

Analyzing, designing and developing ETL strategies and processes, and building blueprints for Informatica development.

Involved in building of the Data Warehouse which included the designing of Data Mart and its development.

Used Informatica client tools - Source Analyzer, Warehouse designer, Mapping designer, Transformation Developer, WorkFlow Manager, Workflow Monitor.

Worked extensively on complex mappings using source qualifier, joiner, expressions, aggregators, filters, lookup and update strategy transformations to develop robust mappings in the Informatica Designer.

Defined and developed technical standards for data movement and transformation as well as review all designs to ensure those standards are met.

Handled extraction of various types of source files Flat files, XML standard source data of different transactions and loading to staging area.

Designed and written the scripts required to extract, transform, load ( ETL), clean, and move data and metadata so it can be loaded into a data warehouse, data mart, or data store.

Created new mappings and enhancements to the old mappings according to changes or additions to the Business logic.

Prepared ETL (Extract, Transform and Load) standards, Naming conventions and wrote ETL flow documentation for Stage, ODS and Mart

Used pentaho data integration to replace previously existing stored procedures and packages with pentaho jobs, thus decreasing their daily runtimes

Created seeding documents for application servers to access DB servers, to maintain and comply with organizational security standards and Created RPM spec files to deploy the pentaho jobs on the servers hierarchically

Have done Design for ETL components and developed database objects.

Perform root cause analysis on all processes and resolve all production issues and validate all data and perform routine tests on databases and provide support to all ETL applications.

Designed and developed the Informatica Power Center mappings, workflow and tasks to load multiples Source Data into different tables in Data Store, Extract Files.

Used IDQ Address Doctor Transformation to Validate and Load the Home Branch Customer data.

Parsed and generated EDI files using Informatica B2B Data transformations and generated EDI X12 Files for Partner Systems.

Developed shell scripts to run Informatica job for each XML file using PMCMD command and if any error occurs while processing each XML, error details are send to given Email ID's and continue to process the next XML file.

Have done the POC to capture the Change Data using Power Exchange in Oracle 11g Database.

Used IDQ's standardized plans for addresses and names clean ups.

Data Analyst June 2015 – April 2016

Ufours IT Solution Pvt Ltd Tamilnadu India

Utilized MS Access to build a data entry and reporting tool.

Created a Data management tool to automate data processing.

Automated Access to extract data from Excel.

MS Access as front and SQL Server as backend.

Built web crawling infrastructure using Python and VBA that extracted event data from websites that was then used to populate a search engine application with information relevant to the hospitality industry.

Developed advanced data cleaning algorithms that standardized account names and addresses as well as contact name deduplication and address verification.

Conduct Data mining on Excel reports.

Documented usage of Automation.

Extracted data from Tableau, SQL Server MySQL and Python to Excel

Scripts used primarily, VBA and SQL queries.



Contact this candidate