Post Job Free

Resume

Sign in

Data Engineer Warehouse

Location:
Irving, TX
Salary:
110k
Posted:
February 26, 2024

Contact this candidate

Resume:

Ambuja Behera Irving, TX, USA,*****

ETL Data Engineer /ETL Developer (Informatica) 972-***-**** ad3xof@r.postjobfree.com

Professional Summary

* ***** ** ********** ** Data Warehousing technology with all phases of Software Development Life Cycle (SDLC) including Business Requirement Analysis, Development, Implementations and Testing of Data warehousing and Database business systems for Telecom domain, Railway, Banking & Finance and Retail Domain.

Implemented data extraction, transformation, and loading processes to support data warehouse initiatives in On-premise and Cloud tools like Informatica Power Center, Informatica IICS/IDMC, AWS Redshift.

Experience integration data to/ from on-premise database and cloud-based solution using IICS/IDMC.

Experience working with cloud base database solution including AWS Redshift, Snowflake cloud.

Experience working with traditional on-premise database including oracle, SQL server and Teradata.

Worked with different non-relational such as flat file, xml file, Jason file.

Developed and maintained data pipelines to ingest and process large volumes of data from various sources and into AWS S3.

Extensively used Slowly Changing Dimension (SCD-Type-1, Type-2) techniques in business application.

Designed and developed complex mappings, Mapplets, tasks and workflows and tuned for processes optimization in Informatica.

Experience in Debugging and Performance tuning of targets, sources, mappings and sessions in Informatica.

Designed and optimized database structures for efficient data storage and retrieval

Collaborated with cross functional teams to understand business requirements and translate them into technical solutions

Conducted data quality checks and implemented data cleansing techniques to ensure accuracy and completeness of data

Created and maintained ETL documentation, including technical specifications and data dictionaries

Created Testcases and participate in UAT and Validation

Utilized SQL, PostgreSQL, Python, and other programming languages to manipulate and analyze data

Experience in loading data into snowflake cloud

Experience Copy/INSERT, PUT, GET commands for lading data into snowflake tables from internal/External stages.

Used Time travel/offset/query profiler in snowflake cloud

Monitored and maintained data warehouse performance to identify and resolve any issues

Optimize SQL queries and mapping to improve performance and resolve bottleneck.

Experience in working on concurrent projects in very demanding and high-pressure situations.

Experience in create RESTAPI’s in IICS.

Stayed UpToDate with industry trends and best practices to continuously improve data engineering processes and tools.

Education

Bachelor of Science, Sambalpur University, India

Certification/Course Completion

Certificate of completion: Snowpro Core COF – CO2 Exam Prep from Udemy (Snowflake)

Skills

ETL TECHNOLOGY

Informatica PowerCenter 10.x,9.x, Informatica Power Exchange 10.x, IICS, Amazon Redshift

DATA WAREHOUSE

Snowflake cloud Datawarehouse, Traditional Datawarehouse

CLOUD

AWS, IICS/IDMC, Redshift, Snowflake cloud, S3 bucket

DATABASES

Oracle 11g/10g,19c, MS SQL Server 2012, MS Access, MySQL, Teradata 16/14, Toad, Amazon Redshift,

PROGRAMMING

SQL, PL/SQL, Shell Scripting (Unix/Linux), Python, PostgreSQL

OPERATING SYSTEMS

Windows, UNIX, Linux

FILES

Json files, Avro, csv, Flat files

BI REPORTING

Tableau

Version control

Git, Bitbucket, Jenkins

Methodology

Agile, Waterfall

Others

TOAD, SQL*PLUS, SQL*Loader, WinSCP, Putty, DB Visualizer, IBM Mainframe, Control-M, GIT, Ultra Edit, Redwood, MS Office, Jira

Experience

ETL Data Engineer SYSCO – Houston, Tx Apr, 2023 – Current

SYSCO is the global leader in selling, marketing & distributing food products to restaurants, healthcare & educational facilities, lodging establishments & other customers who prepare meals away from home.

working with Telematics project Team, collaborating with various team to develop/test/implement/support to convert existing ODI Roadnet to cloud on AWS into Amazon S3 and Redshift aligning with the agile methodology.

Generated Fuel Tax Fact Table for the customer to report in daily Dashboard.

Analyzed business and system requirements.

Working on converting existing informatica process into Realtime cloud process in S3

Working on translate all ETL informatica logic in to SQL in Amazon Redshift

Creating the folder in S3 bucket to store the incoming KAFKA topics on AWS.

Worked in mapping enhancement with existing code based on requirements.

Participated in all development process following from the streaming data coming into S3 to Amazon redshift.

Experience in Python script to use in data ingestion from S3 to redshift with spectrum connector.

Prepared the source target details document for all topic.

Involved in loading the Realtime refresh data into dashboard to check route, driver and event.

Prepared the Automation script for data validation using redshift.

Environment: Informatica Power Center 10.4, Redshift DB, SQL and Putty, UNIX and Teradata, Redwood, AWS and S3, PostgreSQL, Python

ETL Data Engineer /ETL Developer (IICS\IDMC\Snowflake) PNC – Pittsburg/Remote May, 2022 – Apr, 2023

PNC Financial Services is on the list of largest banks in the United States by assets and is one of the largest banks by number of branches, deposits, and number of ATMs. The company also provides financial services such as asset management, wealth management, estate planning, loan servicing, and information processing. worked in Retail and Costing in ABC Costing group.

Analyzed business and system requirements for Data Migration

Working on migration of inbound file informatica process into Oracle Cloud with CAI.

Jira Stories assigned for each release related to Retail and Costing and Participated in all development, enhancement, testing process

Created IICS connection using various cloud connector.

Experience in create REST API IN informatica Cloud process (IICS)

Experience on informatica Cloud process (IICS) to move the files from on-premise Costing/Retails data to snowflake.

Copy/INSERT, PUT, GET commands for lading data into snowflake tables from internal/External stages.

Used Time travel/offset/query profiler in snowflake cloud.

Experienced in clone the table to restore back up in snowflake

Experience in Bulk loading in snowflake cloud with Copy Into command.

Performed load into snowflake using snowflake cloud connector in IICS to support data analytics and insight use case for separate team.

Used informatica cloud Transformations (aggregator, lookup, Joiner, router, union, expression) logic for develop mappings and enhancements.

Developed SCD1 and SCD2 load process.

Extensively used parameters (input and in-out parameter).

Developed informatica cloud task flows (parallel) with multiple mapping task and task flows

Involved in troubleshooting existing ETL bugs

Developing mappings, mapping task and task flows in IICS/IDMC.

Environment: Informatica Power Center IICS, Snowflake cloud, Oracle, SQL and Putty, UNIX, Linux, Redwood

ETL Informatica Developer JPMorgan Chase – Plano, TX June, 2021 – Apr, 2022

JPMC is an American multinational investment bank and financial services. It is one of the oldest and largest private business archives in the United States. Currently working with the Service domain. ETL data integration team and collaborating with various team to develop/test/implement various existing informatica process aligning with the agile methodology.

Analyzed business and system requirements to identify system impacts.

Optimized more than 100 Complex ETL processes for Payment system

Involved in performance tuning by determining bottlenecks at various points like look up, transformation logic, SQL override query.

Analyzed the existing mapping logic to determine the reusability of the code.

Created Unit test plans and did unit testing using different scenarios separately for every process.

Worked in agile minor release cycles the designated database developer.

Managed performance and tuning of SQL queries and fixed the slow running queries in production.

Follow the Agile Process with Bi-Weekly sprint

Attend daily standup call, PBR, Retro

Worked PL/SQL store procedure using bulk collect limit to enhance the performance.

Created testcase, test scenario data set for QA process.

Created Mapping Parameters, Session parameters, Mapping Variables and Session Variables.

Used version control GIT to push code into higher environment.

Environment: Informatica Power Center 10.4.1, Oracle 19c, SQL and Putty, JIRA, Unix. TOAD, GIT. Control M

ETL Developer BNSF – Fort Worth, TX Sep, 2019 – May, 2021

BNSF Railway is one of North America’s leading freight transportation companies, with a rail network of 32,500 route miles in 28 states and three Canadian provinces. To support the BNSF railway data, developing and supporting data load using Power Exchange CDC workflows in Realtime basis. There are more than 500 tables being refreshed everyday using various workflows.

Worked with ODS Realtime Refresh to load the into ODS

Responsible for effective communication between the team and client who were working from different location (Offshore).

Used informatica commands (PMREP, PMCMD) for execution of workflows, develop enhancements.

Extensively used Informatica debugger to figure out the problems in mapping. Also involved in troubleshooting existing ETL bugs

Developing mapping, session and workflow

Testing the workflow

Created Mapping and Session parameters, Mapping and Session Variables.

Developed many complex Informatica mappings using Unconnected and Connected Lookups, Router, Aggregator, Joiner, Update Strategy

Involved in extensive performance tuning by determining bottlenecks at various points like targets,

Dev to Trial Migration-Creating Deployment Groups, migrating files on Unix server

Importing Source/Target Definition

Look up db2 catalog entries for server name/nodes

Monitor Repository services on Enterprise/SAP environments through Admin Console.

Create Group, User and give privileges in Admin console

monitor ODS workflow alerts

Restarting failed ODS workflows

Cold Start ODS workflows

Action taken if infa services are down

Involved during ETL upgrade - Compared to make sure all our Workflows, Mappings and all codes are in sync also compare the performance of the data load from previous vs changed.

Environment: Informatica Power Center 10.2(HF2), DB2, Oracle 11g, SQL and Putty, UNIX and Teradata, Mainframe and Power Exchange.

ETL Informatica Developer Verizon – Irving, TX Sep, 2016 – Sep, 2019

This project handles Verizon wireline (HSI and FIOS) services to customers. Most of the process follows the Extract-Transform-Load concept. Data extracted from the source system is dumped into a raw data table. This data is then manipulated and transformed based on the requirements and moved into appropriate database table. This project integrates data from multiple sources which contains sales, cancel, provision complete, fallouts and billing complete transactions.

Worked with DSL and FIOS Sales Datawarehouse to generate Sales metrics for the customer.

Worked with business analyst for requirement gathering, business analysis and testing, business process descriptions, scenarios and workflow analysis.

Analyzed business and system requirements to identify system impacts.

Created the Detail Technical Design Documents which have the ETL technical specifications for the given functionality, overall process flow for each particular process based on the company Standard documents.

Analyzed the existing mapping logic to determine the reusability of the code.

Handled versioning and dependencies in Informatica.

Developed many complex Informatica mappings using Unconnected and Connected Lookups, Router, Aggregator, Joiner, Update Strategy

Created Mapping Parameters, Session parameters, Mapping Variables and Session Variables.

Involved in extensive performance tuning by determining bottlenecks at various points like targets, sources, mappings, sessions or system. This led to better session performance.

Created UNIX scripts to read/write and ftp files from and to windows servers and UNIX.

Created Unit test plans and did unit testing using different scenarios separately for every process.

Performed ETL and database code migrations across environments using deployment groups.

Worked in agile minor release cycles the designated database developer.

Managed performance and tuning of SQL queries and fixed the slow running queries in production.

Environment: Informatica Power Center 9.x, Oracle 11g, SQL and Putty, JIRA, Unix, Linux and Teradata.



Contact this candidate