Close to * Years of professional experience in IT industry catering to various clients and projects.
Experienced in the areas of data analysis, design, Informatica, Teradata, Exasol development, performance tuning, Upgrading and Development including Requirement Gathering, Analysis, Defect analysis, Root Cause analysis for data / design / code issues and its resolution with permanent fix, Configuration, Migration and deployment of Full Life Cycle Data warehouse project using Fact Tables, Dimension Tables, Star and Snowflake schema modeling.
Having knowledge of the best practices for Informatica PDO (Push down optimization) for Teradata.
Created UNIX shell scripts to run the Informatica workflows and controlling the ETL flow.
Having experience in implementing JSON file read transformation using expression transformation in Informatica 9.1.
Created mappings with transformations like (Stored Proc, SQL Transformation, Transaction Control,XML generator,XML parser)
Having sound knowledge in Teradata Architecture.
Having knowledge in Teradata Utilities (BTEQ, FastLoad, MultiLoad, FastExport, TPT Operators).
Experienced in implementing Slowly Changing Dimensions.
Having sound knowledge on SQL and validating data from various source systems using SQL.
Having knowledge on Exasol Analytical database.
Having sound knowledge on IBM’s BDW (Banking Data Warehouse) model.
Having experience of using Control-M Workload Automation extensively to run ETL jobs, schedule, monitor and to obtain the run time performance statistics.
Having knowledge on bigdata technologies like HDFS, Sqoop, Hive etc.
Having experience in data analytics and dimensional data modeling.
Having experience in handling very large database (VLDB).
Learned Power BI tool for business intelligence reporting purpose.
Having experience using ITSM (IT Service Management) tool for code migration to higher environments.
Having experience working in Agile model and JIRA tool for the same.
Independently perform complex troubleshooting, root-cause analysis and solution development.
Ability to meet deadlines and handle multiple tasks, decisive with strong leadership qualities, flexible in work schedules and possess good communication skills.
Effective in working with various business stakeholders, Support Managers, Senior Project Managers, Project Managers, Business Analysts, QA Analysts and Solution Architects
Title of the Degree with Branch
College/University
Master of Science in Computer Science
VIT University, Vellore, Tamil Nadu, India
ETL Tool
Informatica Power Center 9.1/9.6/10.1
Database
Teradata, Exasol
BigData Technologies
HDFS, Sqoop, Hive
Data Replication Tool
Attunity Replication
BI Reporting Tool
Power BI
Scripting Languages
SQL, Shell, Python (Learning)
Schedulers
Control-M
Version Control
SVN, Gitlab
Code Migration Tool
ITSM, Gitlab CI/CD pipeline
Domain Knowledge
Retail Banking, Agricultural model
Responsible for analysis and understanding of functional requirement specification documents and Understanding existing business model and customer requirements.
Enhancement and new development of Informatica workflows, worklets, mappings, and sessions based on ETL Specification
Enhancement and development of Teradata DDLs, SCD using views, TPT scripts, pack factors based on volumes and performance tuning using collect stats, explain plan and proper joins
Involved in Requirement Analysis, Effort Estimation, Impact Analysis, Code designing, Developing, Deployment of ETL components
Performance tuning of ETL components Informatica and Teradata jobs
Preparation of Test plans and test cases, raising issues/challenges, defect tracking, reviewing the status of the deliverables and migration of objects from lower to upper environment
Preparation Technical Specification Document (TSD)
Production break fix and end user defect analysis and solution
Performed Regression Testing and mismatch Analysis and performed Defect Analysis to find the root causes of these issues.
Being ETL Lead maintained all the documents and tracked all the deliverables according to the schedule.
Helping data modeler for Dimensional modeling based on the ETL specification
Developing new job/job group in Control-M scheduler.
Implementation and deployment using ITSM process
Agile Sprint planning and creating JIRA stories and resolving them on time
Subject Matter Expert (SME) from offshore
Peer code review process.
Mentoring of junior resources
Project Details in Corteva AgriScience
Project# 1
Title
Informatics for R & D
Functional Domain
Agriculture
Project Abstract
Informatics for R & D Platform – It is a data integration project which objective is to integrate the data from the different sources to the one centralized enter prized data warehouse called to maintain single version of truth.
Technical Domain
ETL – Data Warehousing
Tools used
Exasol, AWS, Teradata, Informatica Power Center 10.1, Unix Shell Script, Tivoli Work Load Manager as scheduler
Team size
12
Role
Datawarehouse Engineer and Data Analyst
Period
18 Months (Currently working)
Project Details in Cognizant Technology Solutions
Project# 1
Title
Banking Data Warehouse
Functional Domain
Retail Banking
Project Abstract
It is a three layer architecture – Staging, Integration and semantic. Files from the source will be directly loaded in load tables of staging layer. From staging, data will be moved to integration layer with applied transformation logic on data with other additional audit fields. Integration data will be further moved to semantic layer with all customized transformation logic according to business needs.
Technical Domain
ETL – Data Warehousing
Tools used
Informatica Power Center 9.6, Ab initio, Control – M scheduler, Teradata 14, Unix Shell, SVN, ITSM, JIRA
Team size
6
Role
Senior ETL Developer and Data Analyst
Period
48 Months
Project# 2
Title
ETL Migration
Functional Domain
Retail Banking
Project Abstract
IDN ETL Migration – It is migration project from Ab initio and Sybase based platform to Teradata based platform. ETL tool is built with proprietary tool called Metadata driven Framework (MDF) with the help of Teradata to achieve the primary goal “Feed Per week”.
Files will be SFTPed from the mainframe system and will be loaded into data landing zone and then data will be transformed and loaded into mart layer.
Technical Domain
ETL – Data Warehousing
Tools used
Teradata 14, Event Engine, MDF (Metadata Driven Framework), DDC (Data Development Checklist), CVS
Team size
15
Role
ETL Developer
Period
9 Months
Project# 3
Title
Data Governance Scorecard
Functional Domain
Retail Banking
Project Abstract
Data Governance Scorecard – It is a data governance project which requires to build scorecard for last month, last 3 months, last 6 months and one year about data quality, number of jobs failed in production, about financial data and vendors of Amex and platform uptime and downtime.
Technical Domain
ETL – Data Warehousing
Tools used
Teradata 14, Unix Shell, CVS
Team size
4
Role
ETL Developer and Data Analyst
Period
6 Months
Project# 4
Title
Clinical Integration Platform (CIP)
Functional Domain
Healthcare
Project Abstract
Clinical Integration Platform – It is a data migration project which objective is to migrate the data from the different regional legacy warehouses using Teradata Fast Export utility, Fast Load utility and Informatica into the one centralized enter prized data warehouse called “EdWARD (Enterprise Data Warehouse and Research Depot)” to maintain single version of truth.
Technical Domain
ETL – Data Warehousing
Tools used
Teradata 13, Informatica Power Center 9.1, Unix Shell Script, Work Load Manager as scheduler
Team size
6
Role
Junior ETL Developer
Period
13 Months
I hereby declare that all the information furnished above is true to the best of my knowledge and belief.
Place: Hyderabad SOUGATA KHATUA
Sougata Khatua
ETL Developer, Teradata Developer, Data Analyst
Mobile No: 984*******/912-***-****
Email Id: *************@*****.***
Current Address
Nizampet
Hyderabad, Telengana
PIN: 500090
India
Key Highlights
Educational Qualification
Technical Proficiency
Roles and Responsibilities performed
Professional Project Experiences
Declaration