Post Job Free

Resume

Sign in

Technical Architect Data Modeling

Location:
Pasadena, CA
Posted:
March 26, 2024

Contact this candidate

Resume:

SASWATA DUTTA

Technical Architect

PROFILE

Over 19 years of Data Discovery & Analysis, Data Modeling, ETL Design, Development, Testing, Implementation and Troubleshooting in the field of Data Engineering and Data warehousing.

Domain Experience - 10 years in Airlines, 3 years in BFSI, 3 Year in Retail Banking sector

More than 5 years experience in Agile delivery approach.

Played various roles on Projects that required DW/ETL/Data Architect, Technical Lead, Developer

Proficient in analyzing business requirements & translate them into developer friendly technical design documents.

Experience in Big data processing, SQL,NoSql,OLTP/DW/ODS/Cloud - Data Modeling.

Having 9 yrs of working experience in US. I140 approved and need H1B extension beyond 6 year

CONTACT

PHONE:

626-***-****

WEBSITE:

Saswata Dutta LinkedIn

EMAIL:ad4ku6@r.postjobfree.com

BACHELORS IN ENGINEERING (MECHANICAL)

Institution of Engineers (India)

2000 - 2003

University of Arizona

2022 - 2024

Currently Pursuing MS in data science from University of Arizona

WORK EXPERIENCE

LTIMindtree Ltd USA(For Western Asset Management)

Technical Architect

Oct 2021–Till date

Involved In Western Assets Backoffice operations application support and development

Involved in Abinitio graph development and enhancement

Involved in graph performance improvements, Query tuning, Purging large volume of data

Involved in Analysis of Oracle stored procedure and packages

Involved in writing complex SQL and Stored Procedure and Packages.

Involved In tuning SQL queries

Involved in scheduling jobs using Autosys

Involved in Data modelling and database designing

Involved in understanding and analyzing current systems and applications

Involved in knowledge transitioning and reverse KT

Environment: Abinitio 4.1, Oracle 11g SQL,PL/SQL,Autosys JIL, Unix Shell scripts, jira,svn,CVS,Python,git,Solaris, Linux

Mindtree Ltd Bangalore, India (For Grant Thonton LLC) Technical Architect

Jan-21–Oct-21

Involved in the development of data lake in Azure Data Lake Gen2

Involved in ingestion of files from SharePoint to ADLS gen2

Involved standardization of schema of the files

Involved in cleansing and validation of data

Involved in error logging and messaging

Involved in data modelling and database designing

Involved in end-to-end ETL flow design

Involved in IICS data integration and application integration process flow development

Involved in development of IICS service connector, App connection, Call rest api from process, call DI taskflow from app integration, develop mapping, mapping task etc

Involved metadata tagging and retention policy of data lake

Participate in backlog grooming and story sizing in agile environment

Environment: Azure data lake Gen2, IICS App integration, IICS Data integration, JAVA, SQL Server, SOAP/UI,JSON, REST API, Batch Scripting, Azure Dev Ops, Azure Synapse

Compunnel Inc (For Fidelity Investment) Durham,NC, USA Software Developer

May 2020–Nov-2020

Involved in backend ETL/ELT development of Heart Beat application of Workplace Investment.

Involved in development of complex SQL, PL/SQL code, packages, tune SQL.

Create, maintain, and optimize Oracle stored procedures, functions, Packages, Views, Inline Views, database structures, write complex SQL code.

Develop PL/SQL code using bulk collect and bulk binding for performance improvement

Develop and maintain Unix shell scripts.

Develop and maintain Informatica mapping, mapplet, workflow, worklet using informatica PowerCenter 10.2.

Performs transformation and loading of data according to the business specifications in and Data Warehouse using Informatica transformation like transformation like source qualifier, filter, expression, joiner, sorter, aggregator, router, update strategy transformation, transaction control transformation, HTTP transformation, Java transformation, XML parser transformation etc.

Tuning SQL and PL/SQL code. Tuning informatica jobs and adhering best practice in ETL.

Defines, prepares, executes and implements data validation and unit testing methods to ensure data quality.

Participate in daily scrum meeting and follow agile development process.

Independently analyzes issues, uses judgement to make decisions, and takes necessary action

Participate in scrum calls, backlog grooming, story sizing, retrospective in agile env

Development of JAVA code

Environment: Oracle 12c, Oracle Exadata 12c, Informatica PowerCenter 10, Oracle SQL and PL/SQL, Control-M, Git/Stash, Source Tree, Jenkins, Jira, Tableau, JAVA, Spring boot.

TCG Digital USA (For United Airlines) Solution Architect

May 2019–May-2020

Worked on analysis of 5 different source system and source data profiling.

Worked on source data cleansing, data scrubbing and data conforming.

Translate the user requirements into detailed specifications and design the solution from the detailed specifications

Architecting database and design third normal form database and dimensional database. Involved in creating logical and physical data model.

Writing JAVA code to read data from Schedule Manager application using java api.

Develop complex mapping, mapplet, workflow, worklet, mapplet using informatica PowerCenter 10.2.

Extensive use of Informatica transformation like source qualifier, filter, expression, joiner, sorter, aggregator, router, update strategy transformation.

Develop and maintain complex SQL queries, PL/SQL code with package, procedure, code in oracle database.

Worked on SQL and PL/SQL code using EXPLAIN PLAN, DBMS_PROFILER TKPROF utility.

Develop and maintain Unix shell scripts.

Worked on designing ETL framework for auditing and error handling

Worked on ER modelling and dimensional data modelling. The data model is created based on Kimball methodologies of dimensional modelling.

Worked on ETL mapping design and development to move data from source to staging and staging to target database using Informatica.

Environment: Oracle 10g,11g, Linux, UNIX Shell Scripting, SQL, PL/SQL, JAVA, Informatica PowerCenter 10.1, SQL developer, ER Data modeler, windows 7, SVN, JDK1.6, Groovy, Grails, Tomcat 7, WebLogic 1.3, Webservice, Hibernate 1.0, putty, FILEZILLA, OBIEE 12c, BI publisher.

Interpro INC USA (For Fidelity Investment) Principal Software Engineer

Oct 2018–Jul-2019

Performs transformation and loading of data according to the business specifications in Operational Data Stores and Data Warehouse using Informatica transformation like transformation like source qualifier, filter, expression, joiner, sorter, aggregator, router, update strategy transformation, transaction control transformation, HTTP transformation, Java transformation, XML parser transformation etc.

Create, maintain, and optimize stored procedures, functions, Packages, Views, Inline Views, database structures, and Informatica mapping, workflows, Scheduling Autosys jobs, to adapt to changes in needs and requirements.

Participates in the database designing and data modelling process. Design OLTP data models and star schema.

Tuning SQL and PL/SQL code. Tuning informatica jobs and adhering best practice in ETL.

Defines, prepares, executes and implements data validation and unit testing methods to ensure data quality.

Independently analyzes issues, uses judgement to make decisions, and takes necessary action

Maintain re-useable development standards that help implement

Applies multiple testing techniques to minimize defects and deliver quality code for large and highly complex applications software.

Reviews work of other team members for accuracy, adherence to standards, optimization

Perform tasks required to conform to the Fidelity’s configuration management and software release processes

Environment: Oracle 12c, Oracle Exadata 12c, Informatica PowerCenter 10, Oracle SQL and PL/SQL, Autosys, Git/Stash, Jenkins, Jira, OBIEE

TCG Digital(For United Airlines) Chicago,USA/Kolkata India

Software Engineer/Sr Siftware Engineer/Manager Dec 2006 –Oct 2018

Major Responsibilities:

Translate the user requirements into detailed specifications and design the solution from the detailed specifications

Involved in analysis and design of staging database and target database. Involved in creating logical and physical data model.

Involved in designing ETL framework for auditing and error handling

Involved in ER modelling and dimensional data modelling. The data model is created based on Kimball methodologies of dimensional modelling.

Involved in ETL mapping design and development to move data from source to staging and staging to target database using Informatica

Involved in development of Oracle Stored procedure function, package, complex SQL queries, views, index, table partitioning, database link, load data using SQL loader,

Involved in development of Unix shell scripts.

Involved in unit testing, integration testing, regression testing, UAT.

Participated in providing the project estimates for development team efforts for the off shore as well as on-site.

Involved in POC to migrate on prem application in AWS cloud.

Tuning SQL and PL/SQL code. Tuning informatica jobs and adhering best practice in ETL.

Develop complex mapping, mapplet, workflow, worklet, mapplet using informatica PowerCenter 10.2.

Extensive use of Informatica transformation like source qualifier, filter, expression, joiner, sorter, aggregator, router, update strategy transformation.

Develop and maintain complex SQL, PL/SQL code in oracle database.

Develop and maintain Unix shell scripts.

Environment: Oracle 10g,11g, Linux, C, C++, UNIX Shell Scripting, SQL,PL/SQL, JAVA, Informatica, SQL server, SQL developer, ER Data modeler, windows 7, SVN, JDK1.6, Groovy, Grails, Tomcat 7, WebLogic 1.3, Webservice, Hibernate 1.0,putty,FILEZILLA,OBIEE 12c,BI publisher.

Capstone Project from MS program of University of Arizona

Credit Card Fraud detection:

It is a capstone project from MS program to detect credit card fraud.

Provisioning of EMR cluster

Copying data from S3 to EMR

Moving data from EMR to HIVE

Installing mongodb in EMR

Loading mongodb collection

Writing pyspark code to integrate mongodb and hive data and populate mongodb summary table

Writing python code to read streaming data from kafka and determine fraud trabsaction

Environment

HIVE, MongoDB, Spark/Pyspark, AWS EMR, Kafka, AWS EC2, AWS S3,

python

SKILLS



Contact this candidate