Mobile: +1-726-***-****
Email : ********.********@*****.***
Data Architect
Profile:
•Over 17.5 years of diversified IT experience with 5 years of experience as an Enterprise Data Governance Architect with hands-on expertise in Alation, BigEye, Ataccama, Information Analyzer, Informatica Metadata Management tool; 6 years as a Data Engineer with hands-on expertise in Snowflake implementation, Databricks Implementation on AWS.
•Delivered DAMA-DMBOK policy control, CDE rationalization, and catalog/lineage at scale. Built Python automation for Alation bulk integration APIs and MDE scheduling; operationalized lineage with Manta (SOO/EOO). Worked with Multiple lines of Business within an Enterprise to identify issues of siloed Data Governance and centralize all DG assets.
•2 years of experience in creating visually compelling reports and dashboards using Tableau, Power BI, complex SQL using CTE, and Initial SQL usage.
• 5 years of experience in Design, Developing/customizing and implementing Automation frameworks using SnowPark, Snowflake stored procedures, Alation API, along with Alation Analytics SQL reports, and Data Profiling.
•8 years of experience in the designing and development of Quality Technical Specification Document (HLD and LLD), Source to Target Mapping, and ETL code in Datastage, Informatica.
•Created Technical Solution Architecture document for the migration of on-premises applications to Snowflake and Databricks.
•Implemented Medallion architecture by engineering reusable Snowpark and Snowflake stored-procedure frameworks to design, develop, and maintain end-to-end data pipelines from AWS to Snowflake.
•Extended the same framework to Databricks, leveraging PySpark DataFrames and PySpark SQL for automated ingestion.
•Experience in Power BI report analysis, including strong proficiency in Power BI Desktop, Power BI Service, DAX, and M Query Language to build a Source to Target mapping document to create Data Lineage in Alation.
•Advanced knowledge of data modeling – BIM, LDM, PDM in Erwin, Infosphere Data Architect (Star Schema, Snowflake Schema), data visualization, and data transformation techniques.
•Experience in No SQL databases like Tigergraph, Cassandra, and ArangoDB using graph languages like Gremlin, GraphQL, and Rest API to get live data from Kafka
•Experience with CICD pipelines to support deployment and integration workflows.
•Strong customer-centric approach with excellent analytical, coordination, presentation, and leadership skills.
Professional Experience:
Truist August 2024 – Till Date
Charlotte, NC.
Data Governance Lead
After the high-profile merger of BB&T and SunTrust to form Truist, a comprehensive Data Governance initiative was essential to unify and streamline data management across the newly combined enterprise. The primary objective of this project is to establish a robust and unified Data Governance framework post-merger, harmonizing data definitions, terminology, and standards to ensure consistency and accuracy across Truist's business units and data domains
Responsibilities:
Enterprise Data Governance Leadership: Translate corporate policies/standards into enforceable controls, procedures, and measurable KPIs; embed governance into day-to-day data/analytics delivery with audit-ready evidence.
CDE Lifecycle & Template Onboarding: Drive define steward approve catalog; onboard applications/EUCs/reports into the Enterprise CDE template with owners, classifications, and usage context.
Metadata Management: Develop and manage metadata to provide clear definitions, context, and usage guidelines, facilitating data understanding and accessibility for stakeholders. For all business units of Truist enterprise (more after the merger of BB&T and SunTrust), discuss with owners and Information Architects to collect and finalize critical data elements and catalog them in Alation
Business Term Rationalization (Automation): Use Python/NLP + rule scoring to de-dup/merge overlapping terms (post-merger); align with SOO/EOO; write decisions back to Alation with justification/audit trail.
Steward Enablement & Alation Adoption: Onboard Business Data Stewards; coach on Alation search, Collections, Articles, lineage views, and approval workflows; run office hours and publish quick-starts.
Column- & Table-Level Lineage Engineering: Build API-driven pipelines (Python + ETL/reporting endpoints) to normalize multi-hop paths and publish table/column lineage to Alation; monitor coverage/freshness.
Governance Reporting & Analytics: Deliver SQL-driven dashboards (catalog coverage, lineage %, DQ scorecards, policy exceptions, access reviews) and present trends/actions to governance forums.
AIG May 2024 – Aug 2024
Austin, Texas
Data Integration Lead
AIG embarked on a transformation program to simplify HNI Auto product running on Exigen system (CIP system). New HNI product would be built in Duck Creek system. A conversion engine will be developed to automate the processes for flat cancel policy in CIP and rewrite it in the Duck Creek system.
Responsibilities:
Attend the client workshop sessions to collect all the requirements for Data Migration. Get queries clarified and signed off from Business.
Design a data migration approach to migrate Policy data from AIG system to PCS system and then call Duck Creek REST API services.
Create a Data mapping document to help developers understand how the Duck Creek XMLs will be created, relating the Duck Creek UI elements to source system data and the corresponding conversion logic.
Design a Python accelerator to fetch data from Source Oracle (CIP) and run the conversion logic to generate the Duck Creek compatible XML. For each Policy, there will be a single element in XML
Apple Dec 2023 – May 2024
Austin, Texas.
Data Integration Lead
Apple wants to modernize its existing operational data services, where they want to build a central data store, services, and middleware that extend Siemens Teamcenter Product Lifecycle Management to facilitate data storage, sharing of the product data, as well as the governance of product data
Responsibilities:
Design and develop a data service layer where users and external applications will not have direct access to data, but they can send consumer requests to me and then consume the data using REST API services.
Designed, built, and integrated data into the graph schema in Tigergraph to consume data from Apple Product Management system (TeamCenter). Flattened the REST payloads into graph vertices and edges using Apache Kafka.
Developed a Python Data Quality check framework to automatically create the SQL for each table for multiple columns and run the SQL and load the Data Quality Result table in target system.
Formulated a DQ Logic configuration to be used by the Data Quality check framework.
USAA Oct 2021 – Dec 2023
San Antonio, Texas
Data Architect
IBM partners with USAA in their Modernization journey, maintaining their Data Modeling, Snowflake Cloud Data Migration, and BI capabilities based on Snowflake and Data Governance to unify and streamline data management across the data movement into Snowflake for each line of Business.
Responsibilities:
•Involved with establishing and guiding Data Governance workgroups (data standards, data policies, master data, data Catalog etc.) in cooperation with data stewards. Data Profiling and Proposing Data Quality Rules based on Data Quality dimensions. Gathering business, technical, and operational metadata from different SMEs, Data owners, Information Stewards & Information Architects.
•Migration of Business metadata from IGC to Alation. Planned and phased out the migration effort based on multiple lines of business in USAA.
•Cloud data migration for few LoBs using Snowflake and some LoBs using Databricks.
•Implementation of existing Infosphere Information Analyzer Data Quality rules in BigEye. Discussion with Product vendor BigEye to include new features and test them in USAA.
•Work with client admin of Bigeye and Alation to setup health agent between Bigeye and Alation.
•Involved in Sensitive Data Element (SDE) classification, Critical Data Element (CDE) identification.
•Shouldering the responsibility of developing to help my team gain momentum.
•Led migration of 75 reports to Tableau from SAS-based user-defined reporting and existing Power BI reports; for the 28 existing Power BI reports, required deep analysis of Power BI semantic layers (shared datasets, relationships, DAX measures) and mapping them to Netezza warehouse objects.
•Analyzed PBIX/shared datasets to trace fact/dimension usage, RLS roles, and time intelligence logic; produced a source of truth mapping of Power BI model entities to Netezza tables/views.
Century Link Dec 2020 – Oct 2021
Bengaluru, India.
Data Modeller
IBM works with Century link to provide various information technology and consulting services including Information Governance, Information Architecture, Data Quality and Business Intelligence. IBM maintains and enhances BI capabilities of Century Link by providing technical solutions and its Implementations. Data was migrated from MDM to Cassandra (AWS S3).
Responsibilities:
•Creation of Business Information Model.
•Creation of Logical and Physical Model.
•Creation of Query driven Reporting Layer model based on the Physical Model.
•Shouldering the responsibility of developing to help my team gain momentum.
•High Level Design document & Low-Level Design document and signoff
•Quality Technical Specification Document and Source to Target Mapping
•Performing the migration from development to pre-prod and then production; preparing the test cases for Unit testing and Integration testing
•Facilitating the enhancement of the application as per user’s requirement through change request; interfacing with the users and functional people for gathering the business requirements & functional specifications.
•Building Enterprise Data Platform with Governance using Event Driven Architecture.
•Real Time Data Ingestion using Events and Transactions.
•Apache Kafka Microservices Architecture for message queuing. Worked with Java developers to help them understand how the data needs to flow from Apache Kafka Topics to Cassandra Database which uses AWS S3 as storage.
•End User Rest API's for fetching data for reporting Services. Worked with Java developers to provide SQL that is required to be built for fetching data from Cassandra tables for reporting.
USAA Dec 2017 – Dec 2020
San Antonio, Texas
Data Architect
IBM partners with USAA in their Modernization journey, maintaining their Data Modeling, Snowflake Cloud Data Migration, and BI capabilities based on Snowflake and Data Governance to unify and streamline data management across the data movement into Snowflake for each line of Business.
Responsibilities:
•Creation of Business Information Model, Logical and Physical Model.
•Creation of Query query-driven Reporting Layer model based on the Physical Model.
•Shouldering the responsibility of developing to help my team gain momentum.
•I led the delivery team to provide guidelines on the ETL code delivery, development best practices, review, and analysis.
•Led migration of 75 reports to Tableau from SAS-based user-defined reporting and existing Power BI reports; for the 28 existing Power BI reports, required deep analysis of Power BI semantic layers (shared datasets, relationships, DAX measures) and mapping them to Netezza warehouse objects.
•Analyzed PBIX/shared datasets to trace fact/dimension usage, RLS roles, and time intelligence logic; produced a source of truth mapping of Power BI model entities to Netezza tables/views.
Experience in short from March 2008 - Dec 2016
Great Eastern Life Assurance Jul 2016 – Dec 2017
Singapore, Kuala Lumpur, Kolkata (India).
Data Designer
Responsibilities:
Lead the design and development of Proof of Concept at the client location in Singapore for 4 weeks.
Lead the design and development of Proof of Concept at the client location in Kuala Lumpur, Malaysia on the CDC implementation approach to showcase the success of the approach to be undertaken for the client’s Malaysia operation.
Lead the design and development of an ETL application using DataStage, IBM CDC tool, UNIX, Netezza, and Oracle for Initial Data Load & Delta Data Load; spearheaded the estimation of the Migration & Integration phase, which was first shared with the management team and after editing from the management team, it was shared with the client.
American Insurance Group Japan Nov 2013 to Jul 2016
Kolkata, India
Data Modeler and Developer
IKEA Aug 2012 to Nov 2013
Kolkata, India
ETL Developer
Johnson Controls Aug 2011 – Aug 2012
Kolkata, India
ETL Developer
Marks and Spencer Aug 2010 – Aug 2011
Kolkata, India
ETL Developer
Lexmark and Spencer Mar 2008 – Jul 2010
Kolkata, India
ETL Developer