Post Job Free
Sign in

Data Quality Analysis

Location:
Irving, TX
Posted:
November 03, 2025

Contact this candidate

Resume:

Himadri Nath

*************@*****.*** 469-***-**** US Citizen

https://www.linkedin.com/in/himadri-nath-/

Current Location: Dallas, TX, Open for anywhere in the US.

Work Location Preferred: Dallas, TX

Qualification Summary:

●Himadri has over six years of robust experience in Data Engineering,Data Analysis, Business Analysis, Data Profiling, Data Migration, Data Quality, Data Integration, Data Mapping, Data Modeling, and Metadata Management across all phases of the Software Development Life Cycle (SDLC), including requirement analysis, implementation, and maintenance. He is proficient in Agile methodologies and has hands-on experience with Jira for project management.

●Himadri has a strong command of SQL for querying and analyzing databases, and he has worked with the Snowflake cloud database. He utilized tools like Alteryx and Tableau to blend, prepare, analyze, and visualize data, creating comprehensive reports for decision-making. He performed data profiling, cleansing, and quality checks, importing data from both relational and non-relational databases using Alteryx to ensure data readiness for Tableau.

Summary:

●With over 6 years of strong experience, worked across multiple Lines of Business (LOBs), providing data analysis, gathering business requirements, and serving as a liaison between cross-functional teams, top management, clients, users, and other stakeholders with experience in Banking and Health care Industry

●Strong experience in Data Engineer,Data Analysis, Business Analysis Data Profiling, Data Migration, Data Quality, Data Integration, Data Mapping, Data Modeling, and Metadata Management Services in all phases of Software Development Life Cycle (SDLC) like Requirement Analysis, Implementation, and Maintenance and good experience with AGILE Methodologies and Jira.

●Expertise in validating business rules, identifying missing or low-quality data, and transforming data for optimal use in Enterprise Data Warehouses.

●Skilled in analyzing business processes, eliciting requirements, defining business needs, and translating it to Business Requirement Document (BRD), Functional Requirement Document (FRD), Software Requirement Specification (SRS)/ Technical Requirement Documents (TRD), User Stories, Use-case Specification and Non-functional Requirements.

●Worked on Data Management using Collibra for Enterprise Data Governance projects on areas of Data Quality, Reference data management, Data Dictionary, Business Glossary, Data Discovery, Metadata management.

●Worked in projects involving Data Mapping, Data Profiling, Data Migration, Data Conversion, Data Quality, Data Governance, Data Lineage, Data Integration.

●Experience in interacting with Stakeholders/Customers, gathering requirements through interviews, workshops, and existing system documentation or procedures, defining business processes, identifying, and analyzing risks using appropriate templates and analysis tools.

●Experienced in various phases of Software Development life cycle (Analysis, Requirements gathering, Designing) with expertise in documenting various requirement specifications, functional specifications, Test Plans, Source to Target mappings.

●Experienced in extracting data from various sources, such as databases, APIs, and files, using ETL processes, data integration tools, and scripting languages like SQL.

●Strong experience MS excel Pivot tables and V-Lookup tables for Data Analysis.

●Proficient in data analysis using SQL, Tableau, and other data analysis techniques.

●Used SQL to extract data from Oracle ERP databases to analyze the data.

●Developed robust ETL processes using SSIS, cleansing and enriching data to ensure accuracy and consistency for analytical insights.

●Ensure Data Governance function standards - Metadata, Data Lineage, Data Mappings, Data Ingestion into the EDL appropriately, etc. to support business and technology needs to locate data, understand data, and ensure appropriate use.

●Capable of designing data warehouses and data marts using dimensional modeling techniques, such as star and snowflake schemas.

●Serve as an expert in Business Metadata in Collibra and promote understanding of the data definition at the application, data element and data domain level.

●Proficient in Python, with a strong grasp of the language's syntax, data structures, and libraries.

●Expertise in ad-hoc Python scripts to automate data validation and reporting, addressing immediate business needs and enhancing data quality processes.

●Expertise in Master Data Management, Meta Data, Informatica Business Glossary & Data Quality

●Involved in analyzing the data using MS Excel, SQL and Access and generating different reports to gain more knowledge about the business status.

●Involved in defining the source to target data mappings, business rules and data definitions.

●Worked on the implementation of Metadata Repository, Maintaining Data Quality, Data Clean-up procedures, Transformations, Data Standards, Data governance program, Scripts, Stored Procedures, triggers, and execution of test plans.

●Utilized SSIS to design and implement data integration workflows, seamlessly moving and transforming data from diverse sources for analysis.

Skills:

●Project Management: MS Project, MS Excel, Confidential Rational Portfolio Manager

●Database Management: Oracle SQL, MySQL, MS SQL Server, MS Access, Data Lineage, Hadoop (Hive, PIG), Teradata

●Data Visualization: Tableau, Power BI, Excel, Macros

●Cloud Platform: AWS, Google, Cloud, Azure

●Version Control and Documentation: MS Office Word, MS Project, Advanced Excel, Lotus Notes, GitHub Repository

●ETL: Alteryx, Informatica IDQ, Informatica Power center, SSIS

●Collaboration Tools: JIRA, Confluence

●Testing and Quality Assurance: User Acceptance Testing (UAT), Test Case Design, Test Planning

Education:

Bachelors in management information systems computer science from UT Austin, TX.

Professional Experience:

Comerica Bank,

Dallas, TX Apr 2025 - Present

Senior Data Analyst

Responsibilities:

●Worked across multiple projects focusing on data analysis, data management, and data governance, ensuring data integrity, accuracy, and compliance with enterprise standards.

●Designed and implemented complex SQL scripts and stored procedures in Snowflake to extract, cleanse, and transform large financial and operational datasets, improving query performance and scalability.

●Created and optimized Snowflake data models—including staging, raw, and curated (standardized) layers—using best practices in dimensional modeling, star and snowflake schemas, and surrogate key generation.

●Developed Source-to-Target Mapping (STM) documents to define transformations and business rules, ensuring full traceability and alignment with data governance requirements.

●Conducted data profiling and validation using SQL on Snowflake, Teradata, and Oracle to identify data anomalies, measure data completeness, and drive corrective actions.

●Defined and monitored data quality rules in Collibra integrated with Informatica Data Quality (IDQ) to measure data accuracy, validity, and consistency across key data domains.

●Supported Data Governance initiatives by documenting metadata, lineage, and business definitions in Collibra, establishing enterprise-wide visibility into critical data assets.

●Collaborated with data stewards and governance teams to standardize naming conventions, define data ownership, and enforce data certification policies.

●Built and maintained data dictionaries, ERDs, and conceptual/logical/physical models using ERwin, ER/Studio, and Lucidchart to support Snowflake-based analytics and data integration efforts.

●Tested and validated ETL mappings and orchestration workflows to ensure proper ingestion from RDBMS and flat-file sources into Snowflake, adhering to data mapping and business rule logic.

●Implemented SQL-based data reconciliation frameworks to compare source vs. target data within Snowflake and highlight discrepancies for remediation.

●Partnered with business analysts to translate functional requirements into data models and SQL logic, ensuring accurate representation of KPIs, measures, and reporting metrics.

●Automated data quality checks and metadata updates using Python and SQL tasks within Snowflake, reducing manual intervention and improving audit readiness.

●Created Power BI dashboards connected to Snowflake datasets, visualizing key data quality KPIs, lineage summaries, and domain-specific analytics for stakeholders.

FlagstarBank,Dallas,TX OCT 2022 – March 2025

Senior Data / ETL Analyst.

Responsibilities:

●Designed and automated pipelines ensuring data quality through validation, cleansing, and mismatch resolution across multiple systems.

●Implemented data ingestion and transformation processes with built-in validation, cleansing, and metadata management controls.

●Designed and implemented end-to-end ETL pipelines for ingestion, transformation, and integration of large-scale structured and unstructured datasets across Teradata, Hadoop, and cloud platforms.

●Conducted comprehensive data audits to identify and rectify data quality issues, ensuring accuracy and consistency across datasets.

●Created various data quality mappings in Informatica Data Quality (IDQ) tool and imported them into Informatica PowerCenter as mappings, Mapplets.

●Utilized advanced Excel functions such as VLOOKUP, HLOOKUP, INDEX, MATCH, and array formulas to perform data analysis and streamline workflows.

●Worked with Finance, Risk, and Investment Accounting teams to create Data Governance glossary, Data Governance framework, and Process flow diagrams.

●Integrated data remediation processes with existing ETL workflows, ensuring continuous improvement of data quality throughout data pipelines.

●Built and optimized data models and pipelines using SQL, Python, Spark, and Hive to support advanced analytics, reporting, and governance requirements.

●Developed automation frameworks for data validation, reconciliation, and quality checks, improving pipeline reliability and reducing manual intervention.

●Developed robust ETL processes using SSIS, cleansing and enriching data to ensure accuracy and consistency for analytical insights.

●Demonstrated expertise in writing and debugging complex SQL queries, performing data analysis, and ensuring data quality across multiple Big Data platforms.

●Worked on GCP for the purpose of data migration from Oracle database to GCP.

●Worked on GCP Dataproc, Dataflow, Pub Sub, GCS, Cloud functions, Big Query, Stack driver, Cloud logging, IAM, Data studio for reporting etc.

●Worked on building centralized Data Lake on GCP Cloud utilizing primary services like Cloud Storage, Cloud Functions and Big query.

●Worked on migrating datasets and ETL workloads with Python from On-prem to GCP Cloud services.

●Implemented data partitioning, indexing, and performance tuning techniques to improve processing efficiency in high-volume environments.

●Have Extensive Experience in IT data analytics projects, Hands on experience in migrating on premise ETLs to Google Cloud Platform (GCP) using cloud native tools such as BIG query, Cloud Data Proc, Google Cloud Storage, Clou Composer & Airflow.

●Hands of experience in GCP, Big Query, GCS bucket, G - cloud function, cloud dataflow, Pub/SUB cloud shell, GSUTIL, BQ command line utilities, Data Proc, Stack driver.

●Worked and extracted data from various database sources like Oracle, SQL Server, DB2, and Tera data/Big Data.

●Developed and maintained data quality metrics and dashboards to track and report on data quality issues.

●Utilized SSIS to design and implement data integration workflows, seamlessly moving and transforming data from diverse sources for analysis.

●Used SQL to query into databases and analyze data. Worked on Snowflake cloud database and used tools like Alteryx and Tableau to present study, analyze and present data.

●Tested Complex ETL Mappings and Sessions based on business user requirements and business rules to load data from source flat files and RDBMS tables to target tables.

●Leveraged JQL in automation rules to streamline workflows and reduce manual data handling within JIRA.

●Conducted Data blending, Data preparation using Alteryx and SQL for Tableau consumption and publishing data sources to Tableau server.

TexasHealth Resources ( Healthcare), Dallas, TX

Feb 2019 – Sep 2022

SQL Analyst / BI Analyst

Responsibilities:

●Blended technical and business knowledge with communication skills to bridge the gap between internal business and technical objectives and serve as an IT liaison with the business user constituents.

●Met with Customers to determine User requirements and Business Goals

●Conducted JAD sessions to gather requirements, performed Use Case and workflow analysis, outlined business rules, and developed domain object models.

●Performed Data Analysis and Data validation by writing complex SQL queries using TOAD against the ORACLE database.

●Worked with Informatica Big Data Quality (Informatica Developer and Analyst Tool) to create Column profiles, DQ Rules, Mapplets, Scorecard.

●Designed and developed ETL/ELT pipelines using tools like Azure Data Factory, Talend, Informatica IDMC, and Databricks to ingest and transform structured/unstructured data.

●Built and maintained data pipelines to move data from on-prem systems (e.g., Teradata, Oracle, SAP) to cloud platforms (GCP BigQuery, Snowflake, Azure SQL).

●Performed data modeling and schema design to support analytical workloads, KPI reporting, and regulatory submissions.

●Implemented data quality checks and validation rules within pipelines to ensure accuracy for compliance-driven reporting

●Worked with big data frameworks (Spark, Hadoop, HDFS) for processing large financial and healthcare datasets efficiently.

●Automated pipeline monitoring, alerts, and failure recovery workflows to improve reliability and reduce manual intervention.

●Created SSIS packages to load data from multiple sources to SQL Server using various transformations in SSIS

●Created SSIS packages and used Analysis Services processing task to update the cube with current data

●Created SSIS Reusable Packages to extract data from Multi formatted Flat files, Excel, XML files into UL Database and DB2 Billing Systems.

●Developed numerous Teradata SQL Queries by creating SET or MULTISET Tables, Views, Volatile Tables, using Inner and Outer Joins, Using Date Function, String Function and Advanced techniques like RANK and ROW NUMBER functions.

●Worked on Automating and Scheduling the Teradata SQL Scripts in UNIX using Korn Shell scripting.

●Developed Tableau reports for Operations team to run the daily reports.

●Integrated Snowflake with advanced analytics tools and platforms, such as Python, R, Tableau, and Power BI, for comprehensive data analysis and visualization.

●Created action filters, parameters and calculated sets for preparing dashboards and worksheets in Tableau.

●Developed storytelling dashboards in Tableau Desktop and published them on to Tableau Server, which allowed end users to understand the data on the fly with the usage of quick filters for on demand needed information.

●Utilized Alteryx analytics software to process sophisticated cost models and SQL data conversion.

●Performed data profiling, data cleaning, data quality by importing from both relational and non-relational databases using Alteryx.

●Implemented Python and R solutions for patient data analysis, predictive modeling, and healthcare research, enhancing patient care and outcomes.

●Connected to and manipulated databases using Python

●Utilized Hadoop ecosystem tools such as Hive, Pig, and HBase to perform complex data queries, transformations, and storage operations.

●Experience in Hive and Pig for querying and managing large datasets, providing structured data access and manipulation.

●Conducted data ingestion and ETL processes using Hadoop, integrating various structured and unstructured data sources into the data lake.

●Experience implementing Collibra to automate data management processes.

●Performed an end-to-end Data Lineage assessment and documentation for select CDEs.



Contact this candidate