PRASANNA KUMAR SAPPIDI.
*********@*****.***
Data Engineer
SUMMARY:
SUMMARY:
Over 12+ years of experience in the development of client/server and multi-tiered applications with a strong focus on data management using SQL Server, Oracle, SQL, PL/SQL, and cloud-based technologies.
Extensive knowledge in the Healthcare domain, with specialized expertise in applications like Facets, particularly modules such as Provider and Enrollment.
Hands-on experience in designing, developing, implementing, testing, supporting, and documenting business applications utilizing MSBI, Azure technologies, and Snowflake for optimized data storage and processing.
Proficient in Azure Cloud services, including Azure Data Factory (ADF) for seamless data integration, transformation, and orchestration across multiple data sources, as well as using Azure Synapse Analytics for scalable data warehousing and analytics.
Adept at working with Snowflake, leveraging its cloud data platform for scalable data warehousing, optimized querying, and handling structured and semi-structured data.
Expertise in building ETL pipelines and orchestration processes using Azure Data Factory to automate data extraction, transformation, and loading (ETL) from multiple sources such as SQL Server, Snowflake, Excel, and flat files.
Experience in working with Python, PySpark, and Databricks for big data processing, transformation, and machine learning applications.
Skilled in implementing Coalesce for advanced data integration and transformation, enhancing automation and data workflows in cloud environments.
Strong command over Oracle Architecture and database design concepts such as Normalization, ER Diagrams, and optimization for large-scale systems.
Extensive experience with SQL Server in building complex stored procedures, functions, triggers, and views to process large volumes of data.
Designed and optimized complex SQL queries, stored procedures, and indexing strategies to improve data retrieval performance.
Developed SQL-based data pipelines for efficient ETL processing in Databricks and Ab Initio.
Engineered partitioned tables and indexed views for high-performance query execution in large datasets.
Experience in Databricks for large-scale data transformations using SQL, PySpark, and Scala.
Designed and implemented data pipelines using Databricks, leveraging its capabilities for big data processing and machine learning, ensuring efficient data transformation and analysis.
Worked with AbInitio to develop and optimize ETL workflows, ensuring high data integrity and performance.
Designed ETL pipelines for structured and semi-structured data, ensuring data quality and governance.
Implemented data validation mechanisms within ETL pipelines to monitor and troubleshoot data inconsistencies.
Proficient in ETL tools like SSIS, DataStage, and Informatica, with a track record of designing and implementing efficient data integration and transformation processes.
Expertise in performance tuning for SQL Server and PL/SQL to optimize processing speed for large datasets, ensuring efficient data workflows and reporting.
Developed backend data processing logic using Java to support complex transformations in ETL pipelines.
Built custom Java-based ingestion routines to extract, transform, and load data from multiple sources.
Integrated Java libraries for data validation, cleansing, and enrichment in large-scale datasets.
Implemented Java-based APIs for real-time data processing and integration with cloud-based data platforms.
Solid understanding of HIPAA EDI transactions 834 and 835, ensuring compliance in healthcare data exchanges.
Skilled in troubleshooting Facets Batch Jobs and Tidal Scheduling for efficient batch processing and scheduling automation in enterprise systems.
Worked across various SDLC methodologies including Waterfall and Agile, with a focus on iterative development, testing, and continuous delivery within the data management and cloud space.
Demonstrated strong analytical, problem-solving, and communication skills to collaborate effectively across teams, including Business Engineering, Quality Assurance, and Database Administration.
Microsoft Certifications:
https://learn.microsoft.com/en-us/users/prasannakumarsappidi-0987/transcript/d8x5uoy0qjkw65v
EDUCATION:
Bachelor of Technology: Vijayawada, India - 2009
Master’s from University of North America: Arlington, VA – Sep 2010 to Dec 2012
TECHNICAL SKILLS:
Databases
Oracle 11g/ 10g/ 9i/ 8i, MY SQL, DB2, MS ACCESS, SQL SERVER
Cloud Platforms
Azure Synapse Analytics, Azure Data Factory (ADF), Azure Data Lake Storage, Azure-SSIS Integration Runtime, Snowflake
Languages
SQL, PL/SQL, C, C++, HTML, UNIX Shell Scripting
Database Tools
TOAD, SQL Developer, Oracle Enterprise Manager, SQL*Plus, SQL*Loader, SQL Server Management Studio, Microsoft Visual Studio
ETL TOOLS
DATASTAGE, Coalesce, SSIS, Informatica Power Center 10.x/9.x, Informatica Power Exchange 9.x
Reporting Tools
SSRS, Tableau, SSAS (OLAP Cubes), Power BI
Other tools & Utilities
JIRA, CTU tool, Tidal, Control-M, Business Objects, SQL Developer, OBIEE.
Methodologies
Agile, Waterfall, SDLC
Concepts
Data Warehousing, Performance Tuning, Data Governance, HIPAA Compliance (EDI 834/835)
PROFESSIONAL EXPERIENCE:
HCSC (Health Care Service Corporation). June 2022 to Current
Location: Remote
Role: Data Engineer
Project Summary:
HCSC focused on modernizing the healthcare company’s data infrastructure to improve patient care, operational efficiency, and regulatory compliance. By integrating data from various sources, including electronic medical records and financial systems, a centralized data warehouse was created to streamline reporting and analytics. This solution provided healthcare administrators and clinicians with actionable insights into patient outcomes, resource utilization, and performance metrics, ensuring data quality and adherence to industry compliance standards.
Responsibilities:
Collaborated with the enterprise architecture team to explore and implement Microsoft Fabric for unified data experiences, leveraging Fabric’s integrated services including Lakehouse, Data Factory, Synapse Real-Time Analytics, and Power BI.
Built and tested Lakehouse architecture within Microsoft Fabric to consolidate structured and semi structured data from Azure Data Lake Storage into OneLake for high-performance analytics.
Developed Direct Lake datasets in Power BI on top of Fabric Lakehouse to enable near real-time reporting without the need for data import or refresh, improving business decision timelines.
Implemented data pipelines using Microsoft Fabric Data Factory to orchestrate ingestion and transformation flows, integrating with external sources such as SQL Server and Oracle.
Designed and implemented complex ETL/ELT workflows using SQL Server Integration Services (SSIS), incorporating advanced transformations such as Lookups, Merge Joins, Fuzzy Lookups, Conditional Splits, and Script Components to handle large-scale data integration.
Built data integration workflows in ADF leveraging activities such as Copy Data, Data Flow, and Mapping Data Flow to handle large-scale data transformations and ensure seamless data movement between on-premises and cloud environments.
Created a centralized data lake in Azure Data Lake Storage for scalable storage of structured and semi-structured healthcare data, enabling downstream analytics and reporting.
Utilized Databricks for big data processing, developing scalable ETL pipelines to ensure data integrity and timely delivery
Developed and optimized data processing workflows using Python, PySpark and Databricks to efficiently handle large datasets and perform advanced analytics.
Integrated Fabric with existing ADF and Databricks pipelines to create hybrid workflows and gradually migrate to Fabric-native services.
Leveraged extensive experience with SQL in a Relational Database Environment (SQL Server and Oracle) to develop optimized queries, stored procedures, views, and triggers for data extraction, transformation, and reporting.
Engineered partitioned tables and indexed views to enhance database performance for high-volume datasets, ensuring efficient query execution and data retrieval.
Designed and developed parameterized, drill-down, and sub-reports in SQL Server Reporting Services (SSRS), providing actionable insights to stakeholders.
Migrated legacy DTS packages to SSIS, enhancing workflows with dynamic configurations, error handling, and logging mechanisms to ensure robustness and scalability.
Migrated legacy SSIS packages to Azure-SSIS Integration Runtime, optimizing and deploying them in the Azure environment to leverage cloud scalability and performance.
Designed and developed SQL-based data pipelines for ETL processing in Databricks and Ab Initio.
Created Java-based data ingestion routines to extract and transform data from multiple sources.
Developed complex SQL queries, stored procedures, and indexing strategies for performance optimization.
Built data validation mechanisms to ensure data integrity within ETL processes.
Implemented Java-based API integrations to streamline real-time data processing.
Monitored and troubleshot data pipelines to ensure efficient and timely data delivery.
Worked on Azure Data Factory and Azure Data Lake Storage to store and transform structured and semi-structured data.
Created reusable SSIS packages for complex data transformations, including data cleansing, deduplication, and archiving, adhering to business and compliance requirements.
Built OLAP cubes in SQL Server Analysis Services (SSAS) to support advanced data analytics, enabling multidimensional reporting and decision-making.
Developed scalable ETL pipelines to handle structured and semi-structured data, leveraging script tasks and custom components in SSIS for unique business requirements.
Developed complex transformations in Azure Data Factory, incorporating JSON-based pipeline configurations, custom scripts, and integration with Azure Functions for dynamic data processing.
Contributed to the data warehouse architecture by collaborating with teams to define data modeling strategies, optimizing schema designs, and ensuring consistency across systems.
Provided operational support, resolving incidents, service requests, and planned support activities, including system monitoring and high-availability configurations.
Contributed to organizational planning around Microsoft Fabric adoption roadmap, assessing benefits of unified governance, data lineage, and security across data engineering and BI workflows.
Followed SDLC guidelines, producing comprehensive documentation for all ETL/ELT processes, system designs, and operational workflows.
Implemented change management processes to track and deploy system enhancements while mitigating risks and ensuring compliance.
Environment: AZURESQL Server 2022, Reporting Services (SSRS), Integration Services (SSIS), Analysis Services (SSAS), DTS, Oracle 9i/8i, T-SQL, Windows, XML, MS Excel and MS Access, SAS, Linux.
Client: HealthAxis Tampa FL (remote) July 2021 – June 2022
Role: ETL Developer
Project: Healthaxis Datawarehouse
Description: The project’s goal was to consolidate data from multiple source systems into a centralized data warehouse, utilizing Snowflake’s cloud architecture for storage, SSIS for ETL processes, and Snowflake for advanced data transformation and orchestration.
Responsibilities:
Designed and implemented cloud-based data models for a data warehouse leveraging Snowflake, ensuring scalable, high-performance storage and querying capabilities for real-time data integration.
Analyzed user requirements to identify data sources and map them to Snowflake tables, ensuring efficient data loading and transformation workflows.
Used Coalesce to automate the complex transformation logic, performing data cleaning, enrichment, and aggregation before loading into Snowflake. This helped to streamline ETL operations and improve performance in handling large datasets.
Built SSIS packages for data extraction, transformation, and loading (ETL) tasks, ensuring seamless integration from various source systems (flat files, MS Excel, MS Access) into Snowflake. The transformations were optimized using Coalesce’s advanced capabilities.
Configured Snowflake to handle semi-structured and structured data, implementing best practices for optimizing query performance and minimizing data storage costs.
Developed dynamic SSIS data workflows to load, transform, and cleanse data based on user-defined business logic and requirements, ensuring consistency and accuracy across different data sources.
Developed and optimized ETL workflows using Databricks and Snowflake.
Designed SQL-based data pipelines for efficient data extraction, transformation, and loading (ETL).
Built Java-based data ingestion scripts to integrate data from external sources.
Implemented data quality checks and validation mechanisms for high data accuracy.
Managed Azure Data Factory (ADF) workflows for scalable and automated ETL processes.
Improved query performance by optimizing indexing strategies and partitioned tables.
Leveraged Coalesce to orchestrate and automate data pipelines, improving the data flow from staging to data mart while reducing manual intervention and error-prone processes.
Supported data migration to Snowflake, ensuring smooth data transitions, real-time updates, and maintaining data integrity through automated validation checks during the ETL process.
Created comprehensive documentation for ETL standards, including naming conventions and detailed flow mappings for both staging (ODS) and mart layers in Snowflake.
Collaborated with the team to optimize the ETL pipeline for performance, ensuring efficient data archiving processes and faster data loading times by utilizing Snowflake’s elastic compute and storage scaling.
Environment: Snowflake, Databricks, Azure Data Factory, Reporting Services (SSRS), Integration Services (SSIS), Analysis Services (SSAS), DTS, Oracle 9i/8i, T-SQL, Windows, XML, MS Excel and MS Access, SAS, Linux.
Client: Anthem Inc. Norfolk Nov 2016 – July 2021
Role:SQL Server/ MSBI Developer
Project: Enrollment
Working on multiple applications/projects like NY and WNY Health Homes, UET application, multiple Medicaid, Medicare, Caremore Reports, PCP assignments for all Medicaid markets, enrolling using 834 file as well as proprietary file to process new/existing member data in facets.
Responsibilities:
Managed the migration of SQL Server 2012 databases to SQL Server 2014.
Created tables, stored procedures, and defined functions, including SQL scripts for tuning and scheduling.
Fine-tuned stored procedures for performance improvement by removing unnecessary cursors and using SQL Profiler.
Developed physical and data warehouse models, creating DDL scripts for database schema and object creation.
Generated periodic reports based on statistical analysis of data across various time frames and divisions using SQL Server Reporting Services (SSRS).
Identified unusual data and errors requiring further investigation using data mining techniques.
Developed operational, drill-through, and drill-down reports using SSRS.
Designed Star and Snowflake schemas, identifying facts, measures, and dimensions.
Migrated DTS packages to SQL Server Integration Services (SSIS) and adapted them to use advanced SSIS features.
Migrated data from heterogeneous data sources (DB2, Access, Excel) to SQL Server databases using SSIS.
Created ETL packages with SSIS to validate, extract, transform, and load data into data warehouse and data mart databases.
Conducted client walkthroughs to train users and assist with ad-hoc report requests.
Experienced in defining, designing, and re-engineering enterprise data warehouses and data marts in environments like Teradata and Oracle, managing multiple terabytes of data.
Involved in data modeling and physical database design using tools like Erwin.
Created and delivered personalized reports via MicroStrategy Narrowcast Server to a wide range of users.
Designed and created ad-hoc reports based on client needs using a new reporting tool system.
Provided ad-hoc and pay data reporting for accounting and upper management.
Troubleshot system issues, monitored scheduled jobs, and set up maintenance plans to proactively monitor SQL Server database performance.
Configured transactional replication between production and standby servers for disaster recovery.
Created calculated measures and dimension members using MDX.
Managed user roles and permissions, maintaining security up to date.
Developed Perl and Shell scripts to automate finance billing file processing.
Environment:SQL Server 2012/14 Enterprise Edition, .Net 3.5, SQL BI Suite (SSAS, SSIS, SSRS), OLTP, OLAP, Enterprise manager, PPS, XML, MS PowerPoint, Tableau Desktop 9/10, MS Project, MS Access 2014 & Windows Server 2014, Oracle, Crystal Reports, Cogon's, Business Objects, SharePoint.
Anthem Inc. (Amerigroup) Virginia Beach VA May 2013 - Nov 2016
Role:MSBI/SQLDeveloper
Project: Provider
Amerigroup manages Medicaid, CHIP, Medicare and Long-Term Care programs covering 2 million members in 12 states. Anthem acquires Amerigroup and changes the name as GBD Government Business Division. The goal of the project is to support GBD provider system for Medicare providers as well as Medicaid providers.
The goal of the project is to create multiple provider extracts for provider data management team like loading the KYIP07 files from Central Facets to GBD (Government business Division) for migrating all the KY Medicaid information from Central Facets to GDB system.
GBD uses Custom application for loading the providers into Facets which is called the Provider Roaster Automation. Created the Multiple reports from RAM tool for supporting the business-like Error Reporting and Detailed Report (how many providers loaded through RAM, how many are failed etc.)
WellPoint Medicare member’s information is available in Medisys. Created the extract for sending the GBD Blue (Medicare provider’s information) to Medisys system for assigning PCPS to the members.
WellPoint Maintains the Medicare information in the EPDS (Enterprise Data System) system. The goal of the project is migrating the Medicare data from EPDS to GBDsystem. Actively worked for migrating the Medicare data from EPDS to GBD System.
Care more is another system for WellPoint. Developed an automated provider demographic feed from the internal GBD Facets system to a subsidiary Care More to support the management of the long-term services provider network in Virginia and California
Responsibilities:
Designed, developed, and tested the ETL (Extract, Transform, Load) strategy to populate data from various source systems using SSIS.
Worked on Data Flow tasks such as For Each Loop Container, Execute Process Task, Transfer SQL Server Objects, Jobs, and Logins tasks.
Handled SSIS packages involving FTP tasks, Fuzzy Grouping, Merge, Pivot, Unpivot, and Control Flow Transformations.
Utilized SQL Server Package Configuration for dynamic configurations.
Created analysis reports using Crystal Reports based on key variables stored in the database.
Extracted data from heterogeneous databases like Oracle, Access, DB2, and flat files to SQL Server 2014 via SSIS.
Performed ETL testing and data migration testing for Dimension and Fact Tables.
Conducted testing on Data Quality Rules based on business requirements and report testing.
Developed SSIS packages to extract and load data into SSMS for further tuning.
Created ETL packages to validate, extract, transform, and load data into data warehouses and data marts.
Implemented event handlers and error handling in SSIS packages, notifying users of process results.
Designed ETL jobs for data extraction, transformation, and loading into Data Marts.
Created and optimized databases, tables, indexes, constraints, views, stored procedures, and triggers.
Optimized stored procedures and long-running queries through indexing strategies and query tuning.
Generated reports using Microsoft SQL Server Reporting Services (SSRS) across multiple versions (2014/2012/2008R2).
Participated in report design analysis and interacted with the Team Lead to finalize specifications.
Regularly updated Power BI Dashboards per business requirements.
Performed backup operations and developed customized interactive dashboards using Excel Power BI.
Experienced in designing and modeling SSAS Analytics Multidimensional cubes.
Developed complex stored procedures, views, and temporary tables.
Maintained and developed Tableau dashboards.
Created complex reports with dropdown menus and advanced groupings.
Designed and built dimensions and cubes using star schema with SQL Server Analysis Services (SSAS).
Built high-performance data integration solutions using SSIS for data warehousing.
Integrated and analyzed data from multiple heterogeneous sources, including Oracle.
Environment: MS Access 2007, MS SQL Server- 2014/2012/2010/2008 R2 Enterprise Edition ANSI SQL, Visual Studio 2008, ETL, SSIS, SSAS, SSRS, T-SQL, DTS, SQL Profiler.