Sujan Gottimukkula
Contact: 972-***-****/ **************@*****.***
Data Engineer
Around 14 years of IT experience processing from core SQL development to advanced cloud data engineering with string expertise in Amazon Web Services (AWS), Snowflake, and Microsoft Azure.
Extensive hands-on experience designing and implementing AWS-based data architectures, leveraging services like Amazon S3, AWS Glue, AWS Lambda, and Amazon RDS for building scalable, secure, and high-performance data pipelines.
Strong foundation in SQL Server technologies including T-SQL, stored Procedures, views, Triggers, and performance tuning, with deep expertise in SSIS, SSRS and SSAS.
Proven expertise in building scalable cloud-native ELT pipelines using Snowflake and AWS, enabling batch and near real-time data ingestion using Snowpipe and event-driven architectures.
Strong experience integrating AWS data lake architecture (S3, Snowflake) with modern ETL tools like Matillion and Informatica for enterprise grade data processing.
Extensive experience in Azure ecosystem including ADF, ADLS Gen2, Synapse Analytics, and Databricks, demonstrating strong multi-cloud expertise (AWS, Azure).
Deep understating of data warehousing concepts including dimensional modelling, Star and Snowflake schemas, data transformation, and performance optimization for large-scale datasets.
Proficient in data visualization and reporting tools such as Power BI, Tableau, Crystal Reports and Workday Reporting for delivering business insights.
Experienced in Python Programming for automation, data validation, and integration across AWS, Snowflake, and relational databases.
Strong domain expertise in Property & Causality (P&C) Insurance, with hands-on experience in Guidewire, enabling data-driven decision-making across claims, underwriting, and policy systems; extended experience across Transportation, Media, Education, Healthcare, Retail, and Federal domains, delivering scalable data solutions and supporting Regulatory Reporting requirement (claims, reserves, premiums, and compliance reporting) through accurate data mapping, transformation, and governance frameworks.
Technical Skills:
Cloud Services: Amazon Web Services (AWS), Microsoft Azure – Azure Data Factory (ADF), Azure Data Lake Storage (ADLS), Azure Blob Storage, Azure Databricks, Azure Synapse Analytics
Databases: Snowflake, Azure SQL, MS SQL Server, Oracle Corporation (8i/9i/10g), MySQL, PostgreSQL, Teradata, Amazon Redshift, MS Access
BI Tools: Power BI, Tableau, SSAS (Tabular Model), SSIS, Azure Data Factory, Azure Data Lake
Programming Languages: Python, SQL, PL/SQL, PySpark, Shell Scripting, C#, VB.Net.
ETL Tools: SSIS, Azure Data Factory (ADF), Matillion, Informatica PowerCenter, Pentaho Data Integration.
Reporting Tools: SSRS, Power BI, Power Pivot, SSAS Tabular, Crystal Reports, Tableau, Workday Reporting
Educational Details:
Masters in Information Technology, University of Mary Hardin Baylor, Texas 2012
Bachelors in Information Technology, JNTU, INDIA 2009
LinkedIn: https://www.linkedin.com/in/sujan-gottimukkula-989a0bb7/
Professional Work Experience:
General Services Administration, Frisco,TX, (Sr Data Engineer)
Aug 2022– Till Date
Executed large scale data migration from legacy auction system to Amazon Web Services using Pentaho Data Integration, loaded curated datasets into Amazon RDS and Snowflake.
Developed detailed data mapping documents to transform legacy auction data (bids, listings, users, transactions, payments) into normalized schemas within AWS and snowflake environments.
Automated schema migration, import/export processes, and environment promotions using shell scripting across development, QA, and production environments.
Designed and implemented scalable ELT pipelines using Matillion to ingest and transform auction data from Amazon S3 into Snowflake.
Configured Matillion project-level and environment variables (public/private/shared) to dynamically control data ingestion, transformation logic, and file-based loading processes.
Built batch and near real-time ingestion frameworks leveraging Snowpipe integrated with AWS storage
For continuous loading of auction events such as bids and transactions.
Migrated and optimized legacy SQL workloads from Teradata and Oracle to snowflake using SnowConvert, improving performance and reducing operation overhead.
Utilized snowflake features such as Virtual Warehouse, Time Travel, Zero-Copy Cloning, and Data Sharing to support analytics, recovery, and environment management.
Designed and created tables, external stages, and materialized views Snowflake to support high-volume auction analytics and reporting use cases.
Optimized complex SQL queries and Snowflake workloads, improving query performance for large-scale datasets involving bids, auctions, and user activity.
Developed advanced analytical queries using window functions for use cases such as bid ranking, highest bidder identification, cumulative bidding trends, and auction performance metrics.
Built and orchestrated ETL workflows using Pentaho Spoon and Matillion to handle large-scale structured and semi-structured auction data efficiently.
Implemented performance tuning strategies including partitioning, clustering, and efficient query design to handle high-volume transactional auction data.
Environment:
AWS, SQL, Pentaho Data Integration, Amazon RDS, Amazon S3, AWS Lambda, Tableau, Snowflake, Matillion Cloud ELT/ETL, Python, PowerShell
Vizient Inc, Irving, TX, (Sr Data Engineer)
May 2021– Aug 2022
Orchestrated large-scale data migration to snowflake using ADF Pipelines, enabling seamless data movement from on-prem and cloud sources.
Validated data integrity and performed data quality checks using Python (Pandas, Pyspark) scripts integrated with ADF and Azure Synapse workflows.
Designed and developed modern ETL/ELT pipelines using ADF and replaced legacy SSIS packages with scalable, cloud-native solutions.
Spearheaded database administration tasks on ADW, ensuring optimal performance and data integrity.
Demonstrated expertise in writing complex T_SQL queries, stored procedures, and functions for advanced data transformation, aggregation and analytics.
Designed and optimized complex high performance queries in Azure Synapse SQL pools and PostgresSQL, improving execution time and resource utilization.
Built robust ETL pipelines using ADF and PostgresSQL, handling data cleansing, transformation, and integration across multiple systems.
Led database and schema migrations from legacy systems to PostgreSQL and Azure Synapse, ensuring minimal downtime and consistency using automated deployment strategies.
Collaborated with business stakeholders to design and deliver Power BI analytical solutions backed by Synapse and curated data models.
Delivered Power BI dashboards to executive leadership
Utilized Power Query for data extraction and transformation
Implemented complex DAX calculations for analysis
Designed robust data models for seamless integration
Skilled in manipulating and transforming data using SQL commands like INSERT, UPDATE, DELETE and MERGE to maintain data integrity and consistency.
Environment:
MS SQL Server, Snowflake Cloud Computing, PostgreSQL, SSIS, Python, Power BI, Tableau, Excel, Azure Data Factory.
PricewaterhouseCoopers, Irving, TX, (Sr Snowflake/ Data Engineer)
Feb 2020– May 2021
Designed and implemented data mapping strategies between Guidewire ClaimCenter and downstream systems, using AWS, ensuring seamless and scalable data flow.
Built and optimized SQL queries, stored procedures, and functions in Snowflake for efficient ClaimCenter data extraction and validation.
Developed data transformation logic to map ClaimCenter objects (claims, policies, payments, reserves) to external systems using AWS-based data processing services.
Mapped insurance claims, reserves, and transactions from Guidewire ClaimCenter and PolicyCenter to regulatory reporting frameworks, leveraging AWS data lake architecture.
Designed and developed data pipelines for claims, underwriting, and policy administration using Amazon S3, AWS Glue, and Snowflake in Property & Casualty (P&C) insurance environments.
Built and optimized RLT pipelines using Snowflake and AWS services, improving performance and reducing data processing time.
Developed database objects including stored procedures, view, and constraints in Snowflake to ensure data consistency, integrity, and performance.
Implemented Snowflake Streams and Tasks for CDC and automated scheduling of data pipelines integrated with AW workflows.
Designed and implemented data ingestion pipelines to load data from Amazon S3 into snowflake using staging tables and external stages.
Built event-driven ingestion frameworks using Snowpipe integrated with AWS Lambda and notifications for near real time data processing.
Developed orchestration workflows using Apache Airflow/ AWS setup Functions to manage dependencies, including circular and FK-dependent tables.
Created python-based data validation frameworks to compare and reconcile data between snowflake and other databases such as SQL Server and PostgreSQL.
Extensively worked with Snowflake features including Temporary, Transient, and External Tables, CTEs, subqueries, sequences, and performance tuning techniques.
Environment:
Guidewire PolicyCenter, Guidewire ClaimCenter, Amazon S3, AWS Glue, AWS Lambda, Apache Airflow, Python, SQL, PostgreSQL, Snowflake.
Berkley Entertainments, Irving, TX, (Sr Data Engineer)
Feb 19– Jan 20
Worked on conversion project from CWS (Claims Workstation) to Guidewire Claim Center.
Developed ETL/ELT processes to extract, transform, and load claims data from ClaimCenter into data warehouses, data lakes, and reporting platforms.
Created data transformation logic to map ClaimCenter entities such as claims, exposures, reserves, payments, transactions, and coverages to external systems.
Created Source to Target mapping, percentage of usage of each field and identified the fields that needs to be deprecated.
Created Azure Data factory pipelines for loading the data to Azure SQL database.
Created complex stored procedures that inserts and updates data in different databases.
Created scripts that handles data consistency and error handling across different environments.
Implemented Copy activity, Custom Azure Data Factory Pipeline Activities.
Primarily involved in Data Migration using SQL, Azure SQL, Azure Storage, and Azure Data Factory.
Validated the converted data in both CWS and Guidewire Claim Center.
Environment:
Azure Data Factory, Azure Data Lake, Azure SQL, Python, MS SQL Server 2016, SSIS, Oracle, MS Excel, TFS, Visual Studio 2014, Guidewire Policy Center, Claim Center, Red Gate SQL Search, Red Gate Data Compare.
Perr&Knight, Remote, TX, (Sr ETL Developer - Team Lead)
June 2016–Feb 2019
Responsible for gathering requirements from End user and delivering the same to the team by conducting team meetings.
Involved in designing and created the database by understanding the Clients business entities and categorizing them to different tables.
Automated data validation and reconciliation processes to ensure accuracy and completeness in reports submitted to regulatory bodies.
Collaborated with actuarial, finance, and compliance teams to align data models with regulatory reporting needs.
Worked on multiple ClaimCenter and PolicyCenter end to end conversion projects and in different LOB’s such as Workers Compensation, General Liability, Crime, Inland Marine, Fire & Allied and Commercial Auto etc.
Involved in design and development of Intermediate Schema and reconciliation framework for ClaimCenter, estimated the Losses and calculated the paid losses and Outstanding Losses based on claim rules in the ClaimCenter.
Involved in understanding different Legacy systems and migrate the legacy data successfully as per the business requirement and mainly focused on Guidewire Conversion stream.
Created SSIS Packages, Informatica Mappings to load data from Source to destination database and generate a text file and email the same to the users.
Used Look-up transformation and Slowly changing Dimension transformation to insert the new records and update existing records with no primary key.
Developed Mappings in Informatica using transformations like connected/unconnected lookup, Router, Aggregator, Expression, Stored procedure, Sequence generator, Update strategy, Filter, and Joiners etc. depending upon requirement.
Used the Informatica Designer to create complex Mappings and Mapplets.
Used Informatica Scheduler to schedule the workflows for everyday and on demand load process and worked with Repository manager for creating new folders and giving securities to the folders
Used Informatica as an ETL tool to extract data from source systems to Target system, Source Systems are mainly flat files, TXT files, relational tables and with the Oracle as target.
Created Ad-hoc reports using Report Builder for the client to generate their own reports depending on their requirement, deployed reports onto server and provided access to the users as needed.
Validated SQL Database data against Snowflake Cloud Database using Python Scripts through Jupyter Notebook using Anaconda.
Environment:
MS SQL Server 2016/2012, Snowflake Cloud Computing, SSIS, SSRS, Oracle, Informatica PowerCenter 9.X, MS Excel, TFS, Visual Studio 2014, Guidewire PolicyCenter, ClaimCenter.
Questar Assessments, Apple Valley, MN, (Sr BI Developer)
Oct 2015– June 2016
Created Stored Procedures to import data from csv and text files tables using .fmt files.
Used Merge to perform insert, update and delete operations on a target table based on the results of a join with a source table.
Created various data models in SSAS tabular model and used them to create Power BI templates.
Created dashboards using Power BI to serve different business users.
Used custom Power BI gallery to import visualizations for dashboard purposes.
Published Power BI dashboards to cloud and SSRS on premises Power BI server for user rendering.
Developed Tableau visualizations and dashboards using Tableau Desktop.
Uploaded the created documents to Confluence so that the complete team can have access to it.
Environment:
MS SQL Server 2012/2008R2/2005, SSIS, SSRS, SSAS, Tableau, Power BI, MS Excel, JIRA, SVN, GIT, Bit -Bucket, Confluence, Visual Studio 2012, Red Gate SQL Search, Red Gate Data Compare.
Madison Area Technical College, Madison, WI, (Sr BI Developer)
Jan 15– Oct 15
Worked as a developer in creating complex Stored Procedures, SSIS packages, triggers, cursors, tables, and views and other SQL joins and statements for applications.
Migrated data from various sources including XML Files, flat files, MySQL to SQL Server 2012 by using SSIS and Cognos Data Manager.
Generated database SQL scripts and deployed databases including installation and configuration.
Resolved any deadlocks issues with the databases / servers on a real-time basis.
Responsible for ongoing maintenance and change management to existing reports and optimize report performance.
Created Composite Reports using Workday Reporting, hands on experience with Power BI.
Environment:
MS SQL Server 2012, SSIS, SSRS, MS Excel, Microsoft Power BI, Cognos Data Manager, Workday, Service Now.
United Airlines, Houston, TX, (Sr SQL BI Developer/ DBA)
Jun 2013– Dec 2014
Designed the methods to measure the performance of the SSIS packages and worked on enhancing the performance of SSIS packages.
Used SQL Profiler for troubleshooting, monitoring, and optimization of SQL Server and non-production database code as well as T-SQL code from developers and QA.
Used various transformations like Multicast, Conditional split, OLEDB Command, Script Component and Lookup while loading the data into Destination.
Used Control Flow Tasks like For Loop Container, For Each Loop Container, Sequential Container, Execute Task, Email Task, and Data Flow Task.
Identified the database tables for defining the queries for the reports. Defined report layouts for formatting the report design as per the need. Identified and defined the Datasets for report generation.
Created an ad-hoc reports using SSRS Report Builder.
Used SQL Profiler to Solve Dead Locks/Long Running Queries.
Analyzed VB Script in the existing Crystal Reports and converted them to SSRS.
Created an ad-hoc reports using SSRS Report Builder.
Environment:
MS SQL Server 2005/2008R2/2012, SSIS, SSRS, Crystal Reports 8.5, Team Foundation Server 2010, MS Excel
Dish Networks, Englewood, CO, (SQL BI Developer)
Jan 2013– June2013
Extensively worked on Migrating Teradata into SQL Server Data by using SSIS.
Transferred the MS Excel Sheet Reports into SSRS based reports by migration the data by using SSIS packages and then Use views and tables and stored procedures to develop new reports.
Environment:
MS SQL Server 2008R2/2008, SSIS, SSRS, Team Foundation Server, Windows Server 2005, Teradata, MS Excel
AMC Cookware India Pvt Ltd, Hyd, INDIA, (MS SQL Server Developer)
Jan 09– Aug 10
Worked on optimizing the query by creating various clustered, non-clustered indexes and indexed views.
Environment:
MS SQL Server 2005/2000, MS Access 2000, MS SQL Analysis Services 2005, MS Excel, Visual Studio 2005.