Post Job Free
Sign in

Data Engineer

Location:
Lansdale, PA
Posted:
September 07, 2025

Contact this candidate

Resume:

LINKEDIN / PORTFOLIO

PROFESSIONAL SUMMARY

Data Engineer with 5+ years of experience building scalable data pipelines and transformation frameworks. Specialized in DBT, Snowflake, and PL/SQL, with expertise in designing modular SQL models, implementing testing and documentation, and optimizing warehouse performance. Strong background in data modeling (star and snowflake schemas), CI/CD pipelines, and Git-based workflows. Adept at collaborating with analysts, data scientists, and business teams to deliver reliable, high-quality data for analytics and reporting. SKILLS

Data Warehousing & Cloud Platforms: Snowflake (data modeling, clustering, performance tuning, security), Amazon Redshift, Azure SQL, SQL Server, Teradata, Hadoop

Data Transformation & Orchestration: DBT (modular SQL, testing, documentation, snapshots), Airflow, dbt Cloud, Prefect, AWS Glue, SSIS, Informatica

Programming & Scripting: PL/SQL (stored procedures, functions, packages), SQL, Python, PySpark, Scala Data Modeling: Star schema, Snowflake schema, normalization/denormalization, dimensional modeling, fact/dimension tables

Reporting & Visualization: Power BI, Tableau, SSRS Automation & CI/CD: GitHub, Jenkins, GitHub Actions, Control-M, Power Automate Data Quality & Governance: DBT tests, Informatica DQ, data validation frameworks, schema enforcement Other: Agile collaboration, requirements gathering, cross-functional communication EXPERIENCE

Data Engineer (Client - Nextech Systems – Orpine)

Dec 2023 – Current

Lansdale, PA

Designed and deployed DBT models on Snowflake to transform billing and practice management datasets into clean, analytics-ready schemas, reducing reporting latency by 40%.

Implemented modular SQL transformations in DBT with built-in testing, documentation, and versioning, creating reusable pipelines that improved developer productivity.

Developed complex PL/SQL stored procedures and functions to handle multi-step transformations for Salesforce and NetSuite data integrations, ensuring accuracy across multiple subsidiaries.

Optimized Snowflake queries through clustering keys, pruning strategies, and result caching, improving performance of dashboards accessed by 300+ stakeholders.

Collaborated with analysts to build star and snowflake schemas supporting financial and clinical reporting needs, ensuring data models were business-friendly and scalable.

Migrated legacy ETL workflows into DBT transformations, consolidating data pipelines and improving governance through centralized documentation.

Maintained CI/CD pipelines for DBT projects using GitHub Actions, ensuring automated testing, deployments, and rollback mechanisms.

NIKHIL

PASUPULETI

Lansdale, PA, 19446

510-***-****

*****************@*****.***

Designed audit scripts in Snowflake and PL/SQL to validate Salesforce-NetSuite syncs, enforcing governance and HIPAA compliance.

Automated error reporting and reconciliation tasks using DBT tests and Python scripts, reducing manual QA by 60%.

Partnered with architects to align Snowflake role-based access controls with governance standards.

Created DBT snapshots to track historical changes in practice data, enabling trend analysis.

Version-controlled SQL, DBT models, and PL/SQL code in GitHub, integrating with Jenkins for CI/CD automation.

Delivered real-time Power BI dashboards powered by Snowflake DBT models, enabling leadership to track billing KPIs.

Conducted code reviews and peer sessions for DBT model design, ensuring adherence to best practices.

Documented data pipelines, transformations, and dependencies directly in DBT for end-to-end transparency.

Trained junior engineers in DBT modular design, Snowflake performance tuning, and Git workflows. Tools: Snowflake, DBT, PL/SQL, SQL Server, Python, Salesforce, NetSuite, GitHub, Jenkins, Power BI Data Engineer (Anthem - Blue Cross And Blue Shield) February 2023 – Oct 2023

Atlanta, GA

Built scalable DBT models on Snowflake to support claims analytics, designing star schemas that improved query performance by 30%.

Wrote advanced PL/SQL procedures and functions for claims ingestion and cleansing, automating exception handling.

Migrated 20+ TB of data from Teradata/Hadoop into Snowflake, applying DBT transformations to standardize and document the pipelines.

Optimized Snowflake micro-partitioning through clustering and statistics collection, reducing warehouse costs by 20%.

Integrated Airflow with DBT to orchestrate transformations, ensuring automated runs and timely delivery of datasets.

Applied DBT tests for data validation, implementing schema, unique, and referential integrity checks.

Maintained CI/CD workflows for DBT using GitHub and Jenkins, ensuring peer-reviewed, version-controlled deployments.

Collaborated with data scientists and BI teams to build DBT models powering Power BI dashboards for provider and SLA monitoring.

Designed PL/SQL audit frameworks to validate data consistency before and after migration into Snowflake.

Developed DBT snapshots to track historical claims changes for longitudinal analysis.

Documented DBT models with descriptions, lineage graphs, and dependency mappings.

Partnered with business stakeholders to align Snowflake access and DBT project structure with governance needs.

Trained analysts to query DBT-exposed tables in Snowflake, empowering self-service analytics.

Automated notifications with DBT and Airflow for SLA breaches and pipeline delays.

Supported continuous improvement by tuning SQL transformations and applying DBT macros for reusability.

Delivered real-time Power BI dashboards powered by Snowflake DBT models, enabling leadership to track billing KPIs.

Tools: Snowflake, DBT, PL/SQL, Teradata, Hadoop, PySpark, Airflow, GitHub, Jenkins, Power BI, Informatica Jr. Data Engineer (SAGE IT Inc)

Aug 2022 – Dec 2022

Frisco, TX

Migrated 3TB+ of Amazon Redshift data into Snowflake, implementing DBT transformations to normalize datasets and prepare for BI consumption.

Wrote PL/SQL functions and packages to perform cleansing and enrichment of Redshift data prior to Snowflake ingestion.

Designed DBT modular SQL models for operational KPIs, including incremental load strategies, snapshots, and tests.

Converted Hive SQL queries into DBT models with improved maintainability and governance.

Integrated DBT projects with Prefect for orchestration, automating scheduled transformations.

Built Power BI dashboards powered by DBT-transformed Snowflake datasets for ingestion monitoring and error tracking.

Validated ingested data using DBT tests, Athena SQL queries, and Snowflake stored procedures, ensuring schema correctness.

Developed CI/CD workflows for DBT projects using GitHub Actions, enabling automated testing and deployment.

Applied Snowflake query tuning strategies, including clustering and warehouse optimization.

Designed DBT documentation and lineage graphs for cross-team visibility.

Configured CloudWatch alerts for pipeline monitoring, integrated with DBT notifications.

Documented Redshift-to-Snowflake migration flows and DBT model dependencies.

Supported BI team in migrating legacy reports into Snowflake/DBT-powered models.

Implemented incremental DBT models to reduce warehouse costs by avoiding full reloads.

Conducted peer reviews of DBT code to enforce best practices.

Trained team members in Snowflake optimization and DBT modular design principles. Tools: Snowflake, DBT, PL/SQL, Redshift, Spark, Prefect, GitHub, Python, Athena, Power BI, CloudWatch Peoplesoft Data Analyst (Cleveland State University) Jan 2021 – Dec 2021

Cleveland, OH

Extracted and cleaned large datasets from SQL databases to generate business insights and ensure data accuracy.

Developed Excel-based reports and dashboards for tracking key performance indicators (KPIs) and operational metrics.

Wrote SQL queries to analyze trends, identify inconsistencies, and support business decision-making.

Created basic data visualizations in Excel and Power BI to present findings to stakeholders.

Assisted in building automated data pipelines using scheduled SQL queries and Excel macros for reporting efficiency.

Documented data processes and created user guides for stakeholders to understand the reporting workflows.

Tools: Oracle PeopleSoft, SQL, Excel, Power BI, Tableau (basic), Google Sheets EDUCATION

Masters in Computer science (M.S) (Cleveland State University), Cleveland, OH Bachelors in Computer science (B.S) (Jawaharlal Nehru Technological University), India



Contact this candidate