Post Job Free
Sign in

Data Engineer Azure

Location:
Somerset, NJ
Posted:
January 26, 2025

Contact this candidate

Resume:

MALHAR PATEL

Azure Data Engineer Snowflake Warehousing Specialist

Email - ***********@*****.*** Number - 732-***-****

PROFESSIONAL SUMMARY

Azure Data Engineer with 6 years of experience delivering scalable data engineering solutions in the healthcare domain. Specializes in Snowflake Data Warehousing, Azure cloud services, and real-time data streaming. Adept at building ETL pipelines, optimizing big data architectures, and enabling data-driven healthcare decision-making. Proficient in designing secure and compliant solutions aligned with industry regulations such as HIPAA.

Expertise in Snowflake Data Warehouse for analytics and operational intelligence.

Proficient in building ETL/ELT pipelines using Azure Data Factory (ADF).

Skilled in managing cloud storage solutions with Azure ADLS Gen2.

Experienced in processing and analyzing large datasets using Azure Databricks and Apache Spark.

Competent in implementing real-time data streaming solutions with Apache Kafka.

Hands-on experience with CI/CD pipeline management using GitHub and Azure DevOps.

Advanced skills in Python and SQL for data transformation and automation.

Adept at creating interactive dashboards using Power BI and Tableau.

Strong knowledge of data governance, compliance, and security protocols in healthcare.

Skilled in Agile methodology and cross-functional collaboration to meet project deadlines.

TECHNICAL SKILLS

Cloud Platforms

Azure (Data Factory, Databricks, ADLS Gen2), Snowflake

Data Integration Tools

Azure Data Factory, Snowflake Pipelines

Big Data Technologies

Apache Spark, Azure Databricks

Programming Languages

Python, SQL, Scala

Streaming Platforms

Apache Kafka

Version Control Tools

GitHub, Azure DevOps

BI & Visualization

Power BI, Tableau

Database Systems

Snowflake, SQL Server

Methodologies

Agile, Scrum

PROFESSIONAL EXPERIENCE

Snowflake Developer

CareFirst BlueCross BlueShield Baltimore, MD (March 2022 – Present)

Developed a scalable Snowflake-based data warehouse to enable enterprise-wide reporting and improve analytics for provider performance and claims processing.

Key Responsibilities:

Designed and implemented a Snowflake Data Warehouse to centralize data across claims, providers, and membership records.

Created and optimized complex SQL queries and Snowflake procedures for data transformation and reporting needs.

Configured Snowflake Streams and Tasks to enable real-time data replication and incremental data loading.

Built automated ETL/ELT pipelines using Azure Data Factory to load data into Snowflake from various source systems.

Performed query performance tuning and Snowflake optimization to reduce execution times by 40%.

Designed secure data-sharing mechanisms within Snowflake for third-party vendors.

Collaborated with stakeholders to define data models and schema designs for analytic use cases.

Ensured data security compliance with HIPAA and CMS standards in the data engineering processes.

Deployed Power BI dashboards connected to Snowflake for visualization of KPIs such as claims cycle time and network utilization.

Conducted training sessions for internal teams on Snowflake best practices and SQL optimization techniques.

Environment:

Snowflake, Azure Data Factory, SQL, Python, Power BI, GitHub.

Azure Data Engineer

Oscar Health New York, NY (July 2021 – February 2022)

Project Scope:

Developed a cloud-based data platform to streamline payer operations, enhance claims processing efficiency, and support advanced analytics for member insights.

Key Responsibilities:

Architected a Snowflake Data Warehouse to centralize and manage claims and enrollment data for improved operational analytics.

Built and optimized ETL pipelines using Azure Data Factory, ensuring data accuracy across member and provider records.

Configured Azure ADLS Gen2 for secure, high-performance data storage, reducing data retrieval latency by 30%.

Implemented real-time streaming pipelines with Apache Kafka to process live updates for claim adjudication.

Leveraged Databricks and Apache Spark for advanced analytics and predictive modeling, supporting fraud detection initiatives.

Automated CI/CD workflows using Azure DevOps, enabling faster deployments and minimizing downtime.

Performed query optimization in Snowflake, achieving a 25% reduction in query execution times.

Designed and maintained Power BI dashboards to monitor claims turnaround and payer network performance.

Ensured compliance with HIPAA and CMS regulations in all data workflows.

Conducted stakeholder training sessions on Snowflake and data visualization tools.

Environment: Azure Data Factory, ADLS Gen2, Snowflake, Apache Kafka, Databricks, Apache Spark, Azure DevOps, GitHub, Python, SQL, Power BI.

Data Engineer

Blue Shield of California Oakland, CA (May 2019 – June 2021)

Migrated legacy on-premises data warehouses to the cloud, enabling enhanced reporting capabilities and improving patient care analytics.

Key Responsibilities:

Led the migration of healthcare data warehouses to Azure Synapse, optimizing scalability and cost-efficiency.

Designed and orchestrated complex ETL pipelines in Azure Data Factory to integrate member, provider, and claims data.

Established and managed Azure ADLS Gen2 as a unified data lake for structured and unstructured healthcare datasets.

Implemented Apache Kafka to enable real-time updates in operational dashboards for member services.

Deployed Azure Databricks for data transformation and large-scale data processing, improving data aggregation speed by 40%.

Maintained CI/CD pipelines using Azure DevOps, ensuring consistent code quality and automated deployments.

Collaborated with analytics teams to deliver Tableau dashboards, providing insights into provider performance and cost management.

Wrote Python scripts to automate data validation and auditing processes.

Created documentation and conducted knowledge-sharing sessions for the internal data engineering team.

Enhanced data governance frameworks to comply with HIPAA and Blue Shield’s internal security protocols.

Environment:

Azure Synpase, Azure Data Factory, ADLS Gen2, Apache Kafka, Databricks, Tableau, Python, SQL Server, Azure DevOps.

Education

Bachelor of Science in Information Technology from

New Jersey Institute of Technology



Contact this candidate