Post Job Free
Sign in

Data Engineer Azure

Location:
Atlanta, GA
Posted:
April 01, 2025

Contact this candidate

Resume:

Aruru Chandra Sekhar

Mobile: +1-469-***-****

Email: *******.****@*****.***

Location: Texas

LinkedIn: https://www.linkedin.com/in/chandrasekhararuru4699224076/

Professional Summary:

Lead Data Engineer & Analyst Microsoft CERTIFIED ASSOCIATE Over 17 years of experience in the IT Industry worked on different SDLC in all phases Data Engineering & Analytics using Microsoft Technologies and Technical Expertise on Business Intelligence, Data Warehousing, Data Analysis, Data Modelling, ETL tools, Product Development and interactive reports. Complete understanding in all phases of Software Development Life Cycle (SDLC). 4+ years of leadership experience as a team lead, mentoring and managing resources Proficient in Azure cloud technologies like (Azure Data Factory, SQL Database, Synapse Analytics, Data Lake Storage, Power BI and DAX). 3+ years of hands-on experience with Big Data technologies (Databricks, Spark, Python, PySpark, Kafka). Good knowledge of Data Design Patterns Architecture like Relation Database (OLTP/ODS), Data Warehouse (ETL OLAP), Modern Data Warehouse (Data Lake ELT) and Lakehouse Architecture (Delta Lake ELT) Skilled in data architecture, ETL frameworks, and implementing modern data platforms like Medallion Data Lake, Snowflake, and Delta Lake. Experience in gathering and understanding business requirements and convert into conceptual model, entities and attributes, data dictionary documents and data mapping documents Experience in working with data modelling tool Erwin & Power Designer for design of logical and physical data model. 12 years of experience in health care domain worked in area/modules like provider, bundle payments, Obama health care claims submissions and revenue claims management (RCM)

Technical Skills:

Operating systems : Windows 9x/NT/2000/XP, UNIX.

Database (SQL/NO SQL) : Oracle 11g, MS SQL Server 2017, Spark SQL, Mango DB,

Azure Cosmos DB, KQL and Teradata V2R13.

Data Warehouse : Azure Synapse Analytics, BigQuery, Amazon Redshift, Snowflake

ETL workflow/Tools : SSIS, Informatica Power Center 8.6, Data Build Tool (DBT) & Talend

Azure Cloud Services : Azure SQL Database (PaaS), Azure Storages Blobs, Azure Data Factory(ADF), Azure Synapse Analytics (Azure Data Warehouse), Azure Data lake(ADLS), Onelake, Azure HDInsight’s, HIVE, Azure Analysis Services and Azure data bricks.

Data Architecture Modern : Lakehouse, Delta Lake, One Lake & Medallion Data Lake

Cloud Storages : Azure blob, ADLS Gen2, S3 and Cloud Storage.

Languages known : SQL, PLSQL, Python 3.9

Reporting Tools : Power BI, Tableau 10.5 and SSRS.

Defect Tracking Tools : Rally and ALM 11.

Certifications:

Microsoft Certified: Fabric Data Engineer Associate.

Microsoft Certified: Azure Data Engineer Associate.

Certified in 1Z0-007 Introduction to oracle 9i: SQL from Oracle.

Work Experience:

Working as Data Engineering Consultant in Optum Global Solutions (India), Hyderabad from March 2013 to March 2025.

Worked as Senior Engineer in Collabera, Hyderabad from 2 April 2012 to 25 February 2013.

Worked as Associate Test Lead at Semantic space Technologies Ltd. (SST), Hyderabad from 04 October 2006 to 31 October 2011.

Academic Profile:

Master of Computer Applications from University of Madras, Chennai, Tamilnadu, India. Passed out year December 2009

Bachelor of Commerce –Osmania University, Hyderabad, Andhra Pradesh, India. Passed out year July 1999

Diploma in Advance Software Technology from CMC-ATC, Hyderabad, India. Passed out year August 2000

Employer: United Health Group International Services Pvt Ltd (UHG)

Project title: Life Cycle of Analytics (LCA)

Duration: May 2022 to till date

Client: Excela Health, Boulder and SSM

Domain: Health Care

Software Used: Azure SQL, MS-Fabric, Azure Databricks, Python and Pyspark.

Role: Data Engineer Consultant

Life Cycle of Analytics (LCA): daily will receive raw files with different format type like csv and delimited from client about 50 files for each client will validate file format, file validation, business rules in three level and final load data into data bricks for further analysis and reporting purpose.

Responsibilities:

Understanding and analysing business requirements in PI & sprint grooming

Daily attending stand-up and scrum meetings updating development status and any risk.

Skilled in data architecture, ETL frameworks, and implementing modern data platforms like Medallion Data Lake, Snowflake, and Delta Lake

Data Ingestion importing large volumes of data from various sources into the Fabric environment, often using data pipelines or connectors, dataflow and copy statement.

Data Transformation business requirement and logics are created with the help of fabric Dataflow Gen2, Data Pipeline and Power Query

Data storage in Fabric centralizes data storage with its unified architecture Lakehouse and Onelake.

Create Recurring Scheduling, Event-Driven Scheduling and On-Demand Scheduling Automating batch jobs to run at specific intervals, ensuring timely data updates

Created various notebooks to process bigdata and schedule jobs and workflows on daily, weekly and monthly bases. Created one time configure of interactive spark clusters and job clusters

SQL query optimize ensure data quality check and data validation to meet the business requirement

Written SQL quires to Experience in writing advance PL/SQL or SQL-T Stored Procedures, Functions

Implemented Azure Data Factory (ADF) Pipelines created fresh or initial loads, increment or delta loads and SCD to Orchestrating the ETL Processing on Azure.

Create one time configure of interactive spark clusters, job clusters and pool to reduce cluster spin time.

Experience in data ingestion, preparation, integration and operationalization techniques in optimally addressing the data requirements

Project title: HIM Power BI Migration

Duration: Jan 2021 to Apr 2022

Client: Excela Health, DCH Health System and Mayo Clinic

Domain: Health Care

Software Used: Azure SQL, ASDL, ADF, Synapse, AAS, DAX and Power BI.

Role: Data Engineer Consultant

HIM Power BI Migration: converting exist tableau dashboards to Power BI dashboard

As part of that setting up on premise similar infrastructure in Azure could and also implement new snowflake schema model with Row Level Security (RLS) and integrated Power BI Reports with API’s.

Responsibilities:

Understanding and analysing business requirements in PI & sprint grooming

Daily attending stand-up and scrum meetings updating development status and any risk.

Active participate in data model design and changes are feed back

Based on product backlog user stories creating pipelines, initial loads, increment loads, Data validation, creating user store producers, functions and creating database objects DB scripts and deployments in lower environments.

Hands on experience on installing the selfhosted Integration runtime to migrate data from on premise to cloud.

Hands on experience on creating the generic datasets and linked services for the different sources.

Having good working experience of Storages like Azure Data Lake Storage and Blob Storage

Creating Power BI visualizations to achieve different reports and dash boards

Active participate in defect triage meeting and solution

Experience with relational databases such as MySQL and SQL Server.

developed data strategy for data movement and data integration

Drove data analysis effort for data movement and built ETL designs.

defined level of grain required and star schema concept model for data modeler

developed data strategy for warehouse data movement and data integration strategies for outlook

created concept model for counterparty approval and reporting

created data strategies for ODS and warehouse

created concept and logical tier-1 data model for pahse-1 reference data distribution

Project title: Computer Assisted Coding (CAC)

Duration: Feb 2018 to Dec 2020 date

Client: Excela Health, DCH Health System and Mayo Clinic

Domain: Health Care

Software Used: Azure SQL, ASDL, ADF, Synapse, AAS and Power BI.

Role: BI developer

Computer Assisted Coding: Computer Assisted Coding (CAC), workflow, coder functionality and Clinical Document Improvement (CDI) data into a single reporting product. User Provisioning utilizes internal legacy Auth Admin application.

Computer Assisted Coding (CAC), workflow, coder functionality and Clinical Document Improvement (CDI) data into a single reporting product.

Responsibilities:

Understanding and analysing business requirements in PI & sprint grooming

Daily attending stand-up and scrum meetings updating development status and any risk.

Active participate in data model design and changes are feed back

Based on product backlog user stories creating pipelines, initial loads, increment loads, Data validation, creating user store producers, functions and creating database objects DB scripts and deployments in lower environments.

Creating Power BI visualizations to achieve different reports and dash boards

Active participate in defect triage meeting and solution.

Project title: Veterans Affairs Community Care Network (VA CCN)

Duration: Apr 2017 to Dec 2017

Client: Cleveland Clinic & Presence

Domain: Health Care

Software Used: SSIS, SQL Server 2012, Rally and Tableau10.2.

Role: BI developer

Veterans Affairs Community Care Network: The goal of VA CCN Program is to achieve technology implementation readiness before the start of health care delivery to support the US Department of Veterans Affairs (VA) plans to secure a third party to provide a Community Care Network consisting of providers for medical and surgical inpatient and outpatient services.

Data Repository (Orbit BI) Scope: the Data Repository scope covers requirements for the collection, storage, and distribution of data for the VA. Data is generated and consumed both internally and externally.

Responsibilities:

Understanding business requirements in PI & sprint grooming, daily attending stand-up and scrum meetings.

Based on product backlog user stories creating SSIS Packages, Data validation, Creating user store producers, functions and creating database objects DB scripts and deployments in lower environments.

Active participate in defect triage meeting and solution.

Data Extraction: Pulling data from various sources such as databases, flat files, or APIs.

Data Transformation: Cleaning, aggregating, and converting data into a usable format using SSIS transformations like Derived Column, Lookup, or Conditional Split.

Data Loading: Inserting transformed data into target destinations like data warehouses or DataMart.

Monitoring and Debugging: Checking logs and fixing errors in SSIS packages to ensure smooth execution.

Scheduling Jobs: Automating package execution using SQL Server Agent or other scheduling tools.

Performance Optimization: Tweaking SSIS packages for faster execution and efficient resource usage.



Contact this candidate