Post Job Free
Sign in

Data Engineer Power Bi

Location:
Leander, TX
Posted:
November 26, 2024

Contact this candidate

Resume:

Praveen Rao Gundavarapu

Email: **********@*****.*** - LinkedIn

Phone: 818-***-****

PROFESSIONAL SUMMARY

Senior Lead Data Engineer with 16 + years’ experience in design and building data and Analytics applications & pipelines, including develop & deploy Enterprise level ETL & BI solutions, snowflake and Data ware housing. Work with Bigdata data bricks for cloud applications. developed power BI for enterprise level data driven dashboards and reports

SKILLS

Power shell & shell Programming &T-SQL

Spark RDD, data frame, spark different contexts, UDF & notebook

Azure data factory, data lake, storage

Linux and windows Server Platforms

Hive, beeline, Ambari, Azure HDInsight, Data Lake, Blob storage,

Data Bricks, pyspark, json, Parquet

Power BI, ADF, U-SQL

Python, NumPy & matplotlib

Azure DevOps, Git, YAML, Jenkins

Sales Force marketing cloud, data extracts and campaign data

Snowflake DWH, variant stream and JSON load

Kafka connecter and Topic creation and troubleshoot data load for azure storage

Data warehouse modelling, SQL DW (Azure synopsis), poly base

Build releases & integration

Compute and scale application processing

Azure Functions & EventHub

Integrate snow pipe with Azure, Data mapping and stage& integration

Build AS cubes and deploy using VSTS as well as CICD

EDUCATION

Master of Science in Engineering in Mechatronics, RWTH Aachen Germany

Bachelor of Science and Engineering in Mechanical JNTU Anantapur

CERTIFICATIONs

SQL Server 2008 MCITP Certificate

Hadoop 2.x Administration

Python for Data science (MIT Institute/online)

Azure Databricks(spark) (edX)

PROFESSIONAL EXPERIENCE

NOVEMBER 22 – PRESENT

SR. DATA ENGINEER, EVERESTRE Insurance Remote, New Jersy

Design azure applications and pipelines for global insurance space for analyzing claims data using SQL,

ADF, spark, Databricks and snowflake build dashboards to provide insights for analyzing claims policy

and forecast of the target budget strategy

Design data modelling and schema development for claims approval, Cost of claims data, Continuous Integration and Continuous Delivery (CI/CD), Test Automation, Infrastructure as Code, Secure Coding Practices

Use of kafka and other sources systems ( salesforce cloud platform) api ingestion . load data using synapse and warehouse to connect PBI reports

Ingest using confluence kafka data of survey and reviews to snowflake DB and flatten the data

Load data to stage to curated and build the bi warehouse for loss impact of policies

Build fact and key dimensions for financial reporting needs

Languages and Technologies: Azure Data Factory, Snowflake, Web API, GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, snowflake, kafka, sales force cloud, U-SQL, logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python

NOVEMBER 21 – NOVEMBER 22

SR. DATA ENGINEER, Microsoft Vendor contract, Redmond WA

Design azure applications and pipelines for using flow for analyzing payment data using cosmos, spark, Databricks and snowflake build dashboards to provide insights for analyzing payment billing charge back and provider history

. Design data modelling and schema development for payment approval, Cost of payment data, WorldPay/account update data cleansing and testing for various source Cloud services components like, storage, ADF framework, function app logic app

As a senior lead developer, I am responsible for design and development of the payment applications for C+ E finance group accessing data engineering and data analytics project related to GPS/GF’S payment streams PI and AU data

CLAMOR/DARSY services: - Preparing Asimov data for darsy portal submissions for third party apps development in cloud services. Metrics are stored in Azure and kusto database.

MARCH 21 – NOVEMBER 21

SR. DATA ENGINEER, Halliburton, Houston TX

Design azure applications and pipelines for large datasets using API, json format, for analyzing the fleet, fuel and pump using pyspark, Databricks and build dashboards to provide insights for analyzing fleet metrics

Build big data applications using Azure Data Factory, spark, python and Power BI which offers services to analyze and perform store performance. Use of Google Cloud Storage and BigQuery for extracts and API Designed data bricks application for updating the history of engine performance data using PySpark. Design data modelling and schema development for pe asset data, data cleansing and testing for various source Cloud services components like, storage, ADF logic app and spark used to ingest, load by apply transformation logics for data. Build PowerBI reports for analytics like for data marts & cubes. Build and connect kafka sources for dwh data load to snowflake database

Languages and Technologies: Azure Data Factory, Web API, VSTF/TFS, GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, U-SQL, logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python

JULY 18 – MARCH 21

SR. BIG DATA LEAD ENGINEER, Ahold-Delhaize USA. Quincy MA

As a senior lead developer, I am responsible for design and development of the applications and code deployments. build big data applications using Azure Data Factory, spark, U-SQL, python and Power shell and Web APIs which offers services to analyze and perform store performance. Use of Google Cloud Storage and BigQuery for extracts and API calls. Designed data bricks application using PySpark. Design data modelling and schema development, data cleansing and testing for various sources FTP, Oracle (EDW). Cloud services components like, storage, ADF and spark used to ingest, load by apply transformation logics for data. Build PowerBI reports for analytics like for data marts & cubes. Develop and integrate customer data MDM and marketing campaign with salesforce.

Languages and Technologies: Azure Data Factory, Web API, VSTF/TFS, GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, U-SQL,COSMOSDB logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python

JUN 13 – JULY 18

SR. SOFTWARE DEVELOPMENT ENGINEER, Microsoft Vendor Contract, Redmond WA

C & E finance: Building Data warehouse and data lake analytics: - Build MS finance analytics warehouse using Azure services like data lake storage, U-SQL and Power shell for MS product sales insights and KPI’s by feeding different areas of sources using Azure ADF and spark/python

Surface Device Telemetry &Analytics: - Preparing Asimov data for surface devices. Metrics are stored in Azure database. Data is processed and loaded using Cosmos scripts and Xflow configuration. Wrote DAX queries to derive datasets and build Power BI dashboards to visualize.

Web Analytics: - Office Max Team is assisting Office product website through web analytics. Such as how many page searches, legitimate page visits help to broaden the development of site. Data is tracked via fiddler events in COSMOS cluster. Develop and build ETL pipelines using SSIS and stored procedures to track history of those events. Once loaded to SQL data warehouse build Tabular cube for analyzing its trends

Device Configuration: Use Power shell scripts to re-home, includes install, configure windows and database tools for store devices to be used for RISK analytics application

MSG Campaign Tools: - Design and developed tool named OSCAR database, for tracking machines for vulnerability which uses Global Foundation

Technologies: SQL Server 2010, T-SQL, SSIS, Data warehouse, SSAS, SSRS, Power shell,

NOV 09 – FEB 13

SR. DATABASE ENGINEER, Disney- ABC TV Network, Burbank CA

ETL Architecture, Developing Source to target data mapping (STDM) document defining transformation logic for SAP data using SSIS and SQL stored procedures. Translated business logic to transformation logic to generate pseudo code for the ETL process

Disney Consumer Products: - Develop Online Product Approval (OPA) application database using to manage Product Lifecycle Management (PLM) system used to manage the approval lifecycle for Disney, Marvel, and ESPN consumer products

Technologies: SQL Server 2010, T-SQL, SSIS, Data warehouse, SSAS, SSRS, Power shell,

MAR 07 – OCT 09

DATABASE DEVELOPER, Department of Human Services (Deloitte), Pittsburgh PA

The Allegheny County Department of Human Services (DHS) Office of Children, Youth and Families (CYF) is the county's public office mandated by law to protect children from abuse and neglect and to ensure permanency for children. Supported for the KIDS application and integrated into existing CYF in the IT Management office

Languages and Technologies: SQL server 2005, T_SQL, stored procedures and Triggers, VSTS



Contact this candidate