Praveen Rao Gundavarapu
Email: **********@*****.*** - LinkedIn
Phone: 818-***-****
PROFESSIONAL SUMMARY
Senior Lead Data Engineer with 16 + years’ experience in design and building data and Analytics applications & pipelines, including develop & deploy Enterprise level ETL & BI solutions, snowflake and Data ware housing. Work with Bigdata data bricks for cloud applications. developed power BI for enterprise level data driven dashboards and reports
SKILLS
Power shell & shell Programming &T-SQL
Spark RDD, data frame, spark different contexts, UDF & notebook
Azure data factory, data lake, storage
Linux and windows Server Platforms
Hive, beeline, Ambari, Azure HDInsight, Data Lake, Blob storage,
Data Bricks, pyspark, json, Parquet
Power BI, ADF, U-SQL
Python, NumPy & matplotlib
Azure DevOps, Git, YAML, Jenkins
Sales Force marketing cloud, data extracts and campaign data
Snowflake DWH, variant stream and JSON load
Kafka connecter and Topic creation and troubleshoot data load for azure storage
Data warehouse modelling, SQL DW (Azure synopsis), poly base
Build releases & integration
Compute and scale application processing
Azure Functions & EventHub
Integrate snow pipe with Azure, Data mapping and stage& integration
Build AS cubes and deploy using VSTS as well as CICD
EDUCATION
Master of Science in Engineering in Mechatronics, RWTH Aachen Germany
Bachelor of Science and Engineering in Mechanical JNTU Anantapur
CERTIFICATIONs
SQL Server 2008 MCITP Certificate
Hadoop 2.x Administration
Python for Data science (MIT Institute/online)
Azure Databricks(spark) (edX)
PROFESSIONAL EXPERIENCE
NOVEMBER 22 – PRESENT
SR. DATA ENGINEER, EVERESTRE Insurance Remote, New Jersy
Design azure applications and pipelines for global insurance space for analyzing claims data using SQL,
ADF, spark, Databricks and snowflake build dashboards to provide insights for analyzing claims policy
and forecast of the target budget strategy
Design data modelling and schema development for claims approval, Cost of claims data, Continuous Integration and Continuous Delivery (CI/CD), Test Automation, Infrastructure as Code, Secure Coding Practices
Use of kafka and other sources systems ( salesforce cloud platform) api ingestion . load data using synapse and warehouse to connect PBI reports
Ingest using confluence kafka data of survey and reviews to snowflake DB and flatten the data
Load data to stage to curated and build the bi warehouse for loss impact of policies
Build fact and key dimensions for financial reporting needs
Languages and Technologies: Azure Data Factory, Snowflake, Web API, GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, snowflake, kafka, sales force cloud, U-SQL, logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python
NOVEMBER 21 – NOVEMBER 22
SR. DATA ENGINEER, Microsoft Vendor contract, Redmond WA
Design azure applications and pipelines for using flow for analyzing payment data using cosmos, spark, Databricks and snowflake build dashboards to provide insights for analyzing payment billing charge back and provider history
. Design data modelling and schema development for payment approval, Cost of payment data, WorldPay/account update data cleansing and testing for various source Cloud services components like, storage, ADF framework, function app logic app
As a senior lead developer, I am responsible for design and development of the payment applications for C+ E finance group accessing data engineering and data analytics project related to GPS/GF’S payment streams PI and AU data
CLAMOR/DARSY services: - Preparing Asimov data for darsy portal submissions for third party apps development in cloud services. Metrics are stored in Azure and kusto database.
MARCH 21 – NOVEMBER 21
SR. DATA ENGINEER, Halliburton, Houston TX
Design azure applications and pipelines for large datasets using API, json format, for analyzing the fleet, fuel and pump using pyspark, Databricks and build dashboards to provide insights for analyzing fleet metrics
Build big data applications using Azure Data Factory, spark, python and Power BI which offers services to analyze and perform store performance. Use of Google Cloud Storage and BigQuery for extracts and API Designed data bricks application for updating the history of engine performance data using PySpark. Design data modelling and schema development for pe asset data, data cleansing and testing for various source Cloud services components like, storage, ADF logic app and spark used to ingest, load by apply transformation logics for data. Build PowerBI reports for analytics like for data marts & cubes. Build and connect kafka sources for dwh data load to snowflake database
Languages and Technologies: Azure Data Factory, Web API, VSTF/TFS, GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, U-SQL, logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python
JULY 18 – MARCH 21
SR. BIG DATA LEAD ENGINEER, Ahold-Delhaize USA. Quincy MA
As a senior lead developer, I am responsible for design and development of the applications and code deployments. build big data applications using Azure Data Factory, spark, U-SQL, python and Power shell and Web APIs which offers services to analyze and perform store performance. Use of Google Cloud Storage and BigQuery for extracts and API calls. Designed data bricks application using PySpark. Design data modelling and schema development, data cleansing and testing for various sources FTP, Oracle (EDW). Cloud services components like, storage, ADF and spark used to ingest, load by apply transformation logics for data. Build PowerBI reports for analytics like for data marts & cubes. Develop and integrate customer data MDM and marketing campaign with salesforce.
Languages and Technologies: Azure Data Factory, Web API, VSTF/TFS, GIT, data bricks, Azure Functions, Azure SQL Polybase, BLOB, Data lake, U-SQL,COSMOSDB logic apps, EventHub, data bricks, Azure DevOPS, Jenkins, Azure key vaults and python
JUN 13 – JULY 18
SR. SOFTWARE DEVELOPMENT ENGINEER, Microsoft Vendor Contract, Redmond WA
C & E finance: Building Data warehouse and data lake analytics: - Build MS finance analytics warehouse using Azure services like data lake storage, U-SQL and Power shell for MS product sales insights and KPI’s by feeding different areas of sources using Azure ADF and spark/python
Surface Device Telemetry &Analytics: - Preparing Asimov data for surface devices. Metrics are stored in Azure database. Data is processed and loaded using Cosmos scripts and Xflow configuration. Wrote DAX queries to derive datasets and build Power BI dashboards to visualize.
Web Analytics: - Office Max Team is assisting Office product website through web analytics. Such as how many page searches, legitimate page visits help to broaden the development of site. Data is tracked via fiddler events in COSMOS cluster. Develop and build ETL pipelines using SSIS and stored procedures to track history of those events. Once loaded to SQL data warehouse build Tabular cube for analyzing its trends
Device Configuration: Use Power shell scripts to re-home, includes install, configure windows and database tools for store devices to be used for RISK analytics application
MSG Campaign Tools: - Design and developed tool named OSCAR database, for tracking machines for vulnerability which uses Global Foundation
Technologies: SQL Server 2010, T-SQL, SSIS, Data warehouse, SSAS, SSRS, Power shell,
NOV 09 – FEB 13
SR. DATABASE ENGINEER, Disney- ABC TV Network, Burbank CA
ETL Architecture, Developing Source to target data mapping (STDM) document defining transformation logic for SAP data using SSIS and SQL stored procedures. Translated business logic to transformation logic to generate pseudo code for the ETL process
Disney Consumer Products: - Develop Online Product Approval (OPA) application database using to manage Product Lifecycle Management (PLM) system used to manage the approval lifecycle for Disney, Marvel, and ESPN consumer products
Technologies: SQL Server 2010, T-SQL, SSIS, Data warehouse, SSAS, SSRS, Power shell,
MAR 07 – OCT 09
DATABASE DEVELOPER, Department of Human Services (Deloitte), Pittsburgh PA
The Allegheny County Department of Human Services (DHS) Office of Children, Youth and Families (CYF) is the county's public office mandated by law to protect children from abuse and neglect and to ensure permanency for children. Supported for the KIDS application and integrated into existing CYF in the IT Management office
Languages and Technologies: SQL server 2005, T_SQL, stored procedures and Triggers, VSTS