Post Job Free

Resume

Sign in

Azure Data Engineer

Location:
Ashburn, VA
Posted:
February 01, 2024

Contact this candidate

Resume:

SUMMARY

●Diligent and results-oriented Software Developer with a master’s in computer science and 3 years of professional experience in data engineering.

●Proficient in diverse programming languages including SQL, Python, PySpark.

●Hands-on experience with Azure Data Factory, Azure Data Pipelines, Power BI, and MongoDB.

●Skilled in designing and executing end-to-end ETL workflows, contributing significantly to the success of data-driven projects.

●Experienced utilizing Power BI to create interactive dashboards, visualizations, and reports.

●Proficient in ERD data models creation, SQL query performance tuning using tools like DB Visualizer

●Designed and developed cloud migration solutions for migrating On-Premises databases and ETL using Azure migration tools.

●Developed and optimized CI/CD DevOps pipelines using Azure DevOps, and other related tools like GitHub, and Unix Shell Scripts.

●Built scalable Data pipelines leveraging AWS serverless computing services like Lambda, Glue, SNS etc.

●Demonstrated expertise in utilizing various AWS services, including S3, API Gateway, IAM, CloudWatch, SES, and DynamoDB for comprehensive and secure application implementations.

●Adept with Agile/Scrum, and SDLC methodologies.

STRENGTHS

●Cloud Data Engineering & Analytics

●ETL & Data Lake solutions

●Data Science & Machine Learning

●Data Pipelines & Workflows

Agile (Scrum) Project Lifecycles

TECHNICAL SKILLS

●Languages: SQL, Python, PySpark, Scala, Shell Scripts, Java Script, HTML, CSS

●Data Technologies: Spark, Snowflake, MongoDB, Lakehouse Architecture

●Azure Technologies: Azure Data Factory, Data Pipelines, Power BI, Azure Databricks, Azure Synapse Analytics.

●AWS Technologies: S3, API Gateway, IAM, Lambda, Glue, DynamoDB

●IDE Tools: Visual Studio Code, PyCharm, IntelliJ, DB Visualizer Tool, Power BI

●CI/CD Tools: Git, GitHub, Azure DevOps, JIRA

EDUCATION

Master of Science in Computer Science - University Central Missouri, Lees Summit, MO May 2024

Bachelor of Technology in Computer Science & Systems Engineering, JNTUA, India June 2021

CERTIFICATIONS

Azure Data Engineer Associate, Microsoft Jan 2024

Azure Data Fundamentals, Microsoft Dec 2023

Power Platform Fundamentals, Microsoft Feb 2022

AWS Concepts, Linux Academy by Udemy Aug 2019

Pursuing SnowPro Core

PROFESSIONAL EXPERIENCE

Akhil Software Corp, Ashburn, VA Aug 2022 – Present

Data Engineer - ReAlign Insurance Holdings Inc

●Created Databricks notebooks, Delta Tables following Lakehouse architecture: Bronze, Silver, and Gold.

●Orchestrated ETL data pipelines using Azure Data Factory, Data Flows, Databricks notebooks to extract, transform, and load data from various sources into a centralized data lake/delta lake.

●Disparate data sources in various formats like Text, CSV, JSON, Parquet, and EDW were ingested into Data Lake/ Delta Lake (ADLS) hosted in Azure Cloud using Databricks ETL processes/notebooks written in SQL and Python utilizing Spark Core and Spark SQL libraries.

●Developed CI/CD pipelines using Azure DevOps and Shell Scripts for deploying the artifacts to various environments: test, stage, and production.

●Worked closely with data scientists to understand data requirements and implemented solutions for advanced analytics and machine learning.

●Leveraged JIRA for Scrum, GitHub for Source Control, Azure DevOps for (CI/CD).

●Contributed to the documentation (Confluence) of data engineering processes, ensuring knowledge transfer and team alignment.

Tata Consultancy Services (TCS), Bangalore, India Dec 2020 – Dec 2022

Data Engineer – Raymond James Sept 2022 – Dec 2022

Designed and implemented a real-time data pipeline to process semi-structured data by integrating raw records from various data sources to Azure Synapse Analytics.

Ingested Streaming data into azure Synapse Analytics from the Event Hubs, IoT Hub that route messages to Azure Event Hubs. And used Synapse Analytics capabilities to perform transformations, Aggregations on the data.

Implemented and actively monitored programs to ensure precision and efficacy, including the implementation of alerts and notifications for timely response to any failures.

Data Engineer - Becton Dickinson July 2021 - Aug 2022

●Developed the data pipelines using Azure Data Factory, ensuring the seamless extraction, transformation, and loading of diverse datasets into Azure Data Lake Storage and SQL DW/synapse analytics for the Visualization Purposes.

●Built Data Flows in Azure Data Factory for performing ETL by applying complex data transformations and writing to Azure SQL DW using the activities such as Source, Lookup, filters, aggregations, updates, pivots, Sinks, etc., and linked services such as ADLS, Azure SQL Database, Azure Blob Storage, Key Vaults, Azure SQL Data warehouse, API Service and more.

●Automated the pipelines using scheduled event based tumbling window triggers in Azure Data Factory.

●Engineered end-to-end ETL workflows, contributing significantly to the success of data-driven projects.

●Utilized Power BI for creating interactive dashboards, transforming raw data into visually compelling insights.

●I did a POC on migration of an app hosted on Azure App service to Kubernetes.

●I implemented a Deployment (Stateless Set) for the frontend app and web API by using Docker file, facilitating seamless scalability. Additionally, a Stateful Set was employed for the database, ensuring stable identities and effective management of persistent data.

Junior Data Engineer - Accenture Dec 2020 – June 2021

Built basic ETL that ingested transactional and event data from a web app to Azure Data Warehouse.

Worked with client to understand business needs and translate those business needs into actionable reports in Tableau.

Used Python in spark to distribute data processing on large datasets in Azure data factory using HDInsight’s, improving ingestion and speed.

Supported implementation and active monitoring of pipelines if any failures occur.

ACADEMIC PROJECTS:

Library Management System: MS Project at UCM Aug 2023 – Nov 2023

●Leveraged MongoDB for enhanced scalability and adaptability, revolutionizing the system architecture.

●Executed the project by developing the front end using Angular and Angular Material for a responsive user interface.

●Employed AWS Lambda functions in Python for critical operations such as transactions and user authentication in the back end.

●Utilized AWS services including S3 Hosting, API Gateway, and DynamoDB for scalable and reliable infrastructure.

●Ensured security through IAM roles, CORS configuration, and hashed/salted passwords for access control.

●Integrated AWS SES for email notifications and utilized AWS CloudWatch for effective debugging and monitoring.

●Achieved a robust, scalable, and user-friendly platform for online book management.



Contact this candidate