AKASH NANASAHEB UDE
Ahmednagar, Maharashtra
***********@*****.***
Linkedin: https://www.linkedin.com/in/akash-ude-382888260 SUMMARY
● Data Engineer with 1.5 yrs of experience in building and optimizing data pipeline.
● Having hands on working experience in big data frameworks and Ecosystem including data ingestion and data processing(SPARK, Snowflake, HDFS)
● Have worked with AWS cloud big data technologies including S3 datalake, AWS glue, AWSlambda, cloudwatch.
● Have knowledge of spark ecosystem including spark core, Pyspark, SPARK-SQL
● Developed custom libraries and user-defined functions to perform data transformation and aggregation tasks.
● Built and optimized data processing jobs to ensure efficient performance, and tuned queries to improve overall system performance.
● Experience in designing and implementing data ingestion pipelines for batch processing of data from various sources.
● Possess excellent analytical, problem solving, communication and interpersonal skills with a demonstrated ability to work both independently and collaboratively in team environments. EDUCATION
CGPA, 8.6, Bachelor of Engineering,Electronics And Telecommunications, SPPU 10+2, 78%, Maharashtra state board of Secondary and Higher Secondary Education, Pune 10, 90%, Maharashtra state board of Secondary and Higher Secondary Education, Pune EXPERIENCE
DGCT INFOSOFT, NAGPUR
ServiceNow - Data Engineer 09/2022 - present
● Project name: Retail data analytics
It’s an end to end ETL project to analyze the customer retail data. In this project we built the data migration pipeline using Apache Spark. This project involved analyzing business data to provide insights and actionable recommendations to improve sales, customer satisfaction, and productivity in key areas such as inventory management, pricing, and marketing campaigns. Environment: Spark, AWS(DMS, lambda, Glue, Cloudwatch), s3 datalake, snowflake Roles and Responsibilities:
Daily stand meetings for tasks status discussions.
Apply the technical knowledge in Pyspark to solve day to day tasks.
Working with different file formats like parquet, csv to optimize storage performance
Using transformations in Pyspark / Spark SQL to ETL / clean the data.
Creating the spark glue jobs to apply the transformation on the data
Optimizing the spark job for better performance and monitoring them using cloudwatch and SparkUI
Implementing various transformations on data by understanding the client requirement and building the logic on them
Implement data validation and data quality checks
Write Sql queries in snowflake to extract insights from data as per business requirement DGCT INFOSOFT, NAGPUR
ServiceName - Data Engineering trainee 01/2022 - 05/2022
● Gained a thorough understanding and hands-on experience of big data frameworks and technical tools like
: PySpark, databricks, SQL and different data related libraries.
● Earned knowledge of new technologies like Tableau, Power BI
● Got hands-on-experience with cloud services like AWS, Microsoft Azure
● Worked with data analysis and visualization libraries in python such as Pandas, Numpy, Matplotlib, Seaborn TECHNICAL SKILLS
Languages: Proficient -> Python, SQL, Bash
Database: Oracle, MYSQL, PostgreSQL
Big Data Frameworks: Pyspark, spark-SQL, Hadoop, Sqoop, HDFS
Cloud Technologies: Amazon Web Service(AWS), Microsoft Azure
Data Warehouses: Snowflake, Hive
Tools : PyCharm, putty, data bricks
Technology: Git,Github, Jenkins,Docker
Python libraries: Pandas, Numpy, Matplotlib, Seaborn,Scikit-learn
Other: Problem solving, Data structures and algorithms PERSONAL DETAILS
Address: A/P-Khedle Kajali,Tal-Newasa, Dist-Ahmednagar,414603 Birth Date: 12-01-2000
Languages: Marathi, Hindi, English
Marital Status: Unmarried
Gender: Male
DECLARATION
I hereby declare that all the information in this document is accurate and true to best of my knowledge. Place: Ahmednagar signature
date: (Akash Nanasaheb Ude)