Post Job Free

Resume

Sign in

Data Processing Engineer

Location:
Chicago, IL
Posted:
December 03, 2023

Contact this candidate

Resume:

Kaushik Varma N

Chicago,IL +1-872-***-**** LinkedIn Github Hackerrank Email Website

SUMMARY

● Master’s in Computer Science with over 3+ Years of overall IT experience in Data Engineering.

● Working experience in Hadoop ecosystem (Gen-1 and Gen-2) and its various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager (YARN).

● Experience with components such as Cloudera distribution encompassing components like MapReduce, Spark, SQL, Hive, HBase, Sqoop, Pyspark.

● Good skills on NoSQL Database- Cassandra.

● Proficient in developing Hive scripts for various business requirements.

● Knowledge in Data Warehousing Concepts in OLTP/OLAP System Analysis and developing Database Schemas like Star Schema and Snowflake Schema for Relational and Dimensional Modeling.

● Good hands on in creating custom UDF in SnowFlake Data Warehouse.

● Load and transform large sets of structured, semi-structured and unstructured data from Relational Database Systems to HDFS and vice-versa using Sqoop tool.

● Good Experience on architecture and components of Spark, and efficient in working with Spark Core, Data Frames/Datasets/RDD API/Spark SQL, Spark streaming and expertise in building PySpark and Spark-Scala applications for interactive analysis, batch processing and stream processing.

● Hands-on experience in Spark, Scala, SparkSQL, Hive Context for Data Processing.

● Knowledge on GCP tools like Cloud Function, Dataproc, Big Query.

● Experience on Azure cloud i.e., ADF, ADLS, Blob Storage, Databricks, Synapse etc.

● Extensive working experience in an Agile development Methodology & Working knowledge on Linux.

● Expertise in working with big data distributions like Cloudera and Hortonworks. .

● Experience in tuning and debugging Spark applications and using Spark optimization techniques.

● Knowledge on architecture and components of Spark and demonstrated efficiency in optimizing and tuning compute and memory for performance and price optimization.

● Expertise in developing batch data processing applications using Spark, Hive and Sqoop. EXPERIENCE

Azure Data Engineer Jan 2023 - Nov 2023

ExxonMobil (contract) Houston, Texas

● Engineered SQL scripts to automate query processes, successfully eliminating the need for manual intervention and boosting query volume and accuracy by an impressive 40%

● Designed and implemented an automated system using PowerShell and Azure Cloud Shell, effectively streamlining the deployment of Azure data solutions.

● Crafted advanced data models that enhanced data processing efficiency in Azure, resulting in a substantial 30% increase in overall productivity

● Utilized Azure services such as Data Factory to construct robust data pipelines, enabling the seamless migration of data from legacy SQL servers to Azure Database. Accomplished this through the integration of Data Factories and Python scripting

● Collaborated closely with Analytics and BI teams to develop globally utilized metrics and reports, significantly reducing the need for manual data analysis by more than 15% Associate Software Engineer (Data) Dec 2020 - Dec 2021 Tech Mahindra Hyd, India

● Designed and implemented a highly scalable data model and data warehouse using Snowflake, leading to a remarkable 10% enhancement in data processing speed and a 15% reduction in storage costs

● Engineered and deployed data pipelines to elevate data quality, resulting in a notable 30% increase in data accuracy.

● Streamlined and fine-tuned ETL processes for seamless data loading into Snowflake, achieving a remarkable 50% reduction in data loading time and a 15% enhancement in overall data quality.

● Pioneered the development of highly optimized Spark code using PySpark and Spark-SQL, resulting in an exceptional 20% improvement in data processing speed and unparalleled data accuracy.

● Building and architecting multiple data pipeline, end to end ETL and ELT process for Data ingestion and transformation.

Data Engineer Nov 2019 - Dec 2020

Adiroha Bengaluru, India

● Optimized an existing data pipeline to improve its performance and scalability.

● Created and maintained documentation of data models, ETL processes, and data security policies, resulting in a 30% reduction in onboarding time for new team members and ensuring consistent data governance practices.

● Built a data warehouse in snowflake to capture historical data.

● Conducted data analysis to identify patterns and trends in customer behavior Data Engineer Intern June 2019 - Nov 2019

Adiroha Bengaluru, India

● Developed a new data pipeline to collect and load data from a new data source into the company’s data warehouse.

● Implemented complex SQL queries to extract the data from the data warehouse.

● Optimized SQL Queries using indexes as per the company requirements. PROJECTS

Winkart django,sql,javascript,docker(link) May 2023 - June 2023

● Built a complete end to end ecommerce website where a user can buy apparels.

● Used session keys to implement add to cart function to increment/decrement/delete items in the cart.

● Integrated PayPal payment system so a user can purchase products

● Implemented token based logging in for enhanced security Credit Card Spends Sql(link)

● Retrieved transaction details for each card type when it reaches a cumulative of 100000

● Retrieved which card and expense type combination saw highest month over month growth in Jan-201

● Found which city took least number of days to reach its 500th transaction after the first transaction in that city Snowflake ETL Pipeline (link)

● Created 3 layers to store the data and to capture the CDC data overtime

● For each layer created 3 pipes, 3 tables and 3 Streams to build the continuous data flow from amazon s3

● Used snowflake tasks to automate copying data into tables

● Transformed the data in 2nd layer(curating zone) and sent it consumption layer for performing data analysis More Side Projects

TECHNICAL SKILLS

Languages : Java, Python, C/C++, SQL, JavaScript

Data WareHousing : Snowflake,Pentaho, AWS RedShift Azure Cloud Tools : Azure Data Lake, Azure Blob Storage, Azure VM, Azure Synapse, Data Factory, Azure cosmos Big Data Tools : Hadoop, Hive, Spark, Metastore, Presto, Flume, Kafka Developer Tools : Git,Google Cloud Platform, VS Code, Visual Studio, PyCharm, IntelliJ, Eclipse Libraries : Pandas, NumPy, Matplotlib

ML Frameworks : ScikitLearn, TensorFlow

No SQL : HBase, Cassandra, MongoDB

EDUCATION

Western Illinois University Macomb,IL

Master’s of Science, Computer Science [2022-2023] GPA : 3.29 KL University AP, India

Bachelors in Computer Science [2016-2020] GPA : 3.25 CERTIFICATIONS

AWS Certified Cloud Practitioner



Contact this candidate