Post Job Free
Sign in

Data Engineer Machine Learning

Location:
Overland Park, KS
Posted:
October 15, 2025

Contact this candidate

Resume:

VAISHNAVI NAREDLA

DATA ENGINEER

+1-913-***-**** ****************@*****.*** LINKEDIN

PROFESSIONAL SUMMARY:

• Having overall 4+ years of IT experience in entire SDLC life cycle as well as Data Engineer professional with over one year of experience in Data Science, Data Strategy, Data Mining, Statistical Analysis and Big Data Analytics in Financial Services, Banking, Web analytics and Telecommunications domains.

• Proficient in Python, SQL, and distributed data processing frameworks, with hands-on expertise in Google Cloud Platform (BigQuery, Cloud Storage, Pub/Sub), Azure (Databricks, Data Factory, Azure DevOps), and AWS.

• Chartered Financial Analyst, Financial Risk Management with extensive working knowledge of Banking and Financial Services, Risk Management and IT services.

• Extensive experience in maintaining and analyzing all types of SQL databases. Adept at database management mining specific data from SQL information and working closely with departmental managers to create useful reports.

• Experience with Statistical Analysis and Machine Learning algorithms like Cluster Analysis, Association Rules, Linear and Logistic Regressions, Classification and Regression Trees, Naive Bayes, Text mining, Optimization, Bagging, Boosting, Random Forest, Neural Networks and Other Machine Learning algorithms.

• Working knowledge of Big Data concepts viz. Hadoop/HDFS and Map - Reduce with applications like HBase, Apache Spark and Microsoft Azure ML.

• Expertise in implementation of data science products using machine learning algorithms on AWS cloud architecture. Extensive expertise in statistical and analytics tools and languages like SAS, SQL on massively parallel architecture like Hadoop and Teradata.

• Experience with Big Data analytics and visualization tools like Tableau and Microstrategy 9 for dashboards and automation.

• Experienced in Agile/Scrum environments, leading requirements analysis, system design, coding, testing, and documentation for complex projects.

• Excellent analytical and problem solving skills and ability to work independently and as a part of the team. TECHNICAL SKILLS:

Languages & Programming: Python, R, SQL, SAS (Base, SQL, Stat, Graph, Macros), HTML, CSS, JSP, XML, DHTML Cloud Platforms: Google Cloud Platform (BigQuery, Cloud Storage, Pub/Sub), Azure (Data Lake, Databricks, Data Factory, Azure DevOps), AWS (S3, Redshift, Lambda, EC2)

Data Engineering & ETL: Data Extraction & Transformation (SAS, SQL), Data Modeling, Data Warehousing (SAS Warehouse Administrator), Batch Processing, ML Pipelines, Automation Scripts (Python, SAS Macros), DataBricks Databases: Oracle, DB2, MySQL, Teradata, SQL Server Machine Learning & Analytics: Statistical Modeling, Time Series Forecasting, Anomaly Detection, Optimization Models, Customer Behavior Analysis, Performance Monitoring Visualization & BI Tools: Tableau, MicroStrategy, MS Excel (Pivot Tables, Dashboards), MS PowerPoint Cloud & Deployment: AWS (EC2, Cloud Automation, ML Deployment), End-to-End Tool Operationalization, Azure Databricks

Web Technologie& Frameworks: Django (Python Web Framework), HTML/CSS, JavaScript (Basic), JSP, Apache Spark, Delta Lake, Hadoop,

Workflow & Orchestration: Apache Airflow, Azure Data Factory, Google Cloud Composer, CI/CD Pipelines (Azure DevOps, Git)

Development Tools & Methodologies: UML, Rational Rose, Test Driven Development (TDD), User Acceptance Testing (UAT), Agile (Scrum)

PROFESSIONAL EXPERIENCE:

Client: CVS, NY (Contract)

Role: Data Engineer Sep 2024 – Present

Responsibilities:

• Extensively used SAS and SQL for extraction, transformation and loading of data from Large Scale RDBMS like Oracle and DB2.

• Conducted data manipulation using merging, appending, concatenating and sorting datasets in SAS.

• Created several applications for the purpose of Statistical Modeling and Data mining using SAS/Base, SAS/SQL, SAS/Stat, SAS/Graph and also automated applications using SAS/Macros.

• Involved in administration of data warehouse using Warehouse Administrator functionality of SAS

• Design and development of different data models according to user specifications in the development of databases for small applications.

• Designed and implemented scalable data pipelines in Azure Databricks using PySpark and Delta Lake to process and transform terabytes of raw data from Azure Data Lake Storage (ADLS) into structured formats for downstream analytics.

• Researched and studied Customer spending habits spanning several merchant categories and identified and formed a baseline for launching several merchant centric marketing campaigns and incentives and also resulted in the creating of Merchant Dashboard used by Product executives on a monthly basis.

• Developed and orchestrated end-to-end ETL workflows using Azure Databricks notebooks integrated with Azure Data Factory (ADF), improving data refresh reliability and reducing pipeline runtime by 30%.

• Created and implemented detailed Staffing Model for Security Solutions Business to evaluate and optimize internal and subcontractor staffing requirements based on the observed and forecasted backlogs, pipeline and historical demand analysis using time series forecasting, and optimization models.

• Operationalized and automated end to end staffing tool on the AWS cloud using R & Tableau in conjunction with the cloud architecture team to communicate model outputs and insights with business partners.

• Developed and implemented end to end analytics and machine learning pipeline for the Cloud compute and architecture organization within Disney’s Technology Solution and Services.

• Leveraged Databricks Delta Lake for ACID-compliant, incremental data processing and implemented data versioning to enable point-in-time recovery and rollback of historical data.

• Performed several Banking Center level performance analysis aimed at streamlining and optimizing the sales process and monitoring sales of certain high value checking products.

• Designed and implemented time series forecasting models for estimating server provisioning on the cloud environment which resulted in accurate demand forecasting for cost optimization and budgeting.

• Performed extensive data mining and exploration on the virtual and physical server performance log data and consequently applied anomaly detection algorithm using machine learning tools like R aimed at proactively determining performance issues and reducing downtime. Client: HCL, India

Role: Data Engineer Sep 2021 to Jun 2023

Responsibilities:

• Requirement Analysis and Estimation of project timelines.

• Created Business Logic using Python to create Planning and Tracking functions.

• Developed Python batch processors to consume and produce various feeds.

• Developed internal auxiliary web apps using Django framework with CSS / HTML framework.

• Optimized Spark performance in Databricks by tuning cluster configurations, caching strategies, and partitioning techniques, resulting in a 45% improvement in job execution time and cost savings on compute resources.

• Involved in designing, constructing, and documenting various object-oriented applications using UML, Rational Rose.

• Built and scheduled real-time streaming pipelines using Structured Streaming in Databricks to ingest event data from Azure Event Hubs, enabling real-time analytics and alerting dashboards.

• Collaborated with data scientists to develop and deploy machine learning models within Azure Databricks using MLflow for experiment tracking, model versioning, and lifecycle management.

• Built database Models, Views, and APIs using Python for interactive web-based solutions.

• Used Python scripts to update the content in the database and manipulate files.

• Designed and developed the presentation layer/client end using HTML, XML, DHTML, and JSP.

• Wrote SQL Queries, Store Procedures, Triggers, and functions in MySQL Databases.

• Coordinate with SMEs, other architects, and senior technical staff to identify clients’ needs, document assumptions, build out new requirements, and move code through user acceptance testing.

• Implemented Test Driven Development (TDD) strategy for the project.

• Developed and executed the User Acceptance Testing portion of the test plan. Client: IICL, India

Role: Technical Analyst Intern July 2020 – May 2021 Responsibilities:

• Performed extensive Data mining on large scale Relational Databases like Teradata and DB2, having accessed, created and maintained detail and summary tables.

• Assisted in offering support to other personal who were required to access and analyze the SQL database.

• Performed Path Analysis for measuring visitor drop rates and nodes with heavy bottlenecks to optimize customer experience.

• Adept at database management mining specific data from SQL information and working closely with departmental managers to create useful reports.

• Benefited and Helped in creating and presenting informational reports for management based on SQL data.

• Extensively used MS Excel pivot tables and MS PowerPoint for presenting results and making recommendations to Marketing and Sales teams and senior Executives.

• Helped in mining data from the SQL database that was used in several significant presentations..

• Performed extensive data quality analyses to evaluate existing data processes and recommended process fixes for accurately and effectively generate and measure data for analytics and reporting for leadership teams.

• Created and implemented automated data visualizations for tracking organizational KPIs by creating interactive visualizations and dashboards using Tableau and Micro strategy.

• Responsible for preparing the existing SQL platform for upgrades that were installed as soon as they were released.

• Created and maintained weekly and monthly Management Reports and Dashboard for senior management and Forecasting teams.

• Displayed high level of Project Management and communication skills by interacting with the Line of Business from inception and requirements gathering to presenting final results and suggestions.

• Conceived detailed modeling plan for developing customer value metric for identifying social influencers. Education:

• Masters in Computer Science at University of Central Missouri(Aug 2023 – May 2025)

• Bachelors in Information Technology at JNTU, Hyderabad, India (Aug 2017 – July 2021)



Contact this candidate