Rachana Gardas
913-***-**** ****************@*****.***
PROFESSIONAL SUMMARY
Certified Snowflake Data Engineer with over 4 years of experience building and optimizing scalable, cloud-native data platforms across AWS, GCP, and Azure. Skilled in designing real-time and batch ETL/ELT pipelines using Snowflake, Python, SQL, and Apache Spark. Strong background in data modeling, dbt, pipeline orchestration with Airflow, and automation for analytics and AI/ML use cases. Proven ability to ensure data quality, governance, and security compliance across large datasets. Adept at collaborating in Agile/Scrum environments, working cross-functionally with product, engineering, and analytics teams to deliver reliable, high-impact data solutions and business insights through analytics, automation, and visualization. EXPERIENCE
Snowflake Developer
InterAction 24
Princeton NJ, USA Jan 2025 – Present
Designed and automated scalable batch and real-time data pipelines for migrating from legacy systems (Teradata, SSIS) to Snowflake and BigQuery, enabling secure, high-availability, analytics-ready data for advanced reporting and ML use cases.
Led end-to-end cloud data migration efforts, including strategy, execution, validation, and optimization, ensuring 100% data integrity and minimal downtime to support critical business analytics.
Developed comprehensive automated validation frameworks using Python and SQL, improving schema compliance, data accuracy, and reliability across platforms.
Built modular ETL workflows for historical and incremental data loads using Databricks, Talend, and Spark, supporting real-time business analytics and enhancing dashboard performance.
Implemented functional, integration, and security testing across AWS, Azure, and GCP pipelines to ensure robust data quality and trusted business reporting.
Automated data validation, schema checks, and load verification processes, reducing manual QA effort and accelerating delivery of analytics-ready data products.
Collaborated closely with business stakeholders and cross-functional teams (engineering, product, analytics) in Agile/Scrum environments to deliver scalable, business-focused data solutions.
Documented data models, pipeline architecture, SOPs, and troubleshooting guides to enhance data literacy and empower self-service analytics within business teams.
Championed code reviews and best practices, establishing standards that improved pipeline quality, maintainability, and performance.
Managed user access controls and dataset permissions in Snowflake and BigQuery, and conducted enablement sessions to train teams on secure, independent data exploration.
Modernized legacy SQL Server and SSIS pipelines to cloud-native Snowflake and BigQuery architectures, increasing scalability, reducing operational costs, and supporting evolving data needs.
Developed dbt models to transform raw data in Snowflake into analytics-ready tables, supporting marketing attribution and real-time reporting (Proof of Concept).
Conducted proof-of-concept deployments on AWS EKS for containerized Spark jobs, and evaluated AWS SageMaker for ML model training and real-time inference pipelines.
Implemented data governance standards and security policies across Snowflake and BigQuery to ensure data privacy, access control, and compliance.
Orchestrated batch and streaming data pipelines using Apache Airflow to improve reliability, automate dependencies, and ensure timely data availability. Data Engineer June2020–July 2023
LTIMindtree,
Ohio, USA
Developed scalable, cloud-based data engineering solutions supporting a high-volume supply chain network managing 500+ SKUs, enabling accurate tracking and advanced business analytics.
Automated real-time and batch data workflows to enhance product replenishment, sales forecasting, and demand tracking, driving operational efficiency and improved decision-making.
Designed region-based allocation logic, increasing delivery accuracy by 25% and reducing stockouts by 15%, directly contributing to key business KPIs.
Built and optimized ETL pipelines using Talend, Apache Beam (Dataflow), Databricks (Spark), and Python, reducing data processing time by 40% and accelerating analytics delivery.
Processed and integrated multi-terabyte datasets in BigQuery and Dataflow, ensuring near real-time data availability for executive dashboards and cross- functional business reports
Wrote advanced SQL logic (stored procedures, views, UDFs, CTEs) in BigQuery and Redshift to support complex analytics workflows and automated business reporting.
Delivered clean, business-ready datasets for Power BI dashboards, empowering leadership with real-time, data-driven insights for strategic decisions.
Implemented automated data validation, exception handling, and robust testing frameworks, improving data reliability and minimizing manual interventions.
Collaborated closely with business stakeholders, DevOps, and QA teams to quickly resolve data issues, reducing incident response time by 35% and improving system reliability.
Actively participated in Agile sprints, supporting planning, backlog refinement, and cross-functional deliverables to align with evolving business requirements.
The best practices applied in pipeline architecture, cost optimization, and documentation, ensuring scalable, maintainable, and future-proof data solutions.
Supported data mapping and migration validation initiatives involving IBM DB2, contributing to successful cross-platform reporting and system modernization.
Tuned BigQuery queries using clustering and partitioning strategies, improving performance and reducing costs for large-scale data analysis and reporting.
Conducted exploratory data analysis and collaborated with stakeholders to design dashboards and generate insights that improved forecasting accuracy. SKILLS
• Analytics & Reporting
Power BI (Advanced), Excel (Advanced), Dashboard Design, EDA, Data Storytelling, KPI Metrics, Insight Communication, Tableau (Basic)
• Programming & Scripting
SQL (Advanced), Python (Advanced), Bash, Scala, JSON, REST APIs
• Databases & Data Warehousing
Data warehouse, Snowflake, Google BigQuery, AWS Redshift, MySQL, Oracle (SQL, PL/SQL), IBM DB2 (Basic), DynamoDB, Dimensional Modeling, Data Vault, Star & Snowflake Schemas, NoSQL (MongoDB, DynamoDB), PostgreSQL
• ETL & Data Processing Tools
Talend, Apache Beam (Dataflow), Databricks (Spark), Airflow, ETL/ELT Pipelines, Data Modeling, Data Validation Automation, Data Transformation, Spark, Flink (Basic), Hadoop, Kafka, dbt (Intermediate, Proof of Concept), Dataproc
• Cloud Platforms & Services
Google Cloud Platform (BigQuery, GCS, Dataflow, Dataproc), AWS (S3, Redshift, Lambda, DynamoDB, EKS, SageMaker), Microsoft Azure (ADF, Data Lake, Blob Storage), Cloud Data Lakes
• Orchestration & Workflow Automation
Apache Airflow, Databricks Workflows, Automated Scheduling, Dependency Management
• Infrastructure & DevOps
Infrastructure as Code (Terraform, CloudFormation), Kubernetes (Basic), Docker (Basic)
• Version Control & Collaboration
Git, JIRA, Agile Methodologies, Peer Code Reviews, Cross-Functional Communication
• Machine Learning & Data Science (Basic)
Scikit-learn, Pandas, NumPy, Matplotlib, Seaborn, Data Visualization, Feature Engineering
• Operating Systems
UNIX/Linux, Windows
CERTIFICATIONS
SnowPro Core Certification – COF-C02 Snowflake Issued: July 2025 Expires: July 2027 Credential ID: S104470-250721-COF Scaled Score: 927 Collaboration, Market Place and Cost estimation (Snowflake) 2025 https://achieve.snowflake.com/1fd2c10d-8399-4191-82ad-3d30f53594d6#acc.xDhpnake Data Warehousing (Snowflake) 2025
https://achieve.snowflake.com/08d63a03-d875-478b-b5a5-0402cfe1dc7c#acc.PDdtw0YT PROJECTS
• University Management System
Technologies: Java and MySQL Led the development of a desktop-based application to manage student and faculty records. Designed and implemented database schemas and user authentication modules. Integrated CRUD operations and search filters, improving administrative task efficiency by 50%. Skills Gained: Database normalization, UI logic in Java, and implementing real-time data updates.
• Walmart Sales Data Analysis
Technologies: SQL, MySQL Performed exploratory and statistical analysis on Walmart’s sales data to extract actionable business insights. Wrote optimized SQL queries using joins, aggregates, and subqueries to identify top-selling products, seasonal demand, and performance trends across regions. Improved query performance by 40% using indexing.
Skills Gained: SQL query optimization, data profiling, business insights generation.
• Academic Activity Tracking System
Technologies: HTML, CSS, JavaScript, MySQL Designed and deployed a responsive web application to log, track, and report academic events. Created interactive frontend forms and integrated them with a backend MySQL database for real time updates. Increased event reporting accuracy by 60% compared to manual logs. Skills Gained: Full-stack web development, front-end-backend integration, and basic web security.
• Online Shopping Behavior Prediction
Technologies: Python, Scikit-learn, Pandas, Matplotlib Built machine learning models (KMeans Clustering, Logistic Regression) to predict customer purchase likelihood using behavioral data. Preprocessed over 50,000 records, engineered features, and achieved up to 82% model accuracy. Skills Gained: Data cleaning, feature engineering, ML model building and evaluation, visualization. INTERNSHIP
Greater Coder Technologies June 2020- July2020
During my internship at Great Coder Technologies, I worked on a project involving Sentiment Analysis of Twitter Data. My responsibilities included:
• Data Collection & Preprocessing: Extracted and cleaned Twitter data using APIs and NLP techniques.
• Feature Engineering: Applied text processing techniques such as tokenization, stopword removal, and TF-IDF vectorization.
• Model Development: Built and trained machine learning models using Python (NLTK, Scikit-learn) to classify sentiments.
• Performance Evaluation: Analyzed model accuracy and fine-tuned parameters to improve prediction results.
• Visualization & Reporting: Created data visualizations to present insights and findings effectively. Skills Gained: Python, Machine Learning, NLP, Sentiment Analysis, Data Preprocessing, Scikit-learn, Twitter API, Data Visualization. Summer Engineering Intern
The Smart Bridge, India May 2020-June2020
I worked on a Diabetes Mellitus Prediction using Machine Learning project. My responsibilities included:
Data Acquisition & Cleaning: Collected and preprocessed medical datasets to ensure high-quality input for the model.
Feature Engineering: Selected key attributes affecting diabetes prediction and applied techniques like normalization and feature scaling.
Model Development: Built and trained predictive models using algorithms such as Logistic Regression, Decision Trees, and Random Forest.
Performance Optimization: Evaluated model accuracy using metrics like precision, recall, and F1-score, and fine-tuned hyperparameters for better results.
Visualization & Reporting: Created interactive visualizations to showcase insights from the data and model predictions. EDUCATION
Master of Science in Computer Science Aug 2023 – May 2025 University of Central Missouri, USA