Post Job Free
Sign in

Data Analytics & Cloud Data Engineer Specialist

Location:
Indore, Madhya Pradesh, India
Posted:
November 21, 2025

Contact this candidate

Resume:

Abhijeet Joshi

Mobile: +91-930*******

Email, GITHUB, TABLEAU

Professional Summary

● Served as Data Analyst, Database Developer and Database Admin with a track record of successfully managing the IT projects.

● Designed and maintained robust ETL pipelines using PySpark, DBT, and SQL Server, processing terabytes of structured and semi-structured data across cloud platforms.

● Developed and optimized modular, testable data models in DBT, enabling scalable transformations and reducing data duplication by 40%.

● Built interactive dashboards and data visualizations using Power BI and Tableau, delivering insights to C-level stakeholders and improving decision-making turnaround time.

● Migrated and orchestrated workloads from on-prem SQL Server to cloud platforms like Azure Synapse, AWS Redshift, and Google BigQuery, reducing infrastructure costs by 30%.

● Automated data quality checks and implemented data cleaning workflows, ensuring 99.9% data accuracy across multi-source reporting pipelines.

● Leveraged Snowflake for high-performance data warehousing and analytics, including time travel, zero-copy cloning, and dynamic data masking features.

● Implemented CI/CD pipelines for data workflows using Git-based versioning and scheduled jobs via Azure Data Factory and Airflow.

● Led cross-functional teams in the design of cloud-native data architecture, integrating APIs, event-driven ingestion, and near real-time data processing. Technical Skills

● Languages: Python, SQL, PySpark, Shell Script

● Cloud Tools: Databricks, ADF, Airflow, DBT, AWS Glue, Azure Synapse, Azure Functions, Azure Logic Apps.

● Databases: Snowflake, Redshift, SQL Server, MySQL, PostgreSQL, MongoDB

● Visualization: Power BI, Tableau, QlikView

● Other: Git, Agile/Scrum

Education

● Master of Science in Computer Science (Major) – 2016 California State University, Los Angeles GPA- 3.5

● Bachelor of Engineering in Computer Science - 2012 Shiv Kumar Singh Institute of Technology and Science, India GPA-4.0 Professional Experience

Tata Consultancy Services Indore, Madhya Pradesh

Senior Engineer Database July 2021- Present

● Designed and deployed scalable ELT pipelines using PySpark on Azure Databricks and ADF to ingest data from heterogeneous sources.

● Developed data warehouse models in Snowflake including staging, integration, and mart layers.

● Created DBT models for transforming data, enforcing governance, and CI/CD pipeline integrations.

● Built Power BI dashboards for Sales and Finance domains, integrating row-level security using Azure AD.

● Implemented logging and monitoring for production pipelines using Azure Monitor and Log Analytics.

● Orchestrated end-to-end data workflows using ADF and automated job triggers based on event-based logic. Adobe Systems Noida, Uttar Pradesh

Business Analyst August 2020- June 2021

● Built and managed ADF pipelines and integrated them with Snowflake for data processing and analytics.

● Utilized DBT for transformation logic on customer feedback data to drive insights from Clarabridge models.

● Enhanced Power BI reports for customer experience teams, adding drill-through and KPI analysis.

● Implemented monitoring and alerting framework to proactively manage data quality issues. Capital One Richmond, Virginia

Data Analyst II June 2018 – June 2020

● Led end-to-end Tableau cloud migration as Site Admin, managing 500+ workbooks across 6 departments, replicating data sources, ensuring metadata consistency, and optimizing content structure post-migration.

● Performed predictive analytics and large-scale data handling using Snowflake, Redshift, Teradata, and Databricks, writing Python scripts and SQL queries to extract, transform, and anonymize financial datasets with over 10 million records.

● Engineered data ingestion and analysis pipelines by consuming APIs (DevEx, OpsInsight, Nebula), registering and managing S3 buckets, and integrating with Databricks using Spark and Python for real-time reporting and monitoring.

● Developed dashboards and insights on digital asset usage, automating reporting on table popularity and user interactions across Snowflake and Redshift platforms, significantly improving data discoverability and governance.

Integra Technologies Wilmington, Delaware

Data Analyst Dec 2017 – June2018

● Data Analysis using MySQL for queries from DB2, Excel, Access DB.

● Combining structured and unstructured information for data warehousing and analyzing claims data to help identify cost and recovery. Working with formats like csv, xml, and Excel

● Used services like EC2, S3, DynamoDB, Autoscaling Groups (ASG), RDS, IAM.

● Worked with Management frameworks and Cloud Administration tools. Managed Amazon redshift clusters such as launching the cluster by specifying the nodes and performing the data analysis queries.

● Experienced in Performance Tuning and Query Optimization in AWS Redshift.

● Built and scheduled PySpark batch jobs for data ingestion and transformation. TodaysHR West Hollywood, California

Technical Business Analyst Dec 2016 – Nov 2017

● Executed, Updated, and created MySQL queries to pull data in Teradata.

● Performed assortment coverage reporting using Teradata, MySQL, and advanced MS Excel.

● Created Rich dashboards using Tableau Dashboard and prepared user stories to create compelling dashboards to deliver actionable insights.

● Completed ETL processing, extracting data from Snowflake warehouse and pushing it back to Redshift and Snowflake.

California State University Los Angeles, California Database Developer/ Data Analyst Feb 2016 – Dec 2016

● Designed and deployed end-to-end BI solutions by building metrics, dashboards, and advanced visualizations in Tableau and SQL, streamlining report generation and enhancing stakeholder decision-making through daily, weekly, and monthly insights.

● Built complex SQL scripts, procedures, and data extraction pipelines, automated report workflows, and provided production support to ensure consistent performance of financial systems and BI environments across cross-functional teams.

● Engineered scalable ETL processes using PySpark, integrating and cleaning large volumes of structured and unstructured data for advanced reporting, predictive analytics, and dashboard readiness.

● Leveraged Snowflake for cloud data warehousing, optimizing storage and compute resources while managing data sharing, time travel, and performance tuning across analytics pipelines and business domains.

Warner Bros Interactive Entertainment Burbank, California Database Administrator Intern June 2015 – Aug 2015

● Led end-to-end data analysis and reporting for gaming titles like Batman: Arkham Knight, Mad Max, and Gauntlet, using Vertica, SQL Server, and PL/SQL to deliver actionable insights for game development and user engagement strategies.

● Designed and deployed real-time dashboards in Tableau and Google Analytics to visualize multi-source user-profile data, with incremental refresh and scheduled updates improving analytics availability and accuracy.

● Owned ETL processes and performance tuning, handling data extraction, transformation, and loading from diverse sources into data warehouses; documented ETL design and implemented source-to-target mapping and transformation logic.

● Collaborated cross-functionally with BI teams and stakeholders, delivering ad-hoc and scheduled reports, managing dashboard deployments, and leading mini-projects from requirements gathering through to Tableau publishing and optimization.

Indian Paint Industries Indore, India

Data Analyst Jan 2013 – Dec 2013

● Developed and maintained a real-time inventory website using Spring-Hibernate, Bootstrap, HTML/CSS, with email notifications, social media integration, and dynamic inventory tracking; supported by MySQL and MongoDB for backend data storage.

● Built interactive dashboards and KPIs for executive reporting using Tableau, QlikView, Excel, and Access, delivering actionable financial insights and supporting ad-hoc analysis for business decisions.

● Performed end-to-end data analysis and ETL design across structured/unstructured sources (CSV, XML, Excel), creating BRDs/PRDs, UML diagrams, and leveraging SQL, Python, T-SQL, and Redshift for advanced reporting and business logic.

● Optimized data workflows and ensured quality assurance by implementing stored procedures, views, and robust change management, while also developing tools to transpose and normalize Amazon Redshift key-value data into analytics-ready formats.



Contact this candidate