Srikanth Reddy
Senior Data Engineer Cloud & Big Data Solutions (AWS, Azure, Snowflake)
********************@*****.*** +1-682-***-**** LinkedIn Summary
Data Engineer with six years of experience delivering cloud-based data solutions for logistics, e-commerce, and financial services. Implemented configuration-driven Intelligent Data Platform (IDF) enabling multi-cloud deployment on AWS, Azure, Databricks, and Snowflake, reducing ETL pipeline development time by 70%. Led migration of legacy Alteryx-based manual workflows to AWS- native pipelines, resulting in 90% cost savings. Holds certifications in Snowflake, Databricks, Azure, and AWS. Education
Master of Science in Business Analytics
University of North Texas, G. Brint Ryan College of Business – Denton, TX • May 2022 Bachelor of Technology in Electrical Engineering
Jawaharlal Nehru Technological University – Hyderabad, India • May 2013 Skills
• Programming Languages: Python, SQL, Java, R, PySpark
• Databases: MySQL, PostgreSQL, Amazon Redshift, Azure Synapse, Oracle, MongoDB
• Big Data: Hadoop, Spark, Hive, Databricks, Sqoop, Scala
• Cloud Platforms: AWS (Glue, Lambda, S3, EMR, CloudFormation), Azure (Data Factory, Synapse, Databricks, DevOps)
• ETL Tools: Informatica, Alteryx
• MLOps & DevOps: Docker, Kubernetes, Apache Airflow, Azure DevOps
• Visualization: Tableau, Power BI, QlikView
• Statistics: Regression, ANOVA, Time Series Forecasting, Hypothesis Testing
• Project Tools: Jira, Confluence, Agile Scrum
• Certifications: Snowflake SnowPro Core, Apache Kafka (Confluent Certified), AWS & Azure Fundamentals Work Experience
Pandion Inc – Dallas-Fort Worth Metroplex
Data Engineer, Final Mile Dec 2023 – Present
• Spearheaded a configuration-driven Intelligent Data Platform (IDF) facilitating multi-cloud deployment across AWS, Azure, Databricks, and Snowflake environments; decreased ETL pipeline development timeline by 70%.
• Examined and assessed 15+ diverse data repositories, including databases, APIs, files, and streaming feeds, to formulate robust data ingestion strategies and confirmed data relationships.
• Pioneered fault-tolerant Medallion Architecture data pipelines ingesting 20TB+ monthly with AWS DMS, S3, Glue, Redshift, and Athena; reduced data latency by 30 minutes.
• Engineered a serverless data platform in AWS, replacing Alteryx with native services; decreased data processing latency by 65% and accelerated report generation.
• Upgraded data pipelines integrating AWS and Pricefx, diminishing SLA breaches by 15% through advanced monitoring alongside a 60% increase in data refresh rates.
• Engineered Change Data Capture (CDC) and Slowly Changing Dimension (SCD) Type-2 pipelines in Azure Databricks and Snowflake, orchestrated using Apache Airflow; enhanced data quality by 35%.
• Harnessed Airflow to orchestrate CDC and SCD Type-2 patterns within Azure Databricks, guaranteeing data integrity and improving the accuracy of real-time inventory management reports.
• Documented data flows, data dictionaries, and schema definitions, facilitating 20+ walkthrough sessions; enabled the BI team to derive insights 40% faster resulting in improved decision-making.
• Synchronized BI, product, and offshore teams through Agile best practices, elevating sprint velocity by 15% and streamlining data transparency, which received praise from senior management. Amazon (AWS) – Seattle, WA
Data Engineer Jun 2022 – Nov 2023
Project: Zone Replication & Real-Time Streaming (AWS Route53)
• Spearheaded the creation of a data pipeline for AWS Route53 zone replication across four regions, reducing DNS resolution latency by 25% for users in affected areas.
• Constructed robust CI/CD workflows leveraging CloudFormation, resulting in a 60% surge in deployment frequency, which enabled weekly releases rather than the former monthly release cycle.
• Conceptualized and implemented data quality checks within AWS Glue ETL jobs, decreasing data errors by 45% and ensured compliance with data governance policies, earning team recognition.
• Designed data visualizations with Tableau, SQL, and Python to highlight anomalies and improve data quality, leading to a tangible 90% reduction in data pipeline error rates. Project: GDPR Compliance for AWS Route53 (Europe & China)
• Integrated robust encryption protocols and access controls within data pipelines, fortifying the organization's data security posture across 7 data stores and boosting data quality by 20%.
• Transformed data lifecycle management by automating deletion workflows with Lambda, resulting in a 15% boost in overall compliance score, according to external audits.
• Spearheaded the design and deployment of interactive Tableau dashboards visualizing key compliance metrics, resulting in a verifiable 10% improvement in data accuracy for reporting.
• Designed and maintained an anomaly detection pipeline analyzing 900 million DNS requests weekly and presenting findings to compliance team to fix the three biggest causes of crashes.
• Automated potential violation detection by orchestrating predictive machine learning models, curtailing compliance risk by 35% and delivering key insights for legal teams. University of North Texas System – Dallas, TX
Data Engineer Intern Aug 2021 – May 2022
• Designed a Python script to automate the extraction, transformation, and loading (ETL) of financial data, supporting the ARIMA/SARIMA models and slashing manual effort by 40 hours monthly.
• Engineered the integration of disparate data sources into comprehensive Tableau dashboards, identifying and rectifying 250+ data discrepancies, improving data quality scores by 10%.
• Leveraged SQL queries in PostgreSQL and traced back errors, discovering faulty data mappings, which led to a 20% improvement in data validation and data quality.
• Revamped data cleaning and transformation workflows using Python, decreasing data pipeline execution time by 35% and enabling faster delivery of insights to key stakeholders across British Telecom. Tata Consultancy Services – Hyderabad, India
Data Engineer (Client: British Telecom) Jun 2013 – Jan 2015
• Assembled and maintained sales performance dashboards using Tableau, leading to a 15% reduction in query run time with improved data accuracy to fix frequent crash errors.
• Integrated 3 new data sources into the existing MySQL data warehouse using Informatica, expanding data coverage for real-time customer behavior analytics and improving customer retention by 12%.
• Implemented Python-based predictive models to analyze customer behavior, generating insights that improved customer lifetime value by 15% and directly led to the creation of a new product feature.
• Designed SQL queries to identify and resolve six critical data inconsistencies within customer databases, improving data quality and contributing to a 12% boost in customer retention.
• Unified data solutions with business reporting objectives alongside data analysts and engineers, resulting in the solutions being adopted by three downstream teams and enhanced data influence. Certifications
• Google Cloud Certified Professional – Data Engineer
• AWS Solution Architect
• Snowflake, Databricks, Azure, AWS Fundamentals
• MLOps (MLflow, Feast, SageMaker, Azure ML, Vertex AI)
• Generative AI Pipelines (Vector Databases: Pinecone, Weaviate, FAISS) Requirements Gathering & Elicitation Stakeholder Management Business Process Modeling & Improvement GAP Analysis Use Case & User Story Development Wireframing & Prototyping Functional & Non-Functional Requirements Documentation Data Mapping & Data Modeling Feasibility Analysis & Business Case Creation SWOT Analysis Root Cause Analysis (RCA ) Process Reengineering Agile (Scrum, SAFe, Kanban, LeSS) Waterfall Hybrid Methodologies Design Thinking Lean Six Sigma (DMAIC, DMADV) ITIL (for IT Business Analysts) BABOK (IIBA Business Analysis Body of Knowledge) Jira, Confluence Azure DevOps (ADO) Trello, Asana, Monday.com IBM Rational RequisitePro Jama Connect Microsoft Visio Lucidchart Bizagi ARIS Draw.io Balsamiq (Wireframes) Axure RPFigma Adobe XD InVision UiPath, Automation Anywhere Pega Systems AppianMicrosoft Excel (Advanced: Pivot Tables, Power Query, VBA) SQL (Structured Query Language) Power BI Tableau QlikView / Qlik Sense Google Data Studio (Looker Studio) SAS / SPSS (for advanced statistical analysis) R / Python (for data-driven BA roles) Test Case Documentation UAT (User Acceptance Testing) Coordination HP ALM (Quality Center), Zephyr, TestRail Automation Awareness: Selenium, Postman (API Testing) ERP: SAP, Oracle NetSuite, Microsoft Dynamics 365 CRM: Salesforce, HubSpot, Zoho CRM HRIS: Workday, ADP, SuccessFactors Supply Chain: SAP SCM, Manhattan Associates, JDA Strategic Thinking & Decision-Making Leadership & Mentoring Negotiation & Conflict Resolution Excellent Communication (Written & Verbal) Active Listening & Empathy Change Management Critical Thinking & Problem-Solving Cross-F unctional Collaboration IIBA: CBAP (Cer tified Business Analysis Professional), CCBA PMI-PBA: Professional in Busi ness Analysis Agile: Certified ScrumMaster (CSM), SAFe Agilist Lean Six Sigma: Green Belt, Black Belt Data-Focused: Tableau Certification, Microsoft Power BI Data Analyst Associate ERP/CRM: Salesforce Admin, SAP Certifications Wrike Smartsheet ClickUp Miro MURAL PractiTest JMeter QuickBooks Oracle Financials SAP FICO Blue Yonder (JDA) Blue Prism (RPA) Nintex Primavera P6 Rally (CA Agile Central) Targetprocess Guidewire Infor Epic Systems Coupa Cherwell Service Management Ivanti Service Manager Stormboard ConceptDraw MindMeister Notion SoapUI Katalon Studio Ranorex Pipefy Kissflow WorkFusion Talend RapidMiner Alteryx KNIME Proto.io Moqups Justinmind Caliber RM Helix RM ARIS Express yEd Graph Editor MagicDraw UML, BPMN, SysML DOORS Objecteering TopTeam Analyst CASESpec Informatica PowerCenter Ab Initio Pentaho (Kettle) Talend TIBCO BusinessWorks Bonita BPM jBPM (Red Hat) Lotus Notes Basecamp Zoho Projects CoreLogic Duck Creek Cerner JD Edwards Alteryx, KNIME, ARIS, Justinmind, SoapUI DOORS, Ab Initio, Bonita BPM, Silk Test, Lotus Notes, Duck Creek