Venumadhav Dudi
Email: **************@*****.***
Mobile: +1-937-***-****
Data Engineer
PROFESSIONAL SUMMARY:
Results-driven Data Engineer with 4+ years of experience, demonstrating expertise in analytical thinking and attention to detail, contributing to innovative problem-solving within Agile/Scrum teams. Strong communication skills are used to present technical information.
Specialized in connecting dots across various applications to understand end-to-end views, proficient in Microsoft Office suite, and adept at writing and analyzing complex PL/SQL queries and stored procedures. Team player who can influence.
Proficient in query tools, Oracle Exadata (10g and above), and skilled in effort and financials estimation, enabling data-driven decision-making and ensuring alignment with business needs. Willingness to ask questions.
Adept at cross-functional collaboration, predictive analytics, and DevOps automation, with hands-on expertise in Python, SQL, and Spark, while effectively communicating across the organization to both technical and non-technical audiences.
Seeking to contribute technical depth and strategic insight to high-impact data engineering roles, leveraging strong communication and presentation skills to guide teams for success and identify priorities effectively.
Delivered production-grade solutions using predictive modeling, supporting product and marketing decisions, while working well in a team environment with minimal supervision and managing multiple projects simultaneously.
Migrated monolithic data pipelines to improve scalability, reducing deployment complexity, and demonstrating an analytical mindset, attention to detail, and innovative thinking skills within Agile/Scrum teams.
Strengthened observability in data platforms for proactive anomaly detection, showcasing expertise in problem-solving and the ability to connect dots across various applications to understand end-to-end views.
Drove alignment across data engineering, product, and analytics teams by leading cross-functional initiatives, accelerating project delivery, and ensuring data quality for critical business metrics, demonstrating strong PL/SQL skills.
TECHNICAL SKILLS:
Languages & Tools - Python, SQL, R, PySpark, Git, Docker, Kubernetes, Terraform, Jenkins, PL/SQL
Cloud Platforms - Google Cloud (BigQuery, Dataflow, Composer, IAM, Pub/Sub), AWS (EC2, S3, Redshift, Kinesis), Azure (Synapse, Data Factory, ADF)
Big Data & ETL - Airbyte, Apache Airflow, Spark, Kafka, Hadoop, Hive, Informatica, SSIS, dbt
Databases - Snowflake, MySQL, SQL Server, MongoDB, NoSQL, Oracle Exadata, Oracle 10g
Data Modeling - Star & Snowflake Schemas, Kimball, Inmon, Data Vault
DevOps & CI/CD - GitHub Actions, Jenkins, Terraform, Secret Manager
Monitoring & Compliance - Stackdriver, Prometheus, Datadog, Great Expectations, GDPR, HIPAA
Visualization & Analytics - Tableau, Looker, Power BI, Google Data Studio
Machine Learning - BigQuery ML, Predictive Modeling, A/B Testing, Regression, Decision Trees
Others - Microsoft Office Suite
PROFESSIONAL EXPERIENCE:
Airbyte Jan 2024 – Present
Data Engineer
Responsibilities:
Demonstrated analytical thinking by designing and maintaining scalable ETL pipelines using Airbyte Cloud and BigQuery, improving data availability by 99.9% through automated failovers and syncing from 10+ APIs. This required attention to detail and innovative thinking skills.
Automated ingestion for 15+ data sources and applied Python/SQL transformations, reducing manual workload by 40% and accelerating downstream analytics, showcasing strong PL/SQL skills and the ability to write and analyze complex queries. This improved efficiency.
Enabled proactive incident response by implementing real-time monitoring with Prometheus and Datadog, reducing ETL downtime by 35%, demonstrating proficiency in connecting dots across various applications and business to understand the E2E view.
Accelerated deployment speed by 50% through microservice containerization using Docker and orchestration on Kubernetes, requiring the ability to effectively communicate across the organization depending on the audience i.e. in Tech speak.
Designed DAGs using Apache Airflow, improving ETL success rate by 25% through dependency management and retry strategies, showcasing expertise with Microsoft Office suite usage for documentation and presentations to both technical and non-technical audiences.
Reduced latency by 70% with real-time streaming into BigQuery using GCP Pub/Sub, enabling faster anomaly detection, requiring willingness to ask questions and reach out for assistance as required to ensure optimal performance.
Strengthened data security by implementing IAM, RBAC policies, and encryption, eliminating unauthorized access incidents, demonstrating good skills with identifying priorities and managing multiple projects simultaneously in a fast-paced environment.
Reduced dashboard latency by 60% by building optimized data marts in BigQuery for Looker and Tableau, showcasing know-how working in Agile/scrum teams for prioritization of work and resource assignments to meet project deadlines.
Integrated PyTest and Jenkins in CI/CD pipelines, cutting release regressions by 30% with automated testing, demonstrating strong communication and presentation skills (both to technical and non-technical audience) during sprint reviews.
Boosted data quality by 98% via automated validation checks using Great Expectations and scheduled Python jobs, functioning as a strong team player who can influence and guide the team for success in achieving data quality goals.
McKesson Mar 2022 – Jul 2023
Data Engineer
Responsibilities:
Designed scalable pipelines using Dataflow and BigQuery, enabling real-time analytics and reducing processing time by 40%, demonstrating analytical thinking and problem-solving skills to optimize data workflows and improve efficiency.
Orchestrated workflows with Cloud Composer, achieving consistent ETL success and on-time SLA compliance, showcasing strong PL/SQL skills and the ability to write and analyze complex queries for data transformation and validation.
Unified SAP data from Azure Synapse into BigQuery, improving cross-system reporting speed by 50%, requiring the ability to effectively communicate across the organization depending on the audience i.e. in business speak (exec summary).
Implemented schema validation and lineage using Google Data Catalog, increasing trust in analytics assets, demonstrating proficiency in connecting dots across various applications and business to understand the E2E view of data governance.
Automated GCP provisioning with reusable Terraform modules, boosting infrastructure consistency and reducing provisioning time, showcasing expertise with Microsoft Office suite usage for creating documentation and presentations for stakeholders.
Delivered 10+ dashboards in Looker and Data Studio, empowering business stakeholders with self-service analytics, requiring willingness to ask questions and reach out for assistance as required to ensure accurate and insightful visualizations.
Deployed and maintained scalable APIs using Cloud Functions and Cloud Endpoints, enabling secure access to processed datasets, demonstrating good skills with identifying priorities and managing multiple projects simultaneously to meet business needs.
Cut query costs by 40% and improved speed by tuning BigQuery SQL and applying clustering/partitioning strategies, showcasing know-how working in Agile/scrum teams for prioritization of work and resource assignments to optimize performance.
Reduced latency in ETL monitoring by configuring Stackdriver log-based metrics with real-time alerting, functioning as a strong team player who can influence and guide the team for success in maintaining data pipeline stability.
Led sprint ceremonies, provided estimates, and delivered roadmap items on schedule in Agile environments, demonstrating strong communication and presentation skills (both to technical and non-technical audience) during sprint reviews and planning.
LendingTree Mar 2020 – Feb 2022
Data Analyst
Responsibilities:
Delivered real-time marketing insights by building Tableau dashboards from Snowflake and Redshift, accelerating decision-making, demonstrating analytical thinking and attention to detail in visualizing complex data sets for stakeholders.
Increased marketing ROI by 20% through campaign performance analysis and user behavior modeling, showcasing strong PL/SQL skills and the ability to write and analyze complex queries for data extraction and manipulation.
Created optimized SQL for reporting and automated ETL tasks, saving 10+ hours/week of manual effort, requiring the ability to effectively communicate across the organization depending on the audience i.e. in business speak (exec summary).
Partnered with engineering to define pipeline requirements and reduce data latency for core business metrics, demonstrating proficiency in connecting dots across various applications and business to understand the E2E view of data flow.
Standardized and cleaned datasets using validation rules and data wrangling in Python, ensuring 99% data accuracy, showcasing expertise with Microsoft Office suite usage for creating reports and presentations for management.
Automated daily performance reports using Python scripts and REST APIs, enabling real-time business updates, requiring willingness to ask questions and reach out for assistance as required to ensure data accuracy and timely delivery.
Converted Excel reports into interactive Tableau dashboards, reducing manual update time by 80%, demonstrating good skills with identifying priorities and managing multiple projects simultaneously to meet reporting deadlines.
Analyzed A/B tests to evaluate marketing strategies, influencing product direction with statistically significant results, showcasing know-how working in Agile/scrum teams for prioritization of work and resource assignments to support data-driven decisions.
Validated cross-platform metrics to reconcile dashboards with transactional systems, ensuring C-level trust in data, functioning as a strong team player who can influence and guide the team for success in maintaining data integrity.
Created automated data dictionaries and lineage documentation using metadata APIs, improving data discoverability for business users, demonstrating strong communication and presentation skills (both to technical and non-technical audience) during training sessions.
Certifications:
Tableau Specialist
Azure Fundamentals
GCP Data Engineer
Educational Details:
M.S. in Information Studies - Trine University