SRAVANI
Northlake, TX 317-***-**** *******.***@*****.***
https://www.linkedin.com/in/sravani-komatineni-69b53a259/
PROFESSIONAL SUMMARY
Senior Data Engineer with 9+ years of experience designing and implementing scalable data solutions across cloud platforms (AWS, GCP, Azure). Expert in building end-to-end data pipelines, optimizing data warehouses, and delivering actionable insights through advanced analytics. Skilled in modern data architectures including data lakes, streaming services, and big data processing frameworks. Proven track record of successfully migrating on-premise systems to cloud environments while ensuring data security, quality, and governance.
CORE COMPETENCIES
Data Engineering: ETL/ELT Pipeline Design, Data Modeling, Batch/Real-time Processing, Data Migration
Cloud Platforms: AWS (EMR, EC2, S3, Lambda, Redshift), GCP (BigQuery, DataProc, Cloud Storage, Pub/Sub), Azure (Data Lake, SQL Database, Data Factory)
Technologies: Python, Spark, Airflow, Kubernetes, Databricks, Snowflake, Docker, Terraform, Ansible
Big Data Tools: Hadoop Ecosystem (HDFS, Hive, Pig), Kafka, Matillion, Apache NiFi, Spark, Airflow, MapReduce, Sqoop, Oozie
Languages & Frameworks: SQL, PySpark, Scala, Java, React.js, JavaScript, Python, Java, Scala, PL/SQL, Shell Scripting, Angular.js, Node.js, Spring Boot
Database Systems: Oracle, SQL Server, MongoDB, Cassandra, HBase, Teradata, Snowflake, BigQuery, Redshift
Methodologies: Agile/Scrum, DevOps, CI/CD, Microservices Architecture
Data Visualization: Tableau, Power BI, D3.js, Grafana
Version Control & CI/CD: Git, GitHub, Bitbucket, Jenkins, Cloud Build
PROFESSIONAL EXPERIENCE
AMERICAN FAMILY INSURANCE Madison, WI
Senior Data Engineer May 2023 – Present
•Designed and implemented cloud-native data architecture using Snowflake, Python, and Airflow, reducing data processing time by 65%
•Migrated 20+ ETL workflows from on-premises to GCP cloud services, ensuring zero data loss and maintaining business continuity.
•Built and orchestrated ETL workflows using Cloud Composer (Airflow) and PySpark on DataProc, handling complex transformations over terabyte-scale structured and semi-structured data.
•Architected high-performance data lake solutions using GCP BigQuery, optimizing query performance by 40% through partitioning and clustering
•Developed automated CI/CD pipelines for data infrastructure using Terraform and Cloud Build, decreasing deployment time by 75%
•Built real-time data ingestion workflows using Pub/Sub and Cloud Functions to process 5TB+ of daily streaming data
•Implemented robust data security measures through IAM policies and encryption standards, ensuring regulatory compliance.
•Built and maintained GCP-based data pipelines using BigQuery, Cloud Storage (GCS), Pub/Sub, Cloud Run, and Cloud Functions, enabling scalable and real-time analytics across policy and claims domains.
•Designed RESTful microservices with Spring Boot to seamlessly integrate with supporting subsystems, improving data accessibility.
•Used Terraform to provision GCP infrastructure including GCS buckets, IAM roles, BigQuery datasets, and Pub/Subtopics.
•Deployed containerized ETL and ML workflows using GKE, enabling portable, scalable pipeline execution aligned with cloud-native architecture standards.
•Applied unit and integration testing for ETL logic using PyTest, improving test coverage and reducing pipeline failures during production deployments.
•Created data visualization dashboards using React and D3.js for AI monitoring applications, providing stakeholders with real-time insights.
FIDELITY INVESTMENTS Boston, MA
Data Engineer February 2021 - May 2023
•Led migration of on-premise data warehouses (Oracle, Teradata) to BigQuery, improving query performance by 70% and reducing costs by 40%
•Developed and optimized ETL workflows using Matillion, ensuring data quality and consistency across multiple systems
•Designed and implemented real-time streaming data pipelines using Kafka and Spark on Amazon EMR, processing 3M+ events per hour
•Created batch and streaming data pipelines using GCP services (Cloud Composer, DataProc, BigQuery), optimizing data transformation jobs
•Built automated data quality monitoring solutions using Python and Airflow, reducing manual QA efforts by 80%
•Integrated Spring DAO with Hibernate for efficient data access, optimizing database operations through SQL and HQL query optimization
•Implemented containerized applications using Docker and Kubernetes, improving deployment reliability and scalability
•Developed custom REST APIs using Java Spring Boot, facilitating seamless integration between microservices
SMBC New York City, NY
Data Engineer November 2019 - February 2021
•Designed and implemented cloud-based data pipelines using AWS and GCP services, increasing data processing efficiency by 55%
•Engineered scalable data warehouse structures in Snowflake, enabling faster business intelligence reporting and analytics
•Migrated legacy ETL processes to Cloud Composer (Airflow), reducing job failures by 90% and improving monitoring capabilities
•Developed PySpark jobs in AWS Glue to transform and merge data from multiple sources, enhancing data consistency
•Created monitoring systems using CloudWatch for Lambda functions and Glue jobs, ensuring 99.9% uptime for critical data pipelines
•Built Docker containers for machine learning model deployment, streamlining the ML operations workflow
•Implemented data ingestion strategies from SQL Server to S3 using Sqoop within AWS EMR, optimizing large-scale data transfers
NATIONWIDE Columbus, OH
Data Engineer August 2018 - November 2019
•Configured and maintained Hadoop ecosystem components (Hive, Pig, Sqoop, Oozie) for distributed data processing
•Developed MapReduce jobs in Java for data cleansing and preprocessing, improving data quality by 40%
•Created Power BI dashboards for business users, enabling data-driven decision making across the organization
•Built and published interactive Tableau reports, automating weekly report generation and distribution
•Optimized dashboard performance by implementing best practices in data modeling and query design
•Implemented Azure data solutions including Azure SQL Database migration and Azure Data Factory pipelines
•Provided technical training and support for business users on self-service BI tools
HSBC BANK Hyderabad, India
Data Engineer June 2015 - October 2017
•Designed and implemented BI reporting solutions using Tableau, reducing manual reporting efforts by 70%
•Built complex data pipelines using Azure Data Factory for file system and relational data ingestion
•Implemented data security policies including IAM roles, encryption, and access governance
•Architected modern data solutions using Azure PaaS services to support visualization requirements
•Migrated on-premise Microsoft SQL Server databases to Azure SQL, ensuring minimal downtime
•Designed and implemented database solutions in Azure SQL Data Warehouse, optimizing for analytical workloads
•Created ETL workflows using T-SQL, Spark SQL, and U-SQL for Azure Data Lake Analytics
ZERO TOUCH Hyderabad, India
Software Developer May 2013 - June 2015
•Developed optimized SQL and PL/SQL code for database operations, improving query performance by 30%
•Created SSIS packages for heterogeneous data export from Oracle and Excel to SQL Server
•Implemented database triggers for auditing changes to critical tables and enforcing business logic
•Designed data warehouse schemas (star and snowflake) for analytical processing
•Optimized database performance using execution plans, SQL Profiler, and tuning advisors
•Generated various reports using SQL Server Reporting Services (SSRS)
•Built and maintained ETL processes for data warehousing applications
EDUCATION
Master of Science in Computer Science Northwest Missouri State University
Bachelor of Technology in Information Technology JNTU Kakinada