Jahnavi T — Senior Data Engineer
803-***-**** **************@*****.***
PROFESSIONAL SUMMARY:
Senior Data Warehouse Engineer with 5 years of comprehensive experience in designing and implementing robust data solutions.
Expertly manage Linux-based processes and infrastructure critical for high-performance data warehousing operations.
Proficient in developing and enhancing sophisticated Shell scripts for automating ETL, data load, and extraction processes.
Skilled in optimizing data pipeline performance and implementing architecture improvements within complex data warehouse environments.
Hands-on experience with Oracle development, including database design, query optimization, and performance tuning on Exadata platforms.
Adept at utilizing Python for advanced data processing, automation, and enhancing Linux-based toolsets and jobs efficiently.
Extensive practical knowledge of Unix file systems, permissions, standard tools, and managing operating system environments.
Proven track record in enhancing ETL processes and database operations, ensuring data integrity and system efficiency.
Experienced in using orchestration tools like Airflow with Python to manage and schedule complex data workflows effectively.
Committed to Agile methodologies, driving continuous process improvement and automation initiatives across data engineering teams.
Dedicated to identifying and implementing system and architecture improvements to maximize data warehouse scalability and reliability.
Strong communicator with a passion for delivering innovative data solutions and fostering collaborative team environments.
EDUCATION:
Master of Science in Information Technology @ Indiana wesleyan University TECHNICAL SKILLS:
Programming Languages: Python, Shell Scripting, Perl, Scala, SQL
Cloud Platforms: Oracle Exadata, AWS (S3, Glue, Redshift, Lambda)
Databases: Oracle, PostgreSQL, MySQL, Snowflake, Hive
Big Data Technologies: Hadoop, Spark, Kafka
ETL Tools: Informatica, AWS Glue
Orchestration & Workflow: Apache Airflow, Jenkins
Operating Systems: Linux, Unix
Version Control: Git
Methodologies: Agile, Scrum
WORK EXPERIENCE:
Senior Data Engineer @ Elevance Health Indianapolis, IN Jun 2024 – Present
Designed and implemented scalable data warehouse architectures on Linux-based systems for critical healthcare analytics platforms.
Developed robust ETL pipelines using Python and Shell scripting to process diverse, large-scale healthcare datasets efficiently.
Managed and configured Oracle Exadata environments, optimizing database performance and storage solutions for high availability.
Enhanced various Linux-based toolsets and scripts, automating critical data ingestion and transformation processes comprehensively.
Performed advanced data modeling and optimization within Oracle databases, supporting complex reporting and analytics requirements effectively.
Implemented stringent data quality checks and validation frameworks, ensuring the highest level of data accuracy and governance across the platform.
Utilized Apache Airflow with Python for orchestrating intricate ETL workflows and managing data pipeline dependencies seamlessly and reliably.
Configured secure access controls and managed permissions within Unix file systems to safeguard sensitive healthcare information effectively.
Identified and implemented significant system and architecture improvements, enhancing overall data warehouse reliability and speed considerably.
Developed advanced SQL queries and stored procedures in Oracle for efficient data retrieval and comprehensive analysis of business insights.
Collaborated extensively with cross-functional teams to gather requirements and deliver data solutions aligned with stringent business objectives.
Contributed actively to an Agile development environment, using JIRA for efficient sprint planning and task management methodologies.
Technologies Used: Oracle Exadata, Linux, Python, Shell Scripting, Apache Airflow, Oracle, AWS (S3, Glue, Redshift), PostgreSQL, Git, Jenkins
Data Engineer @ Mastercard Purchase, NY Oct 2021 – Apr 2023
Architected and implemented high-volume data warehousing solutions on Linux platforms for complex financial transaction processing.
Developed advanced Shell scripts and Python programs to automate data ingestion, transformation, and loading into data warehouses.
Managed and enhanced Oracle Exadata environments, ensuring optimal performance and availability for critical financial data applications.
Designed and optimized ETL processes, facilitating the migration of on-premise data warehouse workloads to modern platforms efficiently.
Built robust data pipelines using Spark and Kafka, integrating diverse data sources into centralized data repositories with high throughput.
Implemented comprehensive data security measures, including data masking and encryption policies for sensitive financial information.
Utilized Airflow for orchestrating complex ETL workflows, ensuring timely and accurate delivery of financial data products reliably.
Enhanced various Linux-based toolsets and batch jobs, significantly reducing manual effort and improving operational efficiency for data operations.
Performed in-depth data validation and reconciliation for large financial datasets, maintaining high data quality standards consistently.
Collaborated with Oracle development teams to optimize database schema, indexes, and queries for improved performance and scalability.
Contributed to system and architecture improvements, focusing on scalability and reliability of the data warehouse infrastructure consistently.
Actively participated in Agile ceremonies, ensuring continuous delivery and rapid iteration of data engineering initiatives effectively.
Technologies Used: Oracle Exadata, Linux, Python, Shell Scripting, Apache Airflow, Oracle, Snowflake, Spark (Scala), Kafka, Hadoop, Hive, Jenkins, Git, MySQL
Junior Data Engineer @ Chewy Plantation, FL Nov 2019 – Sep 2021
Developed and optimized complex ETL workflows using Informatica to process large volumes of retail sales and customer data efficiently.
Designed and implemented robust data pipelines for extracting, transforming, and loading data from various relational databases systematically.
Performed comprehensive data transformation and loading into MySQL data warehouses, ensuring data consistency and integrity.
Wrote complex SQL queries and developed stored procedures for reporting and analytics purposes within retail data platforms effectively.
Created robust data validation rules and monitoring frameworks to ensure the accuracy and integrity of critical business data.
Managed and automated daily batch processing workflows using Unix shell scripting, significantly improving operational efficiency.
Maintained detailed documentation for ETL processes and data workflows, ensuring knowledge transfer and system maintainability.
Collaborated closely with business stakeholders to gather requirements and translate them into effective data engineering solutions.
Contributed to the design and implementation of data models that supported comprehensive business intelligence initiatives strategically.
Supported data migration efforts from legacy systems to new data warehouse environments, ensuring seamless transitions without disruption.
Utilized Git for version control and collaborative development of ETL code and scripting across engineering teams effectively.
Participated in an Agile SDLC environment, contributing to sprint planning, daily stand-ups, and review meetings effectively.
Technologies Used: Informatica, Unix, Shell Scripting, SQL, MySQL, Git