Post Job Free
Sign in

Senior Data Engineer - ETL, Oracle, Airflow, Python

Location:
Plano Park, TX, 75074
Salary:
120000
Posted:
April 30, 2026

Contact this candidate

Resume:

Gokul Gude — Senior Data Engineer

972-***-**** **********@*****.***

PROFESSIONAL SUMMARY

A highly experienced Data Warehouse Engineer with around 5 years of experience, specializing in Linux-based data processes and infrastructure management.

Expert in developing and enhancing ETL processes, leveraging strong shell scripting capabilities for robust data warehousing solutions.

Proven proficiency in Oracle database development and optimization, including schema design, query tuning, and stored procedure creation.

Skilled in implementing system and architecture improvements for data warehousing, ensuring scalability, reliability, and performance.

Adept at enhancing Linux-based toolsets, scripts, jobs, and processes to streamline data operations and achieve automation goals.

Extensive experience with Python for data engineering, scripting, and developing complex data pipelines, enhancing efficiency across systems.

Practical working experience with leading ETL tools such as Informatica and orchestration tools like Airflow for end-to- end workflow management.

Deep understanding of Agile methodologies, contributing effectively to project teams and ensuring timely delivery of critical data initiatives.

Passionate about automation and continual process improvement, consistently seeking opportunities to optimize data warehouse operations.

EDUCATION

Master of Science in Computer Science @ The University of Texas at Arlington TECHNICAL SKILLS

Operating Systems & Scripting: Linux, Unix, Shell Scripting, Bash, Perl

Programming Languages: Python, Scala

Databases & Data Warehousing: Oracle, Oracle Exadata, PostgreSQL, MySQL, Hive, Snowflake, Amazon Redshift, Azure Synapse Analytics, MS SQL Server

ETL & Orchestration: Informatica PowerCenter, Apache Airflow, AWS Glue, Azure Data Factory

Big Data Technologies: Hadoop, Spark, PySpark, Kafka, Delta Lake

Cloud Platforms: AWS (S3, EMR, Lambda), Azure (ADLS, Databricks)

DevOps & Version Control: Git, GitHub, Jenkins, Docker

BI & Reporting: Tableau, Power BI

Methodologies: Agile, Scrum

WORK EXPERIENCE

Senior Data Engineer @ Tenet Healthcare Dallas, TX Sep 2024 – Present

Managed and configured robust Linux-based infrastructure to support advanced data warehousing solutions for critical healthcare analytics.

Developed sophisticated shell scripts to automate complex ETL processes, significantly enhancing data extraction, transformation, and loading efficiency.

Designed and optimized Oracle database schemas and stored procedures, ensuring high performance for large-scale healthcare data storage and retrieval.

Implemented Informatica PowerCenter for developing and integrating high-performance data pipelines across diverse healthcare source systems.

Enhanced existing ETL and database load/extract processes using Python scripting, achieving measurable improvements in processing speed and reliability.

Orchestrated intricate data workflows using Apache Airflow DAGs, guaranteeing timely delivery of processed data for reporting and operational insights.

Identified and implemented key system and architecture improvements, focusing on the scalability and resilience of data warehousing environments.

Collaborated effectively with cross-functional teams within an Agile framework to deliver impactful data solutions that met evolving business requirements.

Managed source code using Git and implemented Jenkins CI/CD pipelines for automated deployments across various Linux environments.

Technologies Used: Linux, Shell Scripting, Oracle, Informatica PowerCenter, Python, Airflow, PySpark, Kafka, Tableau, Hadoop, Hive, Git, Jenkins

Data Engineer @ Citigroup New York, NY Mar 2021 – Jul 2023

Implemented and managed Linux-based environments crucial for enterprise-level data warehousing initiatives supporting comprehensive financial analytics.

Developed robust shell scripts for automating critical data refresh, validation, and maintenance tasks within the financial data warehouse.

Designed and optimized Oracle database solutions, including schema design and query tuning, for efficient management of vast financial data.

Utilized Informatica for developing and enhancing ETL processes, facilitating the migration of on-premise financial data into modern data warehousing solutions.

Enhanced database load and extract processes using Python, improving overall efficiency and significantly reducing processing times for large datasets.

Contributed actively to system and architecture improvements, focusing on data warehouse performance and reliability within a hybrid environment.

Ensured stringent data integrity and security within the data warehousing framework by implementing comprehensive validation and access control measures.

Collaborated extensively with development teams to seamlessly integrate data solutions, leveraging Agile methodologies for timely project delivery.

Version controlled code using GitHub and implemented logging and monitoring solutions using tools like Log4j for enhanced visibility.

Technologies Used: Linux, Shell Scripting, Oracle, Informatica, Python, SQL Server, Power BI, Azure Data Factory, Azure Databricks, GitHub

Junior Data Engineer @ Big Lots Columbus, OH Nov 2019 – Feb 2021

Developed and optimized complex ETL workflows using Informatica PowerCenter for efficient retail data integration and processing.

Designed and implemented robust data models in Oracle database environments, supporting critical sales, inventory, and supply chain reporting.

Created and enhanced intricate SQL queries and stored procedures, significantly improving data extraction and load performance for large datasets.

Utilized Unix shell scripting for automating daily batch processing tasks and managing data movement efficiently across various systems.

Implemented rigorous data validation and cleansing rules within ETL processes, ensuring high data quality for downstream business intelligence applications.

Contributed to the architectural design of data pipelines for sales and inventory reporting, integrating diverse source systems effectively.

Managed and performed extensive data migration from legacy systems to Oracle databases, ensuring complete data consistency and integrity.

Collaborated closely with business teams to gather precise requirements and deliver robust data warehousing solutions using Agile principles.

Managed version control using Git and implemented Jenkins for deployment automation within the data processing workflows.

Technologies Used: Informatica PowerCenter, Oracle, SQL, Unix Shell Scripting, Hive, Git, Jenkins



Contact this candidate