Post Job Free
Sign in

Data Engineer - Linux, ETL, Oracle Exadata expert

Location:
Plano, TX, 75025
Posted:
April 30, 2026

Contact this candidate

Resume:

Phaneendra Kumar Srungarapu — Data Engineer

937-***-**** **********.***********@*****.***

PROFESSIONAL SUMMARY:

Possessing over 5 years of extensive experience as a Data Warehouse Engineer, specializing in Linux-based data management and robust ETL processes.

Demonstrated proficiency in implementing, configuring, and managing critical Linux-based infrastructure to support large-scale data warehousing solutions effectively.

Expert in developing and enhancing Shell Scripts and Python-based automation for optimizing operational workflows and system performance efficiently.

Proven track record in designing and optimizing ETL/database load and extract processes, significantly improving data ingestion and transformation efficiency.

Extensive practical experience with Oracle Exadata and various relational databases, ensuring high-performance data storage and retrieval capabilities.

Adept at leveraging Informatica and Apache Airflow with Python for robust data orchestration and complex pipeline management successfully.

Strong understanding of Unix file systems, including mount types, permissions, and standard tools, critical for secure and efficient operations.

Committed to Agile methodologies and continuous process improvement, consistently delivering scalable and maintainable data solutions with precision.

WORK EXPERIENCE:

Data Engineer @ Target Corporation Minneapolis, MN Aug 2024 – Present

Implemented, configured, and actively managed robust Linux-based processes and infrastructure crucial for enterprise- scale data warehousing operations effectively.

Developed advanced Shell Scripts to automate routine administrative tasks, data extraction, and loading procedures within the Linux environment.

Engineered and optimized ETL/database load and extract processes utilizing Oracle Exadata, significantly enhancing data pipeline efficiency and throughput.

Designed and deployed data integration solutions using Informatica, ensuring seamless data flow from diverse sources into the data warehouse effectively.

Identified and implemented critical system and architecture improvements, particularly enhancing the performance of Linux-based data platforms strategically.

Enhanced various Linux-based toolsets, scripts, and scheduled jobs to streamline data processing and ensure operational reliability consistently.

Administered Oracle Exadata databases, focusing on performance tuning, query optimization, and ensuring high availability for critical data assets.

Leveraged Python for developing complex data transformation scripts and API integrations within the Linux data ecosystem proficiently.

Orchestrated intricate data pipelines using Apache Airflow with Python, scheduling and monitoring data movements across the warehouse efficiently.

Applied practical knowledge of Unix file systems, managing permissions and utilizing standard tools for secure and organized data storage.

Collaborated with cross-functional teams using an Agile methodology, providing technical expertise on data warehousing and Linux infrastructure.

Ensured data quality and consistency by implementing robust validation checks and error handling mechanisms throughout the ETL processes.

Technologies Used: Linux, Oracle Exadata, Shell Scripting, Python, Apache Airflow, Informatica, SQL, Unix File Systems, GitHub, Jenkins

Data Engineer @ Liberty Mutual Insurance Boston, MA Feb 2021 – Jul 2023

Designed and developed robust Linux-based ETL pipelines, implementing shell scripts for automating data extraction and transformation processes.

Enhanced critical ETL/database load and extract processes, leveraging Oracle databases for high-volume insurance policy and claims data efficiently.

Implemented system and architecture improvements on Linux platforms, ensuring scalable and reliable data warehousing solutions for the enterprise.

Developed and maintained Python scripts for complex data transformations and business logic within the Linux data environment effectively.

Managed and optimized Oracle databases, performing query tuning and indexing to improve data retrieval performance for analytical reports.

Utilized Informatica PowerCenter to build and maintain comprehensive data integration workflows from various source systems reliably.

Applied practical working knowledge of Unix file systems, including managing permissions and directory structures for secure data storage.

Configured and managed Linux-based infrastructure components to support the data warehousing environment effectively and efficiently.

Collaborated closely with data architects to design and implement dimensional models within the Oracle data warehouse effectively.

Implemented data quality checks and reconciliation processes to ensure accuracy and integrity of all ingested data streams.

Monitored and troubleshot Linux server processes and data pipeline failures, providing timely resolution and root cause analysis.

Participated actively in Agile Scrum ceremonies, contributing to sprint planning, daily stand-ups, and retrospective meetings consistently.

Technologies Used: Linux, Oracle, Informatica PowerCenter, Shell Scripting, Python, SQL, Unix File Systems, Azure Data Factory, GitHub

Junior Data Engineer @ Molina Healthcare Long Beach, CA Jul 2019 – Jan 2021

Supported and enhanced Linux-based data warehouse processes for healthcare claims and member data, ensuring operational continuity effectively.

Developed and maintained Informatica ETL workflows, sourcing critical data from Oracle databases for comprehensive reporting requirements.

Created complex SQL and PL/SQL scripts to perform data extraction, validation, and reconciliation within the Oracle environment efficiently.

Assisted in ingesting large healthcare datasets into on-premise data warehouse systems, including robust Linux servers.

Implemented Shell Scripts to automate data loading tasks and directory management on Unix file systems securely.

Performed data quality checks and resolved discrepancies, ensuring high data integrity for regulatory compliance standards.

Worked with Apache Hadoop and Hive for processing and analyzing historical healthcare data within a distributed Linux environment.

Contributed to initial assessments for migrating on-premise data infrastructure to cloud platforms, planning future transitions.

Enhanced various Linux-based toolsets and processes to improve efficiency of data processing operations consistently.

Utilized GitHub for version control of SQL scripts and Informatica mappings, ensuring collaborative development practices.

Followed Agile SDLC, actively participating in sprint planning, development, and review cycles effectively.

Maintained comprehensive documentation for ETL workflows and database schemas using Confluence, ensuring clarity and accessibility.

Technologies Used: Oracle, Informatica PowerCenter, SQL, PL/SQL, Linux, Shell Scripting, Apache Hadoop, Apache Hive, Unix File Systems, GitHub, Jenkins

TECHNICAL SKILLS:

Programming Languages: Python, Shell Scripting, SQL, Perl, PL/SQL

Data Warehousing: Oracle Exadata, Informatica PowerCenter, Snowflake, Apache Hive, Azure Synapse Analytics, Redshift, Data Lake

Operating Systems & Tools: Linux, Unix File Systems, Standard Unix Tools, Pipes, SSH

ETL & Orchestration: Apache Airflow, Informatica, Azure Data Factory, AWS Glue

Big Data Technologies: Apache Spark, Databricks, Apache Hadoop, Apache Kafka

Database Management: Oracle, PostgreSQL, SQL Server, MySQL

Version Control & CI/CD: GitHub, Jenkins

Business Intelligence: Power BI

Methodologies: Agile (Scrum)

EDUCATION:

MS in Computer Science @ University of Dayton



Contact this candidate