Rasagna Akula — Senior Data Engineer
470-***-**** **************@*****.***
PROFESSIONAL SUMMARY
Results-driven Data Warehouse Engineer with around 5 years of experience specializing in robust Linux-based data infrastructure and scalable data processing solutions.
Expertly implement, configure, and manage critical Linux processes and infrastructure for high-performance data warehousing environments.
Proven ability to identify and implement system and architecture improvements, enhancing overall data platform efficiency and reliability.
Highly proficient in Shell Scripting for automating complex data tasks, managing Unix file systems, and optimizing operational workflows.
Extensive practical experience with relational databases, including Oracle Exadata, for designing, developing, and enhancing ETL/database load processes.
Skilled in Python and Perl for data engineering, scripting, and developing advanced solutions to improve data quality and pipeline performance.
Experienced in utilizing ETL tools like Informatica and orchestration tools such as Apache Airflow with Python for streamlined data operations.
Adept at applying Agile methodologies to project delivery, ensuring continuous process improvement and rapid adaptation to evolving data requirements.
Passionate about automation and continually optimizing data workflows to maximize efficiency and maintain data integrity within warehousing systems.
EDUCATION
Master of Science in Computer Science @ University of Alabama at Birmingham TECHNICAL SKILLS
Operating Systems: Linux, Unix
Scripting & Automation: Shell Scripting, Python, Perl, Apache Airflow, Docker, Jenkins
Database Management: Oracle Exadata, PostgreSQL, MySQL, Snowflake, Hive, SQL
ETL & Data Warehousing: Informatica PowerCenter, AWS Glue, Azure Data Factory, Spark, Hadoop, Data Lake Storage, Data Modeling
Cloud Platforms: AWS (S3, EMR, Lambda, Redshift), Azure (ADLS, Synapse)
Version Control: Git
Methodologies: Agile, Scrum
Business Intelligence: Power BI
WORK EXPERIENCE
Senior Data Engineer @ Regions Financial Corporation Atlanta, GA, USA Sep 2024 – Present
Architected and implemented robust Linux-based data warehousing solutions, ensuring high availability and scalability for critical data assets.
Developed intricate Shell Scripts to automate data ingestion, transformation, and loading processes within a production Linux environment.
Engineered and optimized ETL pipelines using Informatica PowerCenter to integrate diverse data sources into Oracle Exadata.
Managed and enhanced Oracle Exadata databases, including schema design, performance tuning, and complex SQL query optimization for data warehousing needs.
Implemented system and architecture improvements, enhancing existing Linux-based toolsets and processes to boost data processing efficiency.
Utilized Apache Airflow with Python to orchestrate complex data workflows, ensuring timely and reliable data delivery to downstream systems.
Developed advanced Python scripts for data validation, quality checks, and custom data transformations within the data warehouse.
Collaborated with cross-functional teams using Agile methodologies to deliver continuous process improvements and support BI reporting requirements.
Technologies Used: Linux, Shell Scripting, Oracle Exadata, Informatica PowerCenter, Apache Airflow (Python), SQL, Git, Jenkins
Data Engineer @ Vanda Pharmaceuticals Atlanta, GA, USA Nov 2022 – Aug 2024
Designed and deployed data integration solutions within a Linux environment, supporting enterprise data warehousing initiatives.
Developed and maintained Shell Scripts to manage Unix file systems and automate routine data management and extraction tasks.
Constructed and optimized ETL pipelines using Informatica for ingesting and transforming high-volume data into relational databases.
Contributed to Oracle database development, writing complex SQL queries and stored procedures to support data warehouse operations.
Implemented data quality checks and validation frameworks using Python to ensure accuracy and integrity of data in the warehouse.
Enhanced existing ETL/database load processes, identifying bottlenecks and implementing improvements for increased efficiency.
Collaborated on developing and enhancing various Linux-based toolsets, scripts, and jobs to streamline data processing workflows.
Adhered to Agile principles throughout the software development lifecycle, ensuring iterative improvements and timely project delivery.
Technologies Used: Linux, Shell Scripting, Informatica PowerCenter, Oracle, Python, SQL, Git, JIRA Data Engineer @ Easy Buy Bengaluru, India Jun 2021 – Jul 2022
Developed and maintained batch processing pipelines within a distributed Hadoop ecosystem running on Linux servers.
Processed large datasets using Spark (Scala/Python) for complex transformations and aggregations to feed data warehouses.
Constructed a data warehouse using Snowflake, designing efficient schemas and optimizing data loading processes.
Utilized Shell Scripting for automating job scheduling and managing data directories across Unix file systems.
Optimized SQL queries and HiveQL scripts to enhance data retrieval performance from both Snowflake and Hive data stores.
Implemented data ingestion strategies from various sources, ensuring data quality and consistency for reporting purposes.
Performed thorough data quality checks and developed validation routines to maintain integrity of warehousing data.
Participated actively in Agile Scrum ceremonies, contributing to continuous improvement of data engineering practices. Technologies Used: Hadoop, Spark (Scala, Python), Snowflake, Hive, Linux, Shell Scripting, MySQL, Git, Oozie Junior Data Engineer @ Acko Insurance Bengaluru, India Mar 2020 – May 2021
Developed and deployed robust ETL workflows using Informatica PowerCenter to integrate critical business data from disparate sources.
Designed and implemented relational data models in MySQL, ensuring optimal structure for data warehousing and reporting.
Wrote and optimized complex SQL queries and stored procedures for data extraction, transformation, and loading operations.
Performed extensive data cleansing, validation, and profiling to ensure high data quality within the target databases.
Contributed to the enhancement of ETL/database load processes, identifying areas for automation and efficiency gains.
Scheduled and monitored data integration jobs using Control-M, ensuring timely delivery of processed information.
Collaborated on implementing new data integration patterns, focusing on process improvement and maintainability.
Maintained comprehensive documentation for ETL processes and data flows, supporting knowledge transfer and system understanding.
Technologies Used: Informatica PowerCenter, MySQL, SQL, Control-M, Git, Linux