Venkata Susheel Grandhi — Senior Data Engineer
913-***-**** **************@*****.***
PROFESSIONAL SUMMARY
Highly accomplished Senior Data Engineer with 5 years of experience specializing in architecting and optimizing Linux- based data warehousing solutions and infrastructure.
Expertise in developing robust ETL/database load/extract processes using advanced Shell Scripting and Oracle development for high-performance data operations.
Profound practical working experience in setting up and managing complex Linux environments, including deep knowledge of Unix file systems and standard tools.
Skilled in Python and familiar with Perl, leveraging these languages for enhancing Linux-based toolsets, automation scripts, and data processing jobs.
Proven ability to identify and implement significant system and architecture improvements, ensuring scalable and efficient data warehousing infrastructure.
Extensive experience with relational databases, specifically Oracle Exadata, for designing and optimizing critical data storage and retrieval mechanisms.
Adept at utilizing orchestration tools like Apache Airflow with Python for managing intricate data pipelines and ensuring seamless workflow automation.
Demonstrated proficiency in ETL tools, including Informatica, for designing, developing, and deploying enterprise-level data integration solutions.
Committed to Agile methodologies, continuous process improvement, and automation, delivering high-quality, maintainable, and scalable data solutions.
EDUCATION
Master of Science in Computer Science @ University of Kansas TECHNICAL SKILLS
Programming Languages: Python, Perl, SQL
Scripting & OS: Shell Scripting, Unix, Linux, Bash, AWK, SED
Data Warehousing: Oracle Exadata, Informatica PowerCenter, Apache Airflow, Snowflake, Redshift
Databases: Oracle, PostgreSQL, MySQL, Hive, HDFS
Cloud Platforms: AWS (S3, Glue, EMR, Lambda, Redshift, Athena)
Big Data Technologies: Apache Spark (PySpark), Hadoop, EMR
Version Control & DevOps: Git, Jenkins, Docker
Data Visualization: Tableau
Methodologies: Agile, Scrum
WORK EXPERIENCE
Senior Data Engineer @ Tailored Brands, Inc. Houston, TX Jan 2025 – Present
Implemented, configured, and managed critical Linux-based processes and infrastructure for enterprise data warehousing solutions, ensuring high availability.
Developed and enhanced complex Shell Scripts to automate critical ETL/database load/extract processes, significantly improving data ingestion efficiency and reliability.
Architected and optimized data warehouse solutions using Oracle Exadata, ensuring high performance and reliability for large-scale data operations and analytics.
Designed and implemented robust data pipelines using Informatica PowerCenter, integrating diverse data sources into the central data warehouse effectively.
Leveraged Apache Airflow with Python to orchestrate end-to-end data workflows, ensuring timely and automated delivery of critical business intelligence data.
Identified and implemented significant system and architecture improvements within the Linux environment, enhancing overall data platform scalability and performance.
Administered Unix file systems, demonstrating practical knowledge of mount types, permissions, and standard tools to ensure data integrity and security.
Collaborated with cross-functional teams to define data warehousing requirements, translating complex business needs into precise technical specifications and data models.
Enhanced existing Linux-based toolsets and jobs using Python and Shell Scripting for continuous process improvement and operational efficiency across the data platform.
Managed data governance and quality frameworks within the Oracle data warehouse, ensuring accuracy and compliance for all loaded data assets.
Utilized Git for version control and integrated CI/CD pipelines with Jenkins for automated deployment of data warehousing components and configurations.
Worked within an Agile development methodology, actively participating in sprint planning and daily stand-ups to deliver iterative and high-quality solutions.
Technologies Used: Oracle Exadata, Informatica PowerCenter, Linux, Shell Scripting, Python, Apache Airflow, SQL, Git, Jenkins, Snowflake
Data Engineer @ Ally Bank Kansas City, MO Jan 2024 – Dec 2024
Designed and implemented scalable ETL/database load/extract processes using Shell Scripting and Python within a robust Linux environment.
Configured and maintained Linux-based infrastructure for data warehousing initiatives, ensuring optimal performance and resource utilization for critical systems.
Developed and optimized Oracle database schemas and stored procedures, enhancing data retrieval speeds and overall query performance for reporting.
Managed large datasets within relational databases, applying practical knowledge of Unix file systems and permissions for secure and efficient data handling.
Leveraged Informatica for designing and deploying complex data integration workflows, automating data ingestion from various source systems effectively.
Orchestrated critical data pipelines using Apache Airflow with Python, improving data freshness and reliability for analytical reporting and dashboards.
Implemented comprehensive data quality checks and validation frameworks across ETL processes, maintaining high standards for data accuracy and completeness.
Contributed to system and architecture improvements, specifically enhancing Linux-based toolsets and automated processes for data transformation.
Troubleshot and resolved complex data warehousing issues, utilizing advanced Shell Scripting and SQL for efficient problem diagnosis and resolution.
Collaborated closely with data analysts and business stakeholders to understand data requirements and deliver tailored, high-impact data solutions.
Employed Git for version control and integrated solutions with Jenkins for continuous integration and automated deployment in an Agile setting.
Participated in Agile sprints, adhering to best practices for development, testing, and deployment of robust data engineering solutions.
Technologies Used: Oracle, Informatica PowerCenter, Linux, Shell Scripting, Python, Apache Airflow, SQL, Git, Jenkins, PostgreSQL
Junior Data Engineer @ Accenture Hyderabad, India Jul 2020 – Dec 2022
Developed and maintained ETL pipelines using Python and SQL to process diverse datasets within a Hadoop ecosystem on Linux servers.
Ingested and transformed large volumes of data from various sources, including MySQL databases, into Hadoop HDFS for big data analytics.
Executed complex data processing tasks using Hive queries, optimizing them for performance and resource utilization across distributed systems.
Implemented Linux-based scripts for automating data loading and transformation processes, enhancing operational efficiency and reliability.
Designed and developed batch processing workflows, ensuring timely and accurate data availability for reporting and analytical purposes.
Utilized practical knowledge of Unix file systems for managing data storage and access permissions within the Hadoop environment effectively.
Performed extensive data transformation and validation, ensuring data quality and integrity throughout the data lifecycle stages.
Developed monitoring scripts in Shell and Python to proactively identify and resolve issues within the data processing pipelines promptly.
Leveraged Apache Oozie for scheduling and managing complex Hadoop jobs, automating workflows for data ingestion and processing.
Collaborated with data scientists and analysts to support their data requirements, providing cleansed and structured datasets for their models.
Maintained thorough documentation for all developed ETL processes and data models, facilitating knowledge transfer and system understanding.
Adhered to Agile methodologies, contributing to sprint reviews and retrospectives to foster continuous improvement in development practices.
Technologies Used: Hadoop, Hive, HDFS, Linux, Shell Scripting, Python, SQL, MySQL, Apache Oozie, Git Data Analyst Intern @ Dolf Technologies Hyderabad, India Jan 2020 – Jun 2020
Designed and developed ETL workflows using Informatica PowerCenter to extract, transform, and load data from Oracle databases efficiently.
Created and optimized complex SQL queries and stored procedures within Oracle to support data transformation and aggregation needs.
Implemented data cleansing and transformation rules within Informatica mappings, ensuring high data quality for downstream systems.
Developed and maintained data mappings and workflows in Informatica PowerCenter, adhering to best practices for data integration.
Monitored ETL job executions and proactively resolved performance issues within the Informatica and Oracle environments swiftly.
Implemented data validation and reconciliation processes to ensure accuracy and completeness of loaded data into the data warehouse.
Utilized Unix shell scripting for automating various operational tasks, including file transfers and job scheduling effectively.
Collaborated with database administrators to optimize Oracle database performance, supporting efficient data warehousing operations.
Gained foundational experience with basic Hadoop components for initial data storage and processing exploration and analysis.
Contributed to data modeling efforts, assisting in the design of star and snowflake schemas for analytical reporting requirements.
Maintained comprehensive documentation for ETL processes, data flows, and database schemas, ensuring clarity and consistency.
Used Git for version control of ETL scripts and mappings, supporting collaborative development within an Agile team environment.
Technologies Used: Informatica PowerCenter, Oracle, SQL, Unix, Shell Scripting, Hadoop, Git