Ravi Marada
Sr. Data Engineer
********@*****.***
913-***-**** (c)
Experience Summary:
14+ years of IT experience in Administration, Analysis, Design, Development, Testing and Implementation of business application systems in Data warehousing for Industrial sectors ranging from Banking, Retail, Communication, Insurance and HealthCare.
Over 12 years of experience in Redshift, Teradata, ETL methodologies for supporting data extraction, transformations and loading processing through Teradata, IIS, SSIS and DM Express.
4+ years of Databricks Engineering and administration focusing on Workspace, catalog and workflow management, data ingestion, and ETL frameworks.
Worked on Databricks catalog, workspace, schema and compute structuring.
5+ years of AWS Redshift Data engineering, development and administration.
10 plus years of Teradata ETL Development, Administration, Teradata SQL Assistant and data load extract and export utilities of Teradata like BTEQ, Fast Load, Multi Load, Fast Export and Unix shell scripting.
Proficient in Redshift Performance Tuning through PI choice, statistics, query re-writing, skew reduction, compression, and different types of partitioning including Columnar and debugging of existing ETL process.
Leveraged Teradata AWS S3 connection to deep archive older data thereby reduce storage cost on Teradata nodes.
Provided technical solutions and guidance for Business Analysts and Functional requirements. Performed Impact analysis on systems, created data models for projects, and designed logical models and development environments.
Articulated business requirements into technical details through Data Modelling with different Schemas (Star and Snowflake) to fit reporting, query and business analysis requirements.
Designed and Maintained database systems according to Analytical and Reporting requirements using Power BI, Microstrategy, Congos, QlikSense and QlikVeiw.
Integrated Teradata Hadoop HDFS, and manage Hadoop data lake ecosystem like Hive, HBase, and Sqoop through Cloudera Manager.
Worked in all stages of software development cycle, namely administration, design, specification, coding, debugging, testing (test plan and test execution), documentation and program maintenance.
Followed Agile methodology, document and maintain project activities with timelines through Jira and Confluence.
Strong Communication, Presentation and Interpersonal skills with excellent problem-solving capabilities.
Certifications: Teradata Certified Professional
Informatica Power Center 8.x Mapping Design
Education: Masters in Engineering, Gannon University, Erie, PA Batch 2007 - 2009
Bachelors in Engineering, Andhra University, India, Batch 2002 - 2006.
Technical Skills:
Databases
AWS Redshift, Teradata 16.20, 15.1, 14.1, 13.1, V2 R12, SQL Server 2005/2008/2008R2/2012, Oracle 11g, 10g,
Reporting Tools
Power BI, Microstrategy 9.4.1, Cognos, OBIEE 11.1.x, Qlik Sense 13.32.84
Tools
Cloudera Hadoop, Cloudera Manager, Informatica Power Center 10, 9.5 / 8.1, SQL Developer, IBM DataStage, TOAD, HP Quality Center 10, DM Express 7.5, SSIS,, CA Autosys, Crontab,Veritas NetBackup, HP ALM Tool, Python
Data Modeling Tool
E R Studio 16, Erwin 9.5/4.0, MS Visio, Enterprise Architect 4.5
Operating Systems
Windows, Linux, Unix Shell Scripting
Cloud Computing
Databricks, AWS, S3, IAM, EC2, Snowflake Virtual Warehouse
Professional Experience:
Client : Bluestem Brands Tampa FL March 2022 - Present
Position : Sr. Databricks Engineer and Redshift Engineer
Served as Sr. Databricks and Redshift engineer with Bluestem - Finger Hut Brands.
Set-up Databricks catalogs, workspaces and schemas for SQL server and redshift migration projects.
Implemented Databricks security permissions on catalog, group and schema levels using GIT Lab deployments.
Worked on ETL processes for Data ingestion and data export with to and from Databricks.
Handled various ETL process in Databricks including Cloud Object Stores, SaaS Applications and Relational Databases such as Redshift.
Managed loads involving Structured, Unstructured and Semi-structured data such as Json files.
Worked as cluster administrator in databricks with catalog, schemas, compute policies, groups creations and deploying permissions using DevOps pipelines such as CD/CD pipelines.
Application of Terraform code for code deployments in Databricks.
Implemented cluster re-size by addition of new nodes to increase to 12 node cluster.
Managed with redshift cluster for Production and DEV environments.
Extensively managed security related redshift permissions include system, schema and table levels.
Configured workload management groups, cluster parameter groups and queues using AWS console.
Handled back-up and recovery processes using Snapshot processing and scheduling.
Utilized DevOps Integration workload usage with CI/CD pipelines for version controlling and continuous deployments Redshift scripts and Store procedures.
Worked Continuous integration and Continuous Deployment using CI/CD Pipelines through Liquibase and GIT Lab.
Implemented multiple methodology implementation for performance optimization through, analyzing, vacuuming and Concurrency scaling clusters for optimal usage of the Redshift clusters.
Worked on query optimization, query performance using read and write split performance techniques and query tuning of complex queries.
Working on training on Spark and Pyspark coding for batch and streaming load jobs
Client : AbbVie Inc. Waukegan IL Feb 2018 – March 2022
Position : Sr. Redshift and Teradata Engineer
Worked creating Redshift Data shares, External tables, Snapshot creations and Data refreshes between different environments such as Dev, QA and PROD.
Worked on Databricks standard operation procedures for the CDC data injection process and DQ check process, CDC jobs execution status, sandbox table retention process and CDC landing table retention process from SQL Server to AWS platform.
Implemented Enhancement to existing Databricks jobs with respect to Time zone formatting in ETL pipelines for Landing layer.
Worked on new Node additions to current Redshift cluster and upgrade node type of current clusters.
Troubleshooting Redshift issues due to connections, configurations, client tool set up for developers, policy and permissions.
Supporting Redshift daily operations such monitoring memory, grants and permissions, locking, blocks and performance issues.
Automate processes on crontab for privilege user annual recertification and terminated users report Redshift audits using unix shell scripting and python scripts for ETL automations.
Worked on Snowflake platform POC with solid understanding of virtual warehouses, hybrid model of shared-disk and shared nothing architecture of snowflake for compute processing of queries using MPP compute clusters.
Provide technical guidance for integration, testing, design, development, planning of new major, large scale, production systems/databases.
Perform daily maintenance with Teradata tools and utilities such as Viewpoint, PDCR, DBQL monitoring, performance analysis using PDCR tool kit, troubleshooting, problem resolution, and user assistance for Teradata Platform and Teradata Client Setup.
Work with Backup and Restore activities using Data Stream Architecture or DSA.
Manage and maintain all Teradata production and non-production databases and refresh or restore data from one environment to other using DSA.
Responsible for Teradata standards and design of physical data storage, maintenance, access and implement security administration through Teradata Administrator or Teradata Studio.
Plan and maintain workloads through TASM or Teradata Workload management.
Responsible for space management, back-up, monitoring, and maintaining ongoing performance of Teradata platform.
Participate as Subject Matter Expert (SME) as required for design reviews and new technology.
Working knowledge of Cloud computing platforms, with prime focus on AWS and fundamentals of Azure and GCP platforms.
Use Teradata ticketing system of Teradata Support for follow up on issue resolution, incidents and database hardware maintenance activities such as upgrades, patches and change controls.
Client : CVS Health Care Dallas TX Mar 17 –Jan 18
Position : Sr. Teradata Engineer
Design, create and regular tuning of physical database objects (tables, views, store procedures and indexes).
Performing DBA activities include creation of users, spool, temporary, permanent space, checking the tables skewing, table compression, troubleshooting the failure jobs
Composed Automatic jobs for day-to-day admin activities to report on Objects access, and space usage through unix shell scripts and automat them in crontabs.
Worked on metadata management using MDM tools and worked on data transformation on current workloads and work on dependency data structures.
Responsible for change control and Release of Enterprise data models for data warehouse subject areas such as Claims, Member and Eligibility.
Provide documentation to offshore DBA's to monitor system during their shifts and provide 24/7 On-Call support.
Client : New York and Company New York NY May 16 –Dec 16
Position : Sr. Teradata Engineer and Data Modeler
Roles & Responsibilities:
•Worked on building new Data Warehouse for New York and Company retail domain.
•Worked on Developing KPI Dashboard project which serves as one of the cost reduction areas by eliminating wastage in sample order.
•Worked on POC for reports generation using basic Python scripting.
•Capacity planning: Properly identify required hardware, software.
Client : Wells Fargo Fremont CA Mar 15 –Apr 16
Position : Teradata DBA
Roles & Responsibilities:
Working with Regulatory Data Reporting project for banking domain.
Provided Design, Created and Regular Tuning of Physical database objects of Regulatory Data Repository tables, views, indexes to support OBIEE Reports running of Teradata database.
Proactive monitoring Teradata database space utilization and regular clean-up of non-usage objects to release space.
Testing of Stored Procedures and Functions, Unit and Integration Testing of Informatica Sessions, Batches and Target Data.
Client : Bancvue Ltd Austin TX Jan 13 – Feb 15
Position : Teradata DBA
Roles & Responsibilities:
RateWatch Project database tables are designed, create and tuning of database objects such as EDW tables, views, indexes. Also created Join Indexes for faster retrieval of data.
Extensive use of ETL Informatica Tools like Designer in RCX project of transactional data for creating mappings, develop Workflow Manager and monitor the mapping runs in Workflow Monitor.
Working with Release management for test and plan environments with data seeding and push down optimization techniques to improve performance of mappings such as source, target and full Push down optimizations.
Working with ETL leads to formulate ETL approach and appropriately uses Teradata Tools and Utilities.
Client : LORD AND TAYLOR St Louis MO Sep 11 – Dec 12
Position : Teradata ETL Consultant
Client : NIKE Beaverton OR Aug 10 -Aug 11
Position : Teradata Consultant