Tom Jose
**********@*****.***
Oakmont, PA
Professional Summary:
* ***** ** ** ************ experience in analysis, design, development and testing of enterprise grade applications.
8 years of strong SQL scripting and ETL workflow testing experience in both RDBMS and Big Data environment.
Worked on ETL data validation developed using Informatica /SSIS/ Pentaho Data Integration (PDI) ETL environments
Strong data validation experience in Hadoop ecosystem, such as Cloudera distribution, Apache Oozie workflow, Hive and Impala.
Developed and executed detailed ETL related functional, performance, integration and regression test cases, and documentation
Experienced in an Agile environment and team focusing on sprint by sprint deliveries.
Worked with end users to formulate and document business requirements.
Strong knowledge on ETL and BI processes
Worked on Data Analytics and BFS domains.
Exposure of working closely with Customers, Stake Holders and Client Partners.
Working knowledge of Defect management process, Risk Framework and Mitigation, and Enterprise Testing standards
Strong problem solving and technical skills coupled with clear decision making.
Technical Skills:
Databases
MS Sql, Oracle 11g, Netezza
ETL and BI Tools
SSIS, SSRS, Pentaho Data Integration (PDI), Informatica
Manual Testing
Functionality testing, GUI testing, Integration testing, Regression testing, system testing, Bugzilla
Platforms
Linux, Windows
Scripting Languages
Unix shell scripting
Big Data Ecosystems
Cloudera cloud data platform, HDFS, Hive, Impala, Pentaho Data Integration(PDI), Oozie
Professional Experience:
Organization
Duration
Tata Consultancy Services,
2011 – 2020
Ernst Young, Kerala, India
2010 – 2011
Work Experience:
Project I: Digital Advertisement Rating
Employer: Tata Consultancy Services, India
Client: Nielsen Corporation, USA
Role: ETL and Big data Testing Lead
April 2017 – July 2020
Gross Rating Point calculation for advertising campaigns in digital media such as Desktop computers, Mobiles, Tablets, etc. Data arrives from various providers and Panelists. Then after scaling with complex methodology, store the numbers in warehouse. The summary of results are being exported to Oracle reporting DB and shared to stakeholders through different reporting tools.
Responsibilities:
Understand the analytical methodologies to generate test cases to validate the results from each statistical methodologies
Worked with Hortonworks Distribution and Cloudera Distribution.
Performed data analytical testing for the BI systems
Created analytical SQL scripts to validate data between Netezza and Hive.
Validation of data transformations and perform End-to-End data validation for ETL & BI systems
Analyze and understand the ETL workflows developed with Oozie in HDFS cloud system
Generate UNIX shell scripts to manipulate test data and move data from different environment to meet the testing requirements.
Experienced in Agile Methodology and expertise in JIRA software
Design, develop, and test processes for extracting data from legacy systems, production databases, Hadoop, and Open Source platforms
Team size: 6
Technologies Oracle 11g, Netezza, Cloudera, Hive, Apache Spark, Tibco, Shell Script.
Project II: Data Integration and Disaster Recovery
Employer: Tata Consultancy Services, India
Client: JPMorgan Chase, USA
Role: ETL Tester, Big data Functional Tester
2015 – 2016
Being a financial services major, this critical business continuity plan - Disaster Recovery Implementation was to switch to another active data center in case of a disaster. Implemented Control Failover to swap between clusters.
Responsibilities:
Framed the basic design of Disaster Recovery system based on customer inputs
Designed the different approaches for HBase Replication due to their data complexity; and implemented the most suitable approach
Addressed the testing concern of synchronizing the current Production system with DR, in terms of reference & operation metadata
Integrated DR system well with other attached systems (Process orchestration engine and ETL jobs)
Tested Control Failover, to swap between the clusters
Created ETL test cases to validate data between upstream and downstream systems using PDI and Hive Queries (HQL).
Took the ownership of Unit Testing, SIT and ITSM Test Plan documents for UAT
Has been involved in client reviews for test plans and deployment plans
Team Size: 2
Technologies: Shell Script, HBase, Hive, Pentaho Data Integration (PDI).
Project III: Columbia Integration
Employer: Tata Consultancy Services, India
Client: Ameriprise Financial, USA
Role: ETL Tester
2011 – 2015
This project was to integrate asset management reports from their recently acquired bank, to the source bank, using Informatica and MSBI technologies.
Responsibilities:
Provided test case inputs and assisted with production validations
Performed ETL testing using Informatica (Power Center/ Power Mart) (Designer, Workflow Manager, Workflow Monitor and Server Manager)
Strong in ETL data validation developed using Informatica / SSIS ETL environments
Exposure to DB tools: Toad/ PL SQL developer/SQL Plus
Strong in BI report validation developed using SSRS BI
Designed and developed ETL test cases, scenarios, and scripts to ensure quality Data warehouse / BI applications
Experienced on ETL automation tool – QuerySurge
Developed process flows and documentation
Team Size: 5
Technologies: MSBI, INFORMATICA
Project IV: Info-web Application (Internal Financial Application)
Employer: Ernst Young, India
Role: Application Support Engineer
2010 – 2011
Info-web Application is an internal financial application, supported 24*7, and built on .Net and MS SQL. Provided level 2 production application support, based on strict SLA.
Tools: SSRS, SQL Server 2008 R2
Educational Qualification:
Master of Computer Applications (MCA)
Bharathiar University, Coimbatore, Tamilnadu, India
Visa:
H4 EAD valid until August 2022.