Post Job Free

Resume

Sign in

Web Project Data Analyst

Location:
Phoenix, AZ
Posted:
July 18, 2021

Contact this candidate

Resume:

Janarthanan Thoppay Bavanisankar CSM adnstn@r.postjobfree.com 602-***-****

Technical Lead - DWH/BI

Technical Lead with over 16+ years of progressive experience in Data warehousing in Life sciences / Pharmaceutical domain and Banking & Finance domain of IT industry.

Strong Technical experience in data warehousing tools Ab-Initio, Ab-Initio BRE, Bigdata, SSIS, Talend and strong experience in programming languages - Core Java, Linux scripting and Python.

Strong experience in ETL tool using Ab-Initio (with coop version 1.15 to 3.3.5.3) and Ab-Initio products Metadata Hub, BRE. Experience in the different Source / Target systems like Flat files, Multi-file systems (MFS), Ab-Initio queues (Continuous flow Graphs), and Relational database systems RDBMS) - Oracle, Teradata. Implemented Ab-Initio Plans (Conduct>It), component folding and Micrographs.

Lead and architect the solution on migrating traditional ETL Ab-Initio project into open source system - Bigdata Hadoop and Lead the bigdata team in end to end hadoop implementation with the following Hadoop ecosystem - Map Reduce, Pig, Hive, Spark, Kafka, Oozie, Ambari, NiFi

Strong knowledge on Data Warehousing concepts and experience in Data Modeling, Design and Relational & Dimensional Modeling using Erwin & Visio with Oracle, DB2 and MS SQL Server.

Over 6+ years of Experiences in leading projects /teams of around 20 resources (Major Development projects and 24x7 support project model)

Experience in AGILE development methodologies.

Provide an overall architect responsibility including roadmaps, leadership, planning, technical innovation, server upgrade, database upgrade, security. IT governance, etc.

Experience in BI tools like Power BI, SSAS (OLAP Cubes) and SSRS and experience in Microsoft Web and windows Applications using VB.NET, C#.NET and Java web projects using J2EE and JSP

EDUCATIONAL DETAILS

B Tech., (Electronics Engineering) Madras Institute of Technology, 2003

M Tech., (Data Science) Indian Institute of Technology Hyderabad, 2019

CERTIFICATION AND TRAINING

Cloudera Certified Data Analyst in Bigdata HDFS

Certified Scrum Master from Scrum alliance

Microsoft Certified Professional in Web Applications Using VB.NET

TECHNICAL SKILLS

Technical Description

Technical details

Data Warehousing Tools Skills

Ab-initio, Talend, PL/SQL, Erwin, SSIS, Bigdata Hadoop - Mapreduce, Hive, SPARK, Kafka

Reporting Tools

Splunk, SSRS, SSAS, Crystal Reports, Power BI

Database Skills

Oracle, DB2, MS SQL Server 2012

Microsoft Skills

ASP, ASP.NET, VB Script, Power BI, MS SQL Server 2012, SSIS, SSAS, SSRS, ProClarity

Data Science and Machine Learning tools

Spark MLlib, NLTK, Scikit-learn, PyTorch, TensorFlow

Scheduler Tools

Control M, Cybermation

Programming Languages

Shell Scripting, Java, Python, VB.NET, C#.NET, VB Script, JavaScript

Operating System

Unix, Linux, Windows

WORK EXPERIENCE SUMMARY

Organization

Business Application

From Date

To Date

Technology

Role

Atos Syntel

BNFS

03-Jun-

2015

Till Date

Linux, Ab-Initio, DB2, Talend, Java BigData

Hadoop Mapreduce, Hive, Pig,

Cornerstone

Architect & Manager

Wipro

BNFS

25-Oct-

2013

08-May-

2015

Linux, Ab-Initio, DB2, Talend, BigData Hadoop Mapreduce, Hive, Pig

Architect & Manager

Cognizant

BFS,Life Sciences

25-Aug-

2003

11-Oct-

2013

Unix, Ab-Initio, Oracle, Teradata, BigData, HIVE, PIG, SQL Server

2005, Oracle 10g, Cognos BI &

Informatica ETL

Tech Lead & Manager

PROJECTS in Atos Syntel

1. PROJECT TITLE

RETL - MSSU

INDUSTRY

Banking

CLIENT

AMEX

DURATION

February ‘2018 – Till Now

ROLE

Project Manager & Architect

PROJECT

SPECIFIC SKILLS

Python, Ab-Initio both Batch and Continuous flow, BigData Hadoop Mapreduce, Cornerstone, Hive and Spark, Oracle

PROJECT OBJECTIVE:

MSSU (Management Information System ( MIS ) Setup & Servicing Utility) works as a bridge between report setups done by Apps (GIDM,GMIP etc) & Actual Reporting Engines (IBI, GDFS). It extracts, identifies and processes report setups requested through GIDM, GMIP (Global Management Information Processing (GMIP)) applications by users. It processes the data and creates report requests as triggers to the reporting engines (IBI for reports and statements, GDFS for data files) which enables them to generate actual reports. Actual reports are posted into the target system like CIW and GIDM where users can download their reports requested.

PROJECT DESCRIPTION:

MSSU was initially developed in Ab-Initio ETL tool with very limited functionalities and data loaded into DB2. This project is migrating from Ab-Initio ETL process into BigData Hadoop system using HIVE and SPARK SQL with cornerstone as data repository Reports are getting generated to the business team.

RESPONSIBILITIES:

1.Involved in Requirement gathering and source system study

2.Architecting the solution for the requirement and design the application or module

3.Technical Design Documentation, Low level and high level design document creation

4.Involved in coordination with stake-holders to sync up on the project deliverable timelines

5.Construction and implementation of the application and involved in helping the team to resolve the complex issues faced by the team

6.Involved in SIT testing and UAT testing.

7.Handled in Change management activities in ServiceNow and code walkthrough for Production support team.

8.Involved in handling ad-hoc work requests.

9.Involved in production support of Job Monitoring activities.

2. PROJECT TITLE

AMEX – GCC-SAGE-LCV (Loss Charge Volume)

INDUSTRY

Banking

CLIENT

AMEX

DURATION

June '2016 – February ‘2018

ROLE

Project Manager & Architect

PROJECT

SPECIFIC SKILLS

Python, Ab-Initio both Batch and Continuous flow, BigData Hadoop Mapreduce, Cornerstone, Hive and Spark, Kafka, Oracle

PROJECT OBJECTIVE:

SAGE-LCV is a system that recognizes their client companies using AMEX corporate cards which are going into loss and also de-recognizes the companies which are coming back from loss using multiple business rules defined in Amex Business team. It also calculates the Loss charge volume of the companies going in loss using Business rules. Based on these calculations, the incentive to Sales persons has been derived. Loss Charge Volume reports are generated in AMEX Corporate card department to be used for the higher management team and for incentive calculations to sales persons.

PROJECT DESCRIPTION:

SAGE-LCV was initially developed in Abinitio ETL tool with very limited functionalities and data loaded into DB2. This project is migrating from Abinitio Batch and continuous flows into BigData Hadoop system using HIVE, Kafka and SPARK SQL with cornerstone as data repository Reports are getting generated to the business team.

RESPONSIBILITIES:

1.Requirement analysis

2.Designing, ETL process migration and QA reports.

3.Acted as Architect and Manager in this project from offshore.

4.Manages a team of 32 resources.

5.Involved in change management activities in implementing this project

6.Continuously managing the system with migrating this application into Bigdata by market by market.

3. PROJECT TITLE

AMEX – SAGE (Strategic Analytics Global Environment)

INDUSTRY

Banking

CLIENT

AMEX

DURATION

June '2015 – June ‘2016

ROLE

Project Manager & Architect

PROJECT

SPECIFIC SKILLS

Unix, Ab-Initio both Batch and Continuous flow, BigData Hadoop Mapreduce, Cornerstone, Spark, Hive and Spark, Kafka

Oracle

PROJECT OBJECTIVE:

SAGE (Strategic Analytics Global Environment) program is a data warehousing project which collects, stores and manages all the card transaction data for all the countries by applying required business rules on transactional banking daily data. This application captures all the transactions and calculates incentive for sales persons based on card transaction and signing effective date. This project reports to the business team to get the complete picture of the process on a daily basis.

PROJECT DESCRIPTION:

SAGE was initially developed in Abinitio ETL tool and data are loaded into DB2. This project is migrating from Abinitio ETL process into BigData Hadoop system using HIVE and Spark SQL, reports are getting generated to the business team.

RESPONSIBILITIES:

1.Requirement analysis

2.Designing, ETL process migration and QA reports.

3.Acted as Architect and Manager in this project from offshore.

4.Manages a team of 32 resources.

5.Involved in change management activities in implementing this project

6.Continuously managing the system with migrating this application into Bigdata by market by market.

PROJECTS in WIPRO

4. PROJECT TITLE

CITI – AML – E2E Reconilation migration

INDUSTRY

Banking

CLIENT

CITI

DURATION

Oct '2013 – May ‘2015

ROLE

Offshore – Architect

PROJECT

SPECIFIC SKILLS

Unix, Ab-Initio, Talend, DB2

PROJECT OBJECTIVE:

AML E2E (End to End Program) project is a reconciliation program which verifies the Mantas output by applying required business rules on transactional banking daily data and cross checking its output with the Mantas team output (external). This project generates QA reports to the business team to get the complete picture of the process on a daily basis. This project reports the comparison report of comparing the E2E output with Mantas output and the QA report to the business team to get the complete picture of the process.

PROJECT DESCRIPTION:

E2E process was using Abinitio ETL tool. This project migrates the E2E process from Abinitio tool into Talend Open Source tool. A framework has been developed in Talend with a generic process, so that all the graphs for various countries can be changed into Talend in a single framework. This framework consists of 4 main workflows. First step would be Normalization which creates Normalized files to its child workflows. In this step various format of source files from various countries are converted into consistent format files. Second step would be augmentation workflow which applies all the business rules. Third step is the Comparison workflow. Here the augmented source file is getting compared to its corresponding Mantas file. Final / Fourth step is Reporting workflow which provides all the comparison results to the Business team and the detailed output are uploaded into the database. This system covers 238 countries of banking transactional data to be reconciled.

RESPONSIBILITIES:

1.Requirement analysis

2.Designing, ETL process and QA reports.

3.Acted as Architect in this project from offshore.

4.Manages a team of 10 in the Construction & unit testing phase.

5.Involved in change management activities in implementing this project

5. PROJECT TITLE

Capital One – CRG – CEMP Development Team & L3 support teams

INDUSTRY

Banking

CLIENT

CapitalOne

DURATION

Feb '2011 – October ’2013

ROLE

Offshore – Tech Lead

PROJECT

SPECIFIC SKILLS

Ab-Initio ETL graphs; Data warehousing project with Teradata database; BigData

PROJECT OBJECTIVE:

CEMP (Customer Effective Management Program) project is a data warehouse project which stores, analyzes and manages the customer survey information, Maritz is a third party vendor which helps to validate the customer survey for Capitalone. Maritz validates the survey information sent back to CapOne. CapOne studies the customer survey information and performs based on their customer needs.

PROJECT DESCRIPTION:

CEMP has two main ETL processes OUT-BOUND and IN-BOUND process by Abinitio ETL tool. In the OUT-BOUND ETL process, Capitalone sends the customer and their survey information to Maritz. Maritz validates the survey information and sends it back to Capitalone in the IN-BOUND process. All the Ab-Initio graphs are scheduled in the Control M tool.

In addition to the CEMP Development team, I managed two L3 support teams of 14 in size working in 24x7 model which covers all the corporate and Cards related applications in CapitalOne.

RESPONSIBILITIES:

1.Requirement analysis

2.Designing data model, ETL process and BI reports.

3.Acted as tech-lead in this project from offshore.

4.Manages a team of 16 in the Construction & unit testing phase.

5.Involved in change management activities in implementing this project

6.Manages two L3 support teams of 14 in size working in 24x7 model

PROJECTS in Cognizant

6. PROJECT TITLE

EDC Dashboard

INDUSTRY

Life Sciences

CLIENT

Abbott Vascular

DURATION

Jun '2010 – January '2011

ROLE

Onsite – Lead

PROJECT

SPECIFIC SKILLS

Ab-Initio ETL tool; Data warehousing with Oracle database; and BI reports using Cognos 10

PROJECT OBJECTIVE:

This project provides a Dashboard using Merck Clinical Research data.

PROJECT DESCRIPTION:

EDC Dashboard is the data warehousing project which extracts, transforms and loads all the details of the Merck Clinical Research Study data from Medidata Rave database (SQL Server) into CDRT (Clinical Data Reporting Tool) reporting database (Oracle 10g) using Ab-Initio ETL tool. This data includes all the clinical research data in Abbott Vascular on each research project (Study). Using Cognos Reporting tool, Study Metrics Dashboard is generated and displays the clinical study metrics information to the end users. The source database Rave Medidata server is a SQL Server 2005 database and the target database CDRT is an Oracle 10g database hosted in Merck. We designed and developed a data model for Oracle 10g database using Erwin data model tool. Data from the Medidata Rave database is transferred into CDRT database using Ab-Initio ETL tool. In ETL process, source data is loaded into staging tables (STG schema), the data is flattened with referential integrity into integrated tables (IDS schema) and the final data with pre-calculated metrics data and historical functionalities are loaded into DW tables (DW schema). A FM model in Cognos 8.3 is designed using the data source CDRT database and provides ad-hoc reporting to users. Through FM model, Dashboard metrics reports in Cognos 8.3 are designed and implemented which displays the EDC Rave Study Metrics information to the end users.

RESPONSIBILITIES:

1.Requirements analysis and coordinated with the customers to clarify the business requirements.

2.Involved in the design phase of the project & data model design of the data warehouse.

3.Coordinated with offshore to implement the project and acted as the onsite-lead.

4.Coordinated with offshore to implement the SIT testing of this project.

5.Involved in change management activities in implementing this project.

7. PROJECT TITLE

RoADS

INDUSTRY

Life Sciences

CLIENT

Merck – MRL

DURATION

May '2009 - Apr '2010

ROLE

Onsite – Lead

PROJECT

SPECIFIC SKILLS

Abintio ETL Tool; SQL Server 2005 database and OLAP cubes using SSAS; BI Reports using SSRS and Cognos 8.3.

PROJECT OBJECTIVE:

RoADS (Repository of Authoritative Data Sources) is the data warehousing project which collects, stores and manages all the details of the Merck Research pharmaceutical projects from various systems OPX2 (PreClinical and Clinical data), BDRC (Basic Research data) and EPDR (Electronic Portable Data Repository - Clinical data) into a central repository called (RoADS). These details include drug details of the pharmaceutical projects, milestone details of various phases of projects, Risks involved in each phase and cost and effort involved in each phase of the Merck research projects.

PROJECT DESCRIPTION:

This project provides the authoritative data to the Merck Research users through Reports at any time. All the source data are loaded into RoADS data warehouse through ETL process using Ab-Initio on a daily schedule. All the Business rules are applied in the ETL process as per the requirements. After processing the data, the system reports the data to the Users through SQL Server Reports, SpotFire Reports & Cognos Reports. The required data is loaded into MS OLAP Cubes. User analyzes the OLAP cubes data through ProClarity Analytical tool by creating ProClarity reports/views.

RESPONSIBILITIES:

1.Requirements analysis

2.Designing data model, ETL process and BI reports.

3.Acted as the architect in this project.

4.Coordinated with offshore to implement the project and acted as onsite-lead.

5.Involved in change management activities by implementing this project.

8. PROJECT TITLE

OMJ Reporting Services

INDUSTRY

Life Sciences

CLIENT

IMS America

DURATION

Aug '2007 - May '2009

ROLE

Lead Programmer and Onsite – Coordinator

PROJECT

SPECIFIC SKILLS

Ab-Initio ETL Tool; SQL 2005 database and BI reports using SSRS

PROJECT OBJECTIVE:

This project generates the Weekly and Monthly Rx information reports for the OMJ markets (part of JnJ) and sends the reports to OMJ.

PROJECT DESCRIPTION:

This project uses SQL 2005 as database and retrieves the reports using SQL Server Reporting services. In this project, a generic ETL tool and generic reporting tool have been developed. Using these tools, we could implement any kind of ETL process and any kind of report generation process in SQL 2005. Any BI projects in SQL 2005 can reuse these tools.

A generic ETL tool has been implemented to load any kind of source data like Flat files etc., into any database server like SQL 2005, SQL 2000, Oracle, etc.,. This tool has been built using the new features in SQL 2005, SSIS and SSAS.

A generic Reporting tool for generating any kind of reports from any source data has been built with new features in SSRS. The basic criteria to utilize this tool are

1.Data Source (A data source has to be provided. It could be any RDBMS database).

2.Report Template (This gives how the report looks like or the structure of the report)

3.SSRS RDL Template (This gives what data is retrieved to the report and how the data is arranged in report).

4.Type of report (Excel, CSV, PDF or HTML)

Using the above two tools, we could simply implement any kind of projects in Business Intelligence required from our client side. This will reduce our time drastically.

RESPONSIBILITIES:

1.Requirements analysis.

2.Involved in the data model design using Erwin tool.

3.Coordinated with offshore team for Unit testing & system testing.

4.Implement the project into production through Change management process.

5.Automated the process using the scheduler tool cybermation.

9. PROJECT TITLE

Wyeth Oncology

INDUSTRY

Life Sciences

CLIENT

IMS America

DURATION

Jun '2006 - Sep '2006

ROLE

Lead Programmer and Onsite – Coordinator

TEAM SIZE

7

PROJECT

SPECIFIC SKILLS

Ab-Initio ETL graphs in Unix platform, SQL 2005 database and generate Access database.

PROJECT OBJECTIVE:

This project is a data warehouse project which stores and reports the HCO information and market

strategy for the oncology market related to Wyeth products.

PROJECT DESCRIPTION:

A data warehouse is built in SQL Server 2005 database which gets data from different kinds of data HCO information, Outlet information and prescriber information using Ab-Initio ETL graphs. All the complex business rules like Decile Process, Alignment process, Apportion process etc are applied into the ETL process. A data mart in access database is created on monthly basis from the data warehouse and delivered to the client Wyeth.

RESPONSIBILITIES:

1.Requirement analysis.

2.Involved in the data model design using Erwin tool.

3.Designing ETL process.

4.Involved in offshore liaison for this project implementation.

5.Implemented this project into production through change management process.

10.PROJECT TITLE

Custom Delivery Online

INDUSTRY

Life Sciences

CLIENT

IMS America

DURATION

Jun '2004 - Dec '2004

ROLE

Lead Programmer

TEAM SIZE

4

PROJECT

SPECIFIC SKILLS

Web applications and web services using ASP.net, ASP and VB and SQL Server

PROJECT OBJECTIVE:

A new subsystem CD4 online is developed as part of Custom Delivery online. This application is

used as billing system for the requests raised by IMS clients.

PROJECT DESCRIPTION:

IMS clients send their requests to IMS through this online. IMS Business Analysts evaluate the requests; create the bills and charge their clients through the new sub system CD4 online. The evaluated requests send to the IMS development team. The IMS development team completes the requests and sends the results to the clients through this online system.

RESPONSIBILITIES:

1.Requirement Analysis.

2.Constructing the changes in the web project and developed a new subsystem CD4 online.

3.Implement the project into production through change management process.



Contact this candidate