Mahesh Gope ** Hartley Crt
Email : *********@*****.*** Concord – L4K 4W4
Mobile : 416-***-**** ON, Canada
Professional Summary
More than 12 years of IT experience in the implementation of Data warehousing projects using Talend, Informatica Power Center and Power Exchange, Informatica BDE, Data Stage, Alteryx, Syncsort DMExpress, Microsoft SQL Server (Azure), Cognos and Oracle PL/SQL.
Experience in the data migration and data integration.
Experience working on Kubernetes platform using docker, Jenkins and Rancher.
Experience with the spark platform using python and Scala.
Used power exchange for CDC (for capturing the before and after image of the records from the oracle tables)
Experience in working with complex COBOL copybooks (data maps through power exchange).
Experience working as a Data Modeler, both logical and physical and using Erwin data modeler tool.
Experience with Informatica BDE and Hadoop to access data in HDFS and Hive.
Extensive experience as a Team/Tech Lead and leading a group of resources.
Experience in using GitHub, Docker image, bitbucket, Jira board.
Excellent experience in using the Autosys, Tidal scheduling tool.
Technical Competencies
ETL Tools
Talend Studio, Informatica 7.X, 8.X, 9.X, 10.X, Informatica Power Exchange 9.X, Informatica BDE (Big Data Edition), Data Stage, Syncsort DMExpress, SSIS, Alteryx
Reporting Tools
Cognos 8, SSRS
Databases
Oracle 8i/9i/10g/11g, Teradata, Sybase, SQL Server 2008, DB2
Scheduling Tools
Tidal, Autosys, Control M, TWS, Airflow
Functional areas
Banking and Finance, Telecommunications, HealthCare
Programming/Scripting Languages
Scala, Python, Java, SQL and PLSQL, Unix Shell and windows DOS batch scripting
Tools
Toad for Oracle 9.0.1/10.6.1.3, SQL Developer, IntelliJ
WORK EXPERIENCE
Data Enablement & Architecture/ Automated Data Quality Suite Mar 2017 – Till Date
(Senior Data Engineer - Scotiabank)
ADQS: creating the unified view for the manual DQ issues that are getting generated from Collibra as well as for the automated issues that are being created by the business. Data is being ingested after doing various functional requirements implementations like applying technical/business rules to an attribute. This unified view is being used by the dashboard team for the reporting purposes.
-Talend Rest API is being used to generate the access token from Passport URL and this token is then being used for authentication to Collibra.
-Writing the java programs within the Talend Studio to address various business transformations.
-Used various data transformation scripts based on the workflow needs and executed these jobs from TAC, airflow, eclipse.
-Connecting to the collibra end points for various assets to pull the data quality issues (JSON payloads) attributes alongwith the relationships between multiple DQ issues.
-Creating DQ metrics rule for the Data Quality Technical Rule which represents an instance of a business rule applied to a column e.g. what is the % of the total passed records vs total available records, what is the completeness % of the end of day balance etc. Aggregated DQ metrics for all the technical rules is also being calculated for various assets.
-All the DQ metrics and DQ dimensions are being integrated from ADQS to the UC (universal catalog)
DEA : Creating common data lake by sourcing the data from multiple source systems (e.g. basic banking, mortgage, investment etc.). All departments data initiatives will leverage this product, which gives them easy access to trusted, well-understood data assets and services. Use CI/CD pipelines to build the package and deployment using docker image.
Environment/tools : Talend Studio, Spark, Scala, DB2, UNIX, Airflow, Java, Collibra, Python, Azure, Presto
Real Time Customer Insights – Consultant
(Senior Data Engineer - Scotiabank)
Targetting customers for the real time insights campaigns and various nudges for its existing and new customers. Each individual campaigns has been categorized into various LABS. Enterprise Data Lake’s data are being used as a source for all these campaigns.
Designed the solutions for the data stage migration project
Worked on multiple campaigns (e.g. Paperless, Pay-no-Pay, Fraud, CRM) in analyzing and preparing the data for the ESM and DSM team who are responsible for launching various campaigns
In addition, worked on various campaigns activities like Unica and SAS leads campaigns and nudges that involves interacting with AS400 system
Responsible for reviewing the codes like Scala, Python, Data Stage, Informatica, Hive and Unix scripts
Analyzed campaigns data using Pandas.
Environment/tools : Scala, Informatica, Data Stage, HIVE, DB2, Tivoli, Bitbucket, Talend, Python, Azure, Sqoop, UNIX, Alteryx, Pega, Autosys
CIBC – Lead Data Engineer Dec 2013 – 28th Feb 2017
Project # 1 CCT - EDH
Interacting with business to gather requirements and to propose the effective and efficient design and leading the technical team for coding and designing the tasks as per the CIBC standards.
Creating the Enterprise data hub (EDH) to integrate all the source system data in the hub using Informatica BDE and ingesting these data in the HIVE tables.
Used Informatica Power Exchange and syncsort DMX to process complex COBOL Mainframe files based on the copybooks.
Responsible for all kinds of designs, codes (e.g. ETL mappings, DB queries, autosys job designs).
Environment : Informatica BDE, HIVE, Oracle, Power Exchange, Autosys, UNIX, Syncsort DMExpress
Project # 2 BASEL III (Treasury LOB)
The global financial crisis has forced Governments and central banks across the world to overhaul the existing Basel guidelines for banks leading to the drafting of the U.S. Basel III Capital Rules. Consequently, areas like capital adequacy, liquidity risk, stress testing, trading, consumer protection and recovery and resolution planning will come under the purview of strict regulations.
Environment: Informatica Power center 9.0.1, Oracle 11g, Unix, Autosys Scheduling Tool, SQL*Loader
Project # 3 WAL
Worked under the Finance Technology department (EDSF) for replacing the WIL legacy system with the new WAL (wholesale Automotive Lending) system for improving the existing process with regards to the easy access by the end users and other customers like dealers and suppliers
Responsibilities involved in creating the application design document containing the requirements in detail.
Environment: Informatica Power center 9.0.1, Oracle 11g, Unix, Autosys Scheduling Tool, Informatica Power Exchange 9.0.1.
CGI Group Inc. – Tech Lead ( Consultant ) Jan 2012 – Dec 2013
DRG Conversion / DRG TAHP Extract
Tufts is one of the larget healthcare provider in Massachussets, USA .DRG (Diagnostic Related Group) was one of the project related with the different diagnostic related groups.
Cigna Provider Extract
This project was to re-write the existing application written in Oracle using Informatica in order to improve the performance and maintenance. Used power exchange for CDC for capturing the before and after image of the records from the oracle tables.
IHM(Integrated Health Management)
Primary goal of this project was data integration from different departments related with Tuffts. Requirement also involved generating various bills related to Medical, Pharmacy and Enrollment of the customers. The files for each categories were generated on weekly as well as on monthly basis based on the claims made by the customers. Used Java for creating the connections between oracle and the UI.
Environment: Informatica Power center 9.1, Informatica Power Exchange, Oracle 10g, UNIX, Java, Tidal
IBM India Pvt. Ltd. – Application Consultant Aug 2010 – Jan 2012
Bill ODS Consolidation/Bill Presentment
Bill ODS Consolidation basically deals with the merging of wireless and wireline bill i.e. instead of generating separate bills each for wireline and wireless, now a single bill will be generated.
Environment: Informatica Power center 8.6.1/9.1.0, Oracle 10g, Contol-M, Microsoft Visio, UNIX.
RBS (Royal Bank of Scotland) - Software Associate Dec 2008 – Jul 2010
ETL Re-Architecture
ETL Re-Architecture was the extension of the country risk. The objective was to tune the existing code by creating different streamlines.
Country Risk
After RBS acquired ABM amro bank, this project was started for migrating and integrating the data of ABN with the RBS. After the merger of ABN amro bank with RBS, the rules that were followed by the ABN was now going to apply all the rules defined by RBS.
Roles and responsibilities involved creating physical data models by creating various queries out of the logical data models to test various scenarios including the testing of data models and database.
Environment: Informatica Power Centre 8.6.0, UNIX, Oracle 10g, Microsoft Visio, Erwin
Capgemini India Pvt. Ltd. – Associate Consultant Feb 2007 – Dec 2008
Liquidnet
This project was related to the trading firm located in USA and the client was Liquidnet. The main objective of this project was data integration of the firm that Liquidnet has bought and were being merged into its business. PL/SQL used for ETL implementation and Cognos for the reporting tool.
Equities Analysis Datawarehouse
The objective of this Project was to define various business parameters along which a tool can be developed to assist stock investors / traders to evaluate a particular company to determine its investment attractiveness.
VCAS Data warehouse maintenance
Creating mapping for different source systems related to Lehman Brothers business in the market. Informatica used as an ETL tool and cognos 8 for the reporting.
Environment: Informatica 7.1.4, Data Stage, Cognos,Sybase, Perl, UNIX and AUTOSYS, SSIS, SSRS, SSAS, SQL Server, Oracle 9i, PL/SQL