SHAKTI PRASAD MOHAPATRA
(Cloudera Certified Administrator for Apache Hadoop (CCAH))
Address: **** ****** ******, *** ****, Glen Allen VA 23060
Contact: 1-804-***-**** Email:************.*********@*****.***
EXPERIENCE SUMMARY
9 years of experience in Information Technology industry with 4+ years in client facing role focused on Banking
and Financial Service domain as well as Insurance industry having worked for Fortune 500 Clients such as
Capital one, Travelers with multiple skillsets such as Mainframe, Hadoop, ETL, Java, Data Virtualization.
SPECIAL SKILLS
Cloudera Certified Administrator for Apache Hadoop (CCAH))
M102 Certified MongoDB for DBAs
Experienced on working with Big Data and Hadoop File System (HDFS).
Hands on experience in working with Ecosystems like Hive, Pig, Sqoop, Map Reduce, Flume, Oozie,
HBASE, Phoenix, Impala
Strong knowledge of Hadoop, HDFS, Hive and PIG.
Efficient in building hive, pig and map reduce scripts.
Experienced in Data Virtualization tools such as Jboss, Teiid.
Being worked in multiple roles such as designer/developer, Tech lead, BSA, Technology evaluator has
given me immense confidence in adapting new technologies and any challenging roles.
Strong exposure to IT consulting, software project management, team leadership, design,
development, implementation, maintenance/support and Integration of Enterprise Software
Applications, Technology Evaluations in Proof of Concepts.
Strong experience on projects involving mobile applications on Digital and Card Processing.
Extensive experience working on projects with Agile Methodology which includes Story writing, Sprint
Planning, providing updates thru Stand ups, participating in Sprint retro with feature team.
Experienced using VERSIONONE to manage agile projects.
Strong experience on both Development and Maintenance/Support projects.
Good team player with excellent communication skills.
TECHNICAL SKILLS
Technologies Big data/Hadoop, Mainframe Legacy Systems, AB INITIO, Data Virtualization
Tools / Languages: CDH 4.2, Hive, Pig, Sqoop, Flume, Impala, COBOL, JCL, Java Map Reduce,
Jboss DV, Teiid Designer.
Databases: Mongo, DB2 V10.0
Operating Systems: Linux, Mainframe ZOS
Database Tools: Phoenix,OMEGAMON,SPUFI,DB2ADM, QMF, File Manager
DB2,Datastudio,ERWIN, Jboss Studio
Scheduling Tools: Oozie,CA 7, Control M
File Management & File Manager, CAFM, Endevor, ChangeMan
Version Control Tools
PROJECT PROFILE:
Employer : Cognizant Technology Solutions Client: Capital One
Role : Senior Associate 5+yrs
Data Virtualization POC:
Objective of this POC is to leverage the advantages of Data Virtualization tools to create an
abstract layer on top of disparate data sources such as DB2, Teradata, Hadoop, Web Services
without actually physically transferring the data.
Roles and Responsibilities:
Technologies & Tools used: Jboss Studio 8.0.0, Jboss DV 6.1 Beta, Teiid Designer, Eclipse, DB2,
Teradata, Hadoop, Hive, HDFS.
Activities:
Evaluating Jboss Teiid,
Creating connection profiles for various data sources such as DB2, Teradata, Hive,
Flatfile through JDBC/ODBC connections.
Create, Deploy, Execute VDBs to test the abstract layer.
Evaluating Apache JENA, RDB2RDF mapping to be able to create a symbolic layer on
top of the abstract layer(Work in progress).
Customer Exposure View(CEV)
As a part of Customer Exposure View (CEV) project, Capital One is trying to create an application
which will provide 360 degree customer view to the Capital One customer support
representatives. This will help the customer support in real time decision making as they can
access the details around the customer like
What all products customer uses?
Customer’s payment activities for all the products
How the customer is performing?
Roles and Responsibilities:
Technologies used – Hive, Pig, Java MapReduce, Unix Shell Scripting, ControlM
Activities :
Filling the data lake with customer data from various sources
Define ETL process using Pig, Java MapReduce
Load the data into Hive
Enterprise Data Store Enhancements & Maintenances
Database Administration and maintenance for various applications using databases within EDS
(Enterprise Data Store) as well as ADS (Application Data Store) environments which is a near
real time Data Store for consumption of operational data for various real time services as well as
for client facing and front end Applications within Capital One.
Roles and Responsibilities:
Understand System requirement, High level Design, Data Model and transform into
Database design.
Coordinate with project team to ensure optimized Table Design approach to be followed.
Arrange High level Design Review, PDM Review with Platform team.
Design the DB2 Objects across Dev and QA regions.
Reviewing Queries in Batch or Online programs (Stored Procedures) to ensure optimal
performance.
Reviewing Load, Unload and other utility jobs.
Prepare and review Pre Prod Documents.
Preparing Production scripts and working with Production DBA to perform the changes in
production environment.
Working with Production DBA to schedule Imagecopy jobs, RUNSTATs and REORG jobs
as required.
Support QS testing
Support Performance testing and Production Implementations.
Monitoring the query performance using OMEGAMON.
Operational Data Store Service Definition Factory(ODS SDF)
This Project involves developing solutions for the Capital One Cards/ Bank/ Financial/Digital
Services line of business that involve ODS. Often these solutions are part of a enterprise wide IT
implementations that target improve Capital One’s business. Being a data store most of the
projects involving ODS, require ODS to store data that could be provided by any of the Capital
One internal systems or its partners and used by same through real time or batch processes.
Technically ETL processes and mainframe processes are developed to load and retrieve data.
Roles and Responsibilities:
Leading offshore team.
Preparing project statistics, estimation, Analysis,
Design, Coding, Review.
Unit testing for the Online and Batch applications.
System testing support, Performing Implementation tasks, Validations during project
deployment.
Reviewing complex business requirements and System requirements.
Analyzing the requirements and Coordinating with multiple stake holders to
design/propose an optimal solution using Mainframe/ETL technologies.
Reviewing Detail Design, Implementation plan with Clients and external Project team.
Reviewing system test cases, Supporting Integration testing, System testing,
Driving Project Implementation.
Employer : L&T Infotech Ltd. Client: TRAVELERS
Role : Developer/Support Analyst 3yrs 7 months
Personal Line Information Platform (PLIP)
PLIP is data warehouse containing data related to personal lines auto. It contains Current as well
as History data pertaining to the auto insurance.
Following are the major tasks performed on the system.
Collecting data from various upstream sources,
Formatting of the collected data.
Storing it in the database.
Extraction of data from database to various downstream sources on a monthly and
quarterly basis.
This Project also involves enhancements activities to the various processes that stores
data/policies pertaining to the auto insurance in the PLIP (Personal Lines Information Platform)
database. There are mainly 4 processes starting with selecting auto insurance data from the file
sent from the CARS DB, then formatting the data in the way PLIP can accept, then splitting the
same into different state wise and finally insert/update the database. The information stored in
PLIP are used by various downstream systems like FIN DB, PRICING, ISO Auto and Automart for
various financial calculations, Reporting and for Audit Purpose.
The important downstream system is the ISO AUTO. This system extracts the necessary
information from PLIP DB and reformats the extracted data in various steps/processes and
reports the valid data to the Insurance Service Office (ISO) which is a regulatory authority of the
United States. The data submitted to ISO is for compliance and any delay/non compliance would
lead to a heavy penalty to the customer.
The system involves various daily, weekly, monthly, quarterly and yearly jobs which are scheduled
in CA 7 to keep the system up and running.
Roles and Responsibilities:
Coding and Unit testing Scheduling and monitoring of various batch jobs.
Preparing Scheduling Calendar yearly basis. Monitoring of the scheduled and On
Demand jobs. Fixing any production abend, preparing the emergency jobs/procs and
raising request to restart the same.
Handling adhoc request from clients which involves some analysis of any existing
process, providing data requested by client or any other external institutions which
requests some particular data.
Handling enhancements of some existing processes. Enhancement involves gathering
requirements, preparing test cases and then coding, unit testing, supporting system
testing and production deployment.
CERTIFICATIONS
Certification Name Year
M102: Mongo DB for DBAs 2015
Cloudera Certified Administrator for Apache Hadoop(CCAH) 2014
IBM Certification for Database Administrator DB2 10 for z/OS(DB2 612) 2013
IBM DB2 UDB V8.1 Family Fundamentals (DB2 700)
2010
DB2(R) Universal Database V8.1(DB2 703) 2009
IBM Database Administrator DB2 UDB V8.1(DB2 702) 2008
Property and Liability Insurance Principles From AICPCU(INS 21) 2007
ACHIEVEMENTS
Award Name Organization Year
Above & Beyond Cognizant Technology Solutions 2011
Achiever of the Month L&T Infotech Ltd. 2009
Spot Excellence Award L&T Infotech Ltd. 2008
Spot Excellence Award L&T Infotech Ltd. 2007
EDUCATION
Bachelor in Engineering in Comp. Sc. & Engg., Orissa Engineering College, Bhubaneswar,
Odisha, India 2006.