Post Job Free

Resume

Sign in

Big Data Analyst

Location:
Bengaluru, KA, India
Posted:
July 17, 2017

Contact this candidate

Resume:

PRIYANK KUMAR CHAUHAN

Contact: +91-914**-*****

Email: ac06u6@r.postjobfree.com

SUMMARY OF SKILLS

Around 5 year experience in Business Information Management and Expertise in working on Big Data, Business Intelligence and Data Warehousing Technologies

Expertise on Informatica MDM, Informatica Data Quality, Informatica Powercenter, Oracle EDQ, Oracle Data Integrator, Talend, Cognos and SSRS.

Strong skills in Big Data Technologies like Hadoop, HDFS, PIG, HIVE, Flume, Sqoop, Drill, HBase, MongoDB, Zookeeper and Oozie.

Worked Onshore and Offshore with various clients like Santander Bank, Pubmed, Avanss, Dow Jones, MetricStream, Emaar Group, NSC, Edelweiss Commodities and Abila in various Domains like Banking, GRC, IoT, Mortgage, Healthcare, Retail and Hospitality.

Installation & Configuration on Linux/Windows on all environments including Cloud for Big Data, IBM, Oracle, Microsoft and Informatica Tools.

Good understanding of Java especially on OOPs and Client Server technology.

Ease and willingness to learn new technologies, Interest in professional development Extensive experience in the field of User requirement, Requirement analysis, Leadership and decision making.

Self-Starter, Quick Learner and Strong Analytical, Problem solving & Team Building skills. Skills

o Business Intelligence o Big Data

o Data Warehousing o Master Data Management

o Data Governance o Data Quality

o Reporting & Analytics o ETL Solution

EDUCATIONAL CREDENTIALS

Collibra Certified Data Steward

IBM Certified for Business Analytics

Bachelor of Technology (Information Technology), UPTU Technical Skills

Pig, Hive, Flume, Sqoop, Drill, ZooKeeper Hadoop, HDFS, Hortonworks, Cloudera

Informatica MDM HBase, MongoDB

Informatica Data Quality NLP Tools like Semantify, Data RPM

Informatica Power Center Unstructured, Semi-Structured, Structured Data

Oracle Enterprise Data Quality Talend Open Studio

Oracle Data Integrator SQL, PL/SQL, NoSQL

Collibra Data Governance R, Python

IBM Cognos, SSRS Cloud: Azure, AWS

Oracle, SQL Server, DB2, Informix Shell Script, Java, Xml, Web Services WORK EXPERIENCE

Team Lead – BIDW Mphasis Limited, Bangalore Nov 2016 – Till Now

Sr. Business Analyst Sementify Pvt Ltd, Hyderabad Dec 2015 – Nov 2016

Sr. Business Analyst Mettl Induslynk Pvt Ltd, Gurgaon Aug 2015 – Dec 2015

Sr. Software Engineer IMC Global Service Pvt Ltd, Pune Oct 2013 – Jul 2015

Software Engineer TBS Pvt Ltd, Noida Jul 2012 – Sep 2013

Internship I-World Solution, Faridabad Dec 2011 – Jun 2012 PRIYANK KUMAR CHAUHAN

Contact: +91-914**-*****

Email: ac06u6@r.postjobfree.com

PROFESSIONAL PROJECT WORK

Project Name/Client: IoT Transportation Analytics Solution for Avanss, US Technology: NLP, Hive, Sqoop, Drill, ZooKeeper, OOZIE, Reporting and Analytics, HDFS, HDP 2.4 Avanss provides an IoT based complete Transport Management System of Bikes, Cars and Buses. It includes the routes, schedules, boarding, alighting, running status, fuel status, traffic information and availability of Bikes/Cars on station. For buses they provide the running status, schedule, delay and route information at any point of time. An Advance Sensors System had been established to keep track of all the information related to vehicles.

Lead the requirement gathering.

Analyzed raw data in Structured Data Sources.

Implemented ETL to ingest/bulk load data from Database tables to HDFS.

Write the extraction and transformation job workflows & scheduled using Oozie.

Used Apache Drill for quick data validation and Report testing.

Defined Data Model and Business Model.

Used NLP search engine for Adhoc Report generation.

Our task is to provide the statistics and real time information of vehicles in report as well as analytical reports. Project Name/Client: CDO Surge Team for Santander Bank, US Technology: Big Data, Data Governance, Collibra DGC, IDQ, SQL Server Santander Consumer USA Holdings Inc. is a full-service, technology-driven consumer finance company focused on vehicle finance and unsecured consumer lending products. The company has a serviced portfolio of more than $50 billion has more than two million customers across all credit grades. Data Governance is one of the critical groups in the organization that manages the quality of data that is maintained and furnished to the various holding companies as well as the reporting bureau’s in addition to the CCAR filing that they are responsible for. Comprehensive Capital Analysis and Review (CCAR) is a US regulatory framework introduced by the Federal Reserve in order to assess, regulate, and supervise large banks and financial institutions - collectively referred to in the framework as Bank Holding Companies (BHCs). The assessment is conducted annually and consists of two related programs: CCAR & DFAST stress testing. The capital structure is stable given various stress-test scenarios and planned capital distributions, such as dividends and share repurchases, are viable and acceptable in relation to regulatory minimum capital requirements.

Involved in the Business Data Strategy and Solutions.

Implementing Data quality, and master data management to enriching data lineage with more business value.

As part of the DG group here, my responsibilities are defining Data Quality rules for various items.

Maintain quality controls across the enterprise as well as the data submitted to the Feds.

Key areas of implementation would include the CCAR - Fed Edit Checks and KDE’s, Risk & Finance Data Management (RFDM) KPI’s and All Loans & Lease Losses (ALLL) KDE’s evaluations. Project Name/Client: Healthcare Analytics Solution for Pubmed, US Technology: NLP, PIG, Hive, HBase, OOZIE, ZooKeeper, Reporting and Analytics, HDFS, CDH 5.4 PubMed is a service of the US National Library of Medicine. PubMed compromise over 26 million citations for biomedical literature from Mediline and life science journals. PubMed citations and abstracts include the fields of biomedicine and health, covering portions of the life sciences, behavioral sciences, bioengineering, chemical sciences, and medical, nursing, dental, veterinary, healthcare and preclinical sciences journal articles. PubMed also provides access to additional relevant web sites and links to the other NCBI molecular biology resources.

Analyzing raw data in Unstructured Data Sources.

Aggregation and Data Transformation using User Defined functions.

Used PIG to pre-process, validate and transform data.

Used OOZIE to automate data loading from HDFS to HBASE.

Understand Business terms and Defined Business Model.

Used Semantify NLP for semantic and faster data extraction and Report generation.

Linked the related document in Report, so User can also view the referential document from report. PRIYANK KUMAR CHAUHAN

Contact: +91-914**-*****

Email: ac06u6@r.postjobfree.com

Project Name/Client: GRC BI Reporting Solution for Metric Stream and Dow Jones, US Technology: NLP, HBase, Hive, OOZIE, ZooKeeper, Reporting and Analytics, HDFS, HDP 2.4 Governance, Risk and Compliance BI Reporting Solution provides unique data for monitoring a range of risks associated with third-parties. They help financial institutions and businesses meet regulatory requirements on Anti- Money Laundering, Anti-Bribery and Corruption, Economic Sanctions, Third Party Due Diligence and Commercial Risk operations.

Requirement gathering /solution architecture.

Data Analysis for structured data and semi-structured source systems.

Created HBase tables to load large sets of structured and semi-structured data.

Written Hive queries for data analysis using Canned Report to meet the business requirements.

Enable users to generate Adhoc Reports and faster extraction of data using Semantify NLP. Project Name/Client: BI Solution for Mettl, India

Technology: Talend, SSRS, SQL Server, PostgreSQL, MongoDB Mettl, India's leading online skill assessment and online certification platform, has introduced an innovative new tool to its successful lineup. The company's new case study simulator aims to bring case study methodology to online assessments.

Understanding Business Problem.

Data Validation, Data Integration and Data Quality using Talend Open Studio.

SQL and PL/SQL code optimization and fine tuning.

Created Reporting data mart by integrating data from MongoDB, Postgres and SQL Server.

Setup BI Reporting solution for Traditional Reports.

Adhoc Report generation on Client requirement.

Generated Assessment Reports.

Project Name/Client: MDM Solution for Emaar, Dubai Technology: Informatica MDM, IDD, IDQ, Powercenter, Oracle Emaar IT serves the multitude of business by Emaar group in Dubai, such as Real Estate, Hospitality, Malls and Recreation. This Project is to provide a single 360 degree view of unique customers spanning across various source systems using Informatica MDM, IDD, IDQ and Power Center.

Successfully installed and configured MDM 9.X on Linux.

Thoroughly involved in understanding of MDM architecture and business processes.

Lead the requirement gathering /solution architecture for MDM.

Developed reusable, efficient and optimized solution with Incremental Update Strategy for 50+ different domain servers with various types of databases.

Performed data cleansing and Data Standardization using Data Quality & Reference table.

Well versed with data quality features like Analyst & transformation like Key Generator, Standardizer, Parser, Labeler, Case Converter, Match, Consolidation etc.

Prepared logic for data rejection management along with the reason for rejection.

Designed /configured and implemented Informatica MDM engagements.

Applied Address Doctor transformation for Address Validation and Standardization.

Effectively conducted Informatica MDM Hub configurations Data Mappings (Landing, Staging and Base Objects), Data validation, Match and Merge rules.

Used Arabic and Demo population for fuzzy match rules.

Instrumental for data integration, Data Quality, data architecture and Master Data Management, project life cycle Phases, best practices and processes.

Integrated external business application with MDM hub using Informatica Developer.

Created Entity Objects, Relationships and Hierarchies for 360 degree view of data in IDD.

Created IDD Application, Defined Physical, Logical Schema and deployed IDD application to enable user for search and view consolidated records.

PRIYANK KUMAR CHAUHAN

Contact: +91-914**-*****

Email: ac06u6@r.postjobfree.com

Project Name/Client: Data Warehousing & Reporting Solution for NetFORUM, US Technology: Informatica Power Center, Informatica Data Quality, SSRS, SQL Server Abila provides netFORUM Enterprise which is membership management, enterprise system that allows you to coordinate association management, e-Marketing, e-Commerce, content management, and business operations in a Web-based environment.

Lead the requirement gathering and Data Analysis using Data Profiling.

Data Integration from legacy data to NetFORUM master database.

Implemented SCD Type 2 strategy to handle delta detection.

Used Address Doctor to validate Person/Organization Address.

Used Reference tables for Data Cleansing for various Abila clients.

Generated financial Reports using SSRS.

Project Name/Client: Data Warehousing Solution for NSC, US Technology: Oracle Enterprise Data Quality, Oracle Data Integrator, Oracle The National Student Clearinghouse is the nation's trusted source for education verification and student educational outcomes research. More than 3,500 colleges and universities, enrolling over 98% of all students in public and private U.S. institutions, participate in the Clearinghouse. Participating institutions provide access to actual enrolment and degree information on each of their students to us. Clearinghouse services are designed to facilitate an institution's compliance with the Family Educational Rights and Privacy Act, The Higher Education Act and Other laws.

Requirement Analysis and Mapping Document

Analyzing the raw data using Data Profiler.

Performed Data Validations using Audit feature based on Business Rules.

Data Integration and Data Quality for distributed Source Systems to Master Database.

Created Match and Merge rules for Data Consolidation and Golden Records. Project Name/Client: Reporting Solution for S2B, Canada Technology: IBM Cognos 10.X, SQL Server

S2B was an auction platform where user can post their requirement/project and other interested user can bid on that. Here we had three different types of actors, Supplier, Buyer and Management. Buyer can create a bid for projects; Supplier can bid on buyer’s Project, Management will take care of transactions. I have generated Reports on transaction of Buyer and Supplier for Analysis. Management can do analysis on all the transaction & actions by Any Buyer or Supplier on monthly as well as yearly basis by reports.

Thoroughly involved in understanding Business Model and User Requirement.

Actively participated in defining Data Model

Developed Stored Procedures and Triggers to fulfill user requirements.

Created Framework in Cognos.

Worked on Metric Studio for generating Score Card.

Generated Reports using Report Studio, Query Studio. Declaration

I hereby declare that the information furnished above is true to the best of my Knowledge and belief. I have the relevant records to prove the same.

Date:

Place: Bangalore Priyank Kumar Chauhan



Contact this candidate