Post Job Free
Sign in

Data Project

Location:
Sacramento, CA, 94203
Salary:
$ 85/hr
Posted:
October 25, 2017

Contact this candidate

Resume:

*****.********@*****.***

908-***-****

Summary:

Having 15+ years’ IT experience, performed multiple roles as Data Analytics architect, Application architect, Technical lead, Team lead, Data Analyst, Lead Enterprise Data Warehousing and Developer Legacy mainframes.

Have 4+ years’ experience in statistical modelling, data analytics and Machine Learning, build predictive models, mine the data for useful patterns using R-Programming/Python.

Having 7+ years’ experience in Data Modeling, Data Warehouse Design and Development on EDW (Enterprise data warehouse), DI (data integration) MDM (Master data management) and BI (Business Intelligence) applications using IBM Infosphere DataStage.

Having 4+ years’ experience in Developing, Enhancing, Maintaining, Testing, performance tuning legacy and web applications using DB2, Oracle, CICS, JCL.COBOL technologies.

Integrating data from various sources, explore data, Pre-Process data for Machine learning models, performed Feature engineering, Train the models, Test the models and Evaluate the models.

Worked with multiple clients GAP, XEROX, Citi and DHCS worked with various functionalities like Manufacturing, Order management, inventory management, Supply chain, custody operations, sales and marketing and Medicaid Management information systems, provided application solutions, proposed and implemented various best practices.

Proficient in predictive machine learning models such as Linear-Logistic Regression, Decision trees, Support Vector Machines, KNN, Random Forest, XGBoost, clustering (K-means), dimensionality reduction (PCA and Singular Value Decomposition), recommendation systems (Collaborative filtering) and ensemble modelling techniques (Boosting, Stacking and Bagging).

Strong experience in Data Modeling, Data Analysis, Data Mining (DM), Machine learning (ML), Big Data and NoSQL technologies, prepared statistical models for org wise application maintenance estimations, process monitoring, and build model for understanding customer survey results.

Extensive hands-on experience in planning and leading of all phases of the Software Development Life Cycle including requirements gathering, analysis, design, architecture, construction, testing, deployment, production support transition and production support. Practiced waterfall and agile project management practices.

Have designed ETL frameworks for Audit Balance & Reconciliation and prepared docs for ETL design standards.

Worked with a special team to define the next generation delivery model that includes industrialized service model. Activities includes analyzing industry trends, defining the delivery model, assessing engagements for the gap analysis for the desired state, coach teams on the delivery model, train them on the tools, and facilitate implementation and help teams realize reduced cost of development and maintenance.

Technical Summary:

Programing Languages

R, Python, SQL, XML, COBOL

Databases

Oracle, DB2, NOSQL

Tools-Utilities

Rstudio, Github, ETL, Infor sphere, Data stage Quality stage, Erwin, File-Aid, KNIME

Technologies

Data Analytics, Machine Learning, EDW, ETL and Mainframes

Operating systems

UNIX, WINNT, Z/Os and Amazon EC2

Techniques

Regression, Classification, Clustering, Random forests, Bagging, Boosting and NLP

Industries

Retail, Logistics Supply chain, Banking and Medicaid

Professional Experience:

XXXX, Sacramento, CA (Sep 2016 – Present)

Problem statement:

To build and enhance pre-payment predictive modeling solution in California Medicaid Management Information System (CA-MMIS)

To minimize the impact of fraud, waste, and abuse in DHCS CA Medicaid payments, while helping ensure that healthcare benefits continue to be delivered to those who need them most.

Responsibilities:

Use k-nearest neighbor algorithm whose distance metric is optimized by a genetic algorithm in detecting two types of fraud: inappropriate practice of service providers and identifying “doctor-shoppers”

Prepare Ensemble scoring model by combining multiple models at various granularity levels, each has a unique ability to represents specific aspects of the data and maximize the prediction accuracy.

Use Association rules and a neural segmentation algorithm for fraud detection

Perform network and social link analysis, to identify a bad actor, and their associations, to identify anomalies and fraudulent schemes, Social Network Analysis identifies organized fraud activities by modeling relationships between entities.

Examine claims by member, provider, service, and other attributes with the goal of identifying and assigning an alert and risk score that prioritizes claims for further review, used fraud propensity scores to look for suspicious billing patterns

Detect mathematical formulas, to look for outliers, Improbable service sequences, Recurring referral patterns,

Perform provider peer comparisons which reveals patterns of abnormally high or uncommonly frequent billing for similar services

XXXX, NYC, NY (Aug 2015 – Aug 2016)

Problem statement:

To segment Customers of High net worth by considering behavioral needs and attitudinal measures related to financial literacy, willingness to accept financial advice, drivers of time, and emotive behaviors such as risk appetite for investing to market suitable banking products.

Responsibilities:

Considered 300+ customer attributes, spanning measures of investable income, age, geography, and socio-economic status

Understanding rationale behind problem statement, business Objectives from domain prospective and converted business problem statement into analytics the problem definition.

Connected to different sources, extracting the desired data, integrate, preprocess, clean data, replace missing values with most appropriate values.

Exploring & analyzing the data statistically and identifying the significant features for modeling. Perform feature engineering by exploring, summarize large size data of different shape using R-Programming.

Design and develop predictive models and explanatory models with most possible accuracy using regression (linear & logistic), classification, clustering, Random Forests and Ensemble Algorithms.

Text analysis for grouping high-net worth customer's e-mails, to feed into business process flow.

Evaluation of models for Bias and variance problem. Test and Evaluate models for best parameters using K folds.

Helping the business to interpret relations between attributes, performing hypothesis test and building Explanatory models.

XXXX, NYC, NY (Nov 2014 – Aug 2015)

Problem statement:

To build a Data analytics intelligent model, to classify large pools of loans, according to risk of each mortgage loan within given period for subprime and agency mortgages by considering various attributes such as zip code, FICO score, loan-to-value (LTV) ratio, initial interest rate, initial balance, type of mortgage, deal ID, time of origination, defaults, foreclosure, modification, real estate owned (REO) and prepayment events and use it to estimate risk of the current customer.

Responsibilities:

Data elements validation using exploratory data analysis (univariate, bivariate, multivariate).

Missing value treatment, outlier capping and anomalies treatment using statistical methods.

Create dummy variables for certain datasets to fit into the regression.

Select the variables to be used in the model, by performing forward stepwise regression, R-square and VIF values.

Fit a logistic regression model using the entire subprime dataset.

Evaluate predictions outputs, evaluate model’s performances, fine turned parameters used in the models and used the outcome of the as an input to build better models.

XXXX, Rochester, NY (April-2009 to March-2014)

Project Overview:

NAOSC-DWH (North America Outbound Supply Chain) Data warehouse solution is designed to develop ODS (Operational Data Store) and EDW to maintain Parts, Supplies and Equipment information for Logistics, Transportation, Shop Floor Operators to measure various business metrics. Build generic competency on DataStage suite and develop best practices in solution implementation that is to provide high end consulting & architecting services of IBM DataStage products. Enabling rapid design, development and deployment of data warehouse/ data mart by re-using routines and transforms and reusable frameworks.

Technology Overview, EDW-BI. consists of IBM Infosphere DataStage 8.7, AIX Server and Oracle 10g, Erwin. Applications transactional activities technologies consists of Client server applications using Java and oracle through multi-tier and legacy mainframe applications and its databases.

Key Roles and responsibilities as Senior Architect

Lead EDW development activity of 2 projects each of the project effort each 15K man hours.

Review and approve source-to-target mapping, solution design, data profiling rule's, SQL/ETL tuning & optimization

Defined ETL project scope, provided sizing and estimated for development and production support

Analyzed the data generated by business process, defined the granularity, finalized fact and dimension attributes.

Collaborate with business users to identify business needs/requirements and transform those to design/develop ETL data solutions and ensure best practices, policies and procedures are utilized.

Guided the team to use Information Analyzer to perform initial data quality assessment, identify & measure existing data defects by setting up the rules and used Data Quality stage to Cleanse, Correct & Match input data.

Lead the transition activity with least cost through India to India transition, online meetings for business discussions with clear agenda, achieved the cost effectiveness with proper planning and execution.

Works with client to establish priorities and severity levels and monitors work queues and makes work assignments to team members, ensuring a balanced workload and timely delivery.

Preparing Root Cause Analysis from client perspective and tracks actions to achieve higher service availability

Identifying the training needs (technical, functional, process) and preparing the team for future challenges

Ensures required commitment documentation is created, approved, and stored in the required repositories

Works with senior management/team to establish performance goals and gather and maintains metrics provides reports, provide suggestions for the improvements.

Actively involved in all the designs, actively help the team in all technical difficulties, automated repetitive tasks, tuned for performance data stage applications and achieved higher productivity

XXXX (Maintenance and Enhancements), SFO, CA (Oct-2001 to March-2009)

Project: XXXX (SCaN and PROMPT) & Roles: Technical lead, Team leader (Projects)

Engagement Overview:

SCaN (Supply Chain Navigator), is a web-based system for tracking and forecasting the movement of merchandise from distribution of a purchase order (PO) to the arrival of the merchandise at a Gap destination.

PROMPT (Purchase order management and Product tracking system), is a critical system which Keeps track of creation, approval and maintenance of POs, Maintenance of retail styles, Management of Vendors and invoice information and interacts with other GAP crucial systems.

Technology overview: It is a client server applications front end is Forte/JAVA. Back end is DB2, Transactional activities are performed through COBOL/CICS/DB2/MQ online and batch programs, technologies used in the projects include ESBXML, Mercury quality center, Oracle, ETL, data stage, TOAD and UNIX.

Key Roles and responsibilities in GAP Engagement

Setting up the meetings with the business to understand the requirements and proposing the cost effective technical solutions, involve them in user acceptance testing.

As a Techno Functional SME, of GAP supply chain IT, participated in several client initiated workshops as part of application modernization and reviewed various proposals,

Lead SCaN and PROMPT application support team worth 50K hours of annual maintenance activity.

Prepare the functional, technical and database designs, Coding the programs. Unit testing the changes, led the user acceptance testing and performance turned legacy applications, in multiple projects worth total man hours of 30K over the period of 48 months

Work closely with Development, Infrastructure, DBAs, Business Analysts, Testing, Other Functional, EDI Teams, Data integration, Conversion teams and Production support teams to monitor & track current tasks and projects.

Project: XXXX (Quality Assurance System): (Jan-2008 to Feb-2009)

Project Overview:

The objective of this project was, to build a centralized system of record (SOR) for the Merchandise QA audits of Gap Inc.

This application enables to select the shipments for raising an audit and recording the audited results into this system, this system is replacing current paper based inspection report for shipments at factory location conducted by QA auditor. Shipment related information will be flown from different systems like Purchase order management system, Vendor Profile system, Merchandise Hierarchy to QAS by using the ETL scheduled batch jobs, the total effort of this project was 7K man hours over a period of 12+ months.

Key Roles and responsibilities as a Tech lead

Prepared High level ETL technical specifications to translate business rules & functionality requirements.

Oversee day-to-day ETL development work of offshore resources, including design and code reviews and guided the team towards successful implementation

Created common components to be used across different streams and standardized the code

Responsible for data analysis, requirement gathering, source-to-target mapping, frequency analysis, process flow diagram and documentation

Developed complex data stage ETL jobs according to business requirements and defined its frequency.

Project: XXXX Profile GAP M&E: (Oct-2006 to Dec-2007)

Project Overview:

The objective of this project was, to build a centralized application (using MDM-Master Data Management solution) as system of record for all merchandising, non-merchandising vendors and Factories with which GAP is dealing, without disturbing the existing multiple applications/modules across different department functionalities, once it was build it publishes the profile details using standard XML EBO to the ESB queues for which other applications subscribed for. The total effort of this project was 8K man hours over a period of 12+ months

Key Roles and responsibilities as a Tech lead

Interacted with existing application's business community, understood and documented GAP vendor’s data life cycle.

Actively involved in collaborative discussions with business users and finalized attributes of proposed system entities and its relationships.

Involved in designing the Data model and prepared the mapping document for attributes of vendor profile application from that of various source systems and designed the logic to transform, defined the data selection priorities as it has same/similar attributes of different systems.

Designed and developed ETL process using Data Stage designer to load data from multiple source systems data.

Lead the conversion testing team and validated the data bridged to VP from all the destination systems.

Wrote the test plan and scripts for the conversion testing & lead testing team for system and integration testing.

Leading the development team both offshore and on-site, reviews all configurable and deliverables make sure all the work products are assigned tracked reviewed delivered and implemented.

Worked closely with the project manager, in preparing the project plans & track the status of the project.

Working closely with Development, Infrastructure, DBAs, Business Analysts, Testing, Other Functional, Conversion and Production support teams to monitor & track current tasks and projects.

Managing Stakeholders from project initiation through smooth Implementation

Escalation of critical issues to Management for their support to resolve the issues quickly.

Preparation of transition plan for production support team after implementation and lead the transition effort

Education and Certifications

Degree/ Certification

Qualification/Course

College / University / Institution

Year of Passing

PMP

PMP (Project Management Professional)

Project Management Institute PMI

2011

Masters

Master of Foreign Trade

Pondicherry university

2005

Bachelors

BTECH CIVIL ENGINEERING

Nagarjuna university

1990

DB2 DBA

DB2 DBA for Z/OS

IBM

2008

Other

Machine Learning

Stanford University thru Coursera

2013

Other

SAP CRM Training

Questivity-Santa Clara

2009



Contact this candidate