Post Job Free

Resume

Sign in

Data Analyst

Location:
Manhattan, NY, 10001
Salary:
80/hr
Posted:
September 25, 2023

Contact this candidate

Resume:

Samuel Ebenezer Bolmalla

adzxwh@r.postjobfree.com

469-***-****

Professional Summary:

** + years of professional IT experience with domain knowledge in Banking & Finance, Healthcare, Insurance, Telecommunication, Retail, Media, and E-Ticketing

Experience in Data Architecture, Data modeling and Data Analysis at Enterprise Level

Performed data quality checks and involved in cleansing the incoming data feeds and profiled source systems data as per business rule using Informatica Data Quality

Extensive experience in EDW, Database Administration, MSBI and Analytics, Application design and development; and ETL design, implementation plan and test strategies

Experience as a Senior ETL developer from architectural perspective in SSIS using C# and Informatica tools.

Extensive experience in writing and supporting developers with stored procedures, SQL queries, UDFs, triggers using T-SQL and PL SQL programming.

Experience in SQL Server Installation, Configuration, Performance Tuning, Client/Server Connectivity, Query Optimization, Backup/Recovery, Running DBCC Commands and Database Mail Configuration

Experience in performance tuning, maintenance, and Administration by creating Jobs, Alerts, SQL Mail Agent, and schedule DTS/SSIS Packages.

Data Warehousing ETL exposure of using Informatica PowerCenter Client tools – Mapping Designer, Repository manager, Workflow Manager

Excellent in design, development of ETL SSIS Packages for integrating data using OLE DB connection from heterogeneous sources like (Excel, CSV, Oracle, flat file, Text Format Data) by using multiple transformations provided by SSIS such as Data Conversion, Conditional Split, Bulk Insert, Merge, and union all.

Experience in designing reports using SQL Server Reporting Services (SSRS) and creating various types of reports like drill down, drill through, Parameterized, Cascading, Conditional, Table, Matrix, Chart and Sub Reports. Designed Dashboards, Scorecards and reports using multiple KPI’s, OLAP charts and SSRS Reports

Working experience in Big data and Hadoop environments (Impala, Hive, Scala, and Spark Commands)

Hands on experience in AWS (S3 bucket) and Azure Technologies with Power BI.

Experience in ER Modeling and Dimensional Modeling for OLTP and OLAP systems.

Expertise in creating Databases, Views and User defined data types, functions, Cursors, and Indexes.

Excellent analytical, communication and interpersonal skills. Proficient in technical writing with presentations and a good team player with Agile methodologies.

Experience working with Team Foundation Server (TFS and SCCM)

TECHNICAL AND DOMAIN SKILL

Database : MS SQL Server 2019, Oracle 12c & Oracle 9i, Teradata, Bigdata, Hadoop, Impala

ETL : SSIS, Informatica, DataStage

Data Modeling Tools: Erwin, MS Visio and MS Excel -FSDM

Reporting : SSRS, Tableau, QlikView, Business Objects

Web Application Development: Net Framework 4, ASP.Net, Workflows, WCF Web Services

Operating Systems : All version of Windows OS

Other Tools/Utilities : Varicent (ICM tool), Infragistics, System Centre Configuration Management.

Programming languages : C#, Visual Basic, HTML5, CSS3, JavaScript, Power shell, Scala, Java, Python, TSQL

EDUCATION

MS in Computer Science Osmania University, Hyderabad

BS in Computer Science Osmania University, Hyderabad

PROFSSIONAL EXPERIENCE:

Client : Flagstar Bank, NY July 2023 to Sep 2023

Project: CIO – Business Intelligence

Role : Lead Data Analyst/SQL Developer

Description: CIO-BI is a Business Intelligence group serves is an integration component of the bank’s data warehouse bringing in and distributing data across organizations to support loan origination and servicing systems, CCAR -regulatory data and maintains credit life cycle.

Responsibilities:

Understand the architecture of the Loan Origination and Servicing systems data model and gather requirements for the migration of the data.

Load data using SSIS packages from nCino upstream system and provide it to the downstream systems for reporting purposes.

Used Pandas and NumPy for developing various machine learning algorithms and utilized machine learning algorithms such as linear regression, multivariate regression, naive Bayes, Random Forests, K-means, & KNN for data analysis.

Used pandas, NumPy, seaborn, scipy, matplotlib, scikit-learn, NLTK in Python for developing various machine learning algorithms.

Responsible for building reports in Power BI from the Scratch.

Create views using SQL and TSQL to provide data for tableau reporting teams.

Write code to identify loans with high-risk rate combining data from multiple source systems.

Coordinate with organizations internal data teams to synchronize data standards and data governance policies.

Work on the current schemas and create new database objects to enhance and extend the data warehouse.

Create Stored Procedures and views using SQL programming language to perform scan of the new files against existing customer records using temporary tables.

Managed relationship between tables in Power BI using star schema.

Create column mapping document to load data from IBM/COBOL flat files to SQL Server destinations using file layout documents.

Load data using derived columns, data conversion using ExecuteSQL tasks and Dataflow tasks.

Script task and C# programming to perform file validations and access the text and streams of the file to perform ETL with data governance standards.

Involved in troubleshooting batch jobs hosted on SQL Server Agent using Master package and custom audits.

Involved in change release management for CAB approvals and coordinate with DBA team for deployments.

Resolve production data issues daily to maintain business continuity.

We work closely with Stakeholders and product owners to deliver expected requirements.

Technical Environment: MS SQL Server, Python, Oracle, nCino and Tableau

Client : Charles Schwab, NJ Sep 2020 to June 2023

Project: AST - Database Services Group and GCP Cloud Datawarehouse

Role : Database Lead

Description: AST Data Services Group provides distributed database support to AST (Advisor Services Technology) applications and Systems which includes onboarding of the new clients, data delivery with end-to-end services and support back-office applications. It also includes infrastructure related support with migrations and performance monitoring. Provides support for risk, audit, and regulatory purposes.

Responsibilities:

Write SQL queries, Stored Procedures and create views using T-SQL and PL/SQL to onboard new clients and support existing customers.

Analysis of the database architecture to create data delivery and data flow documents.

Understand and document functional specifications of the requirement.

Migration activities from Teradata to Google Cloud Platform Big query database

Involved in troubleshooting and enhancing queries in Teradata and Big query to perform ETL loads.

Troubleshoot SSIS packages and run the loads manually fixing the error in the input file.

Support currently running AST databases and involve in enhancement and performance tuning of the existing systems.

Generating various capacity planning reports (graphical) using Python packages like NumPy, matplotlib.

Create ETL packages to generate daily reports and monthly reports.

Automate the delivery of the data by making needed changes in the configuration tables and route tables using XML.

Clean data and processed third party spending data into maneuverable deliverables within specific format with Excel macros and python libraries such as NumPy, SQL Alchemy and matplotlib.

Create column mapping document and ICD documents.

Execute pre-implementation and post-implementation queries to perform validation.

Create PL/SQL packages and deploy code in non-prod environments using Jenkins and check-in the code in TFS.

Migration of production data after performing thorough dry runs

Job monitoring on SQL Server Agent and Control M tool and fix if any issues.

Provided inputs to Informatica team to create mappings and mapplets to transform the data according to business rules.

Attend regular daily calls/meetings with Application Owners, Sales Team & Stakeholders

Technical Environment: Oracle, MS SQL Server, Python, Linux

Client : FedEx, Memphis TN Dec 2019 to Jun2020

Project: ECDS (Enterprise Customer Data Services)

Role : Data Analyst

Description: ECDS maintains the data across enterprise for all the customer accounts. Primarily deals with customer account management.

Responsibilities:

Analysis of the Upstream data model and data flow within the enterprise

Understand and document functional specification document.

Support currently running Applications database and involve in enhancement and performance tuning team.

Write SQL Queries, Stored Procedures and create views and perform data validation.

Incorporate ETL strategies to load the data addressing data quality and data governance.

Update data dictionary to map columns and create documents to load ETL data.

Provided design and mapping documents for Informatica team.

Work with Sales Team and MDM teams to implement data quality.

Trouble shoot defects raised during loading and testing in a timely manner.

Attend regular daily calls/meetings with Application Owners, Sales Team & Stakeholders

Technical Environment: Oracle, Teradata, MS SQL Server, DataStage

Client : Department of Transportation, Washington D.C Sep 2018 to Nov 2019

Project: National Registry of Certified Medical Examiners

Role : Database Lead

Description: The National Registry of Certified Medical Examiners (National Registry) is owned by DOT’s Federal Motor Carrier Safety Administration (FMCSA) and implements regulations requiring commercial motor vehicle drivers to be examined by a certified medical examiner to be deemed medically qualified to operate a commercial motor vehicle. There are also three testing organizations who exchange XML files in batch with National Registry to allow medical examiners to take the certification test and to transmit test results.

Responsibilities:

Analysis of the legacy National Registry data model and current method of data migration from legacy system

Creation of the normalized data model for the new National Registry application

Understand and translate business needs into data models supporting long term solutions.

Work with Application Development team to implement data strategies, build data flows or pipelines and develop conceptual data models.

Create logical and physical data models using best practices.

Creating SQL Queries and Stored Procedures, troubleshooting and Query Performance Tuning.

Create scripts & repeatable processes to migrate data from legacy to target data model.

Setup process of reverse engineering of physical data models from databases and SQL scripts

Work with business analysts to map data from legacy to target data model.

Support Testing Engineers during system testing phase and Support QA and user group in acceptance testing

Provide fixes for the defects raised during testing in a timely manner.

Analyze impact to change in data model, make modifications to the data model and migration scripts.

Implement the target data model, views, and stored procedures.

Migrate Production Data from the legacy National Registry to the New National Registry system.

Attend regular scrum calls/meetings with Application Owners & Stakeholders

Technical Environment: Oracle, MySQL, Appian Cloud/AWS S3

Client: JPMC Bank, Newark DE Sep 2017 to Sep 2018

Project: CCAR/CCEL

Role: Data Analyst

Description: CCAR modeling process includes decisions about the types of quantitative models to use, scenario and model development, champion/challenger model creation, results reporting, and integration with the operating plan. enable management to review or challenge every step of the capital planning process – including scenario creation, loss modeling and net interest income projections – not just the results. The forward-looking projections help bank management understand the cause and effect of the stress test results.

Responsibilities:

Created views to populate data as per the business requirement and build the semantic layer from the base tables based for the modeling team data consumption.

Worked on defects generated from Informatica Data quality tool.

Developed workflows in Informatica Data quality for validation and cleansing.

Responsible for requirement gathering and analysis to prepare Interface Component Document with the business logic.

Create validation reports at various stages of the data load from source systems (Teradata) to target (Hadoop) Environment

Was part of Risk and Finance Integration Systems team to process data used for CCAR data modeling process

Contributed for CDD (Conceptual Design Document) for the implementation of CECL with data collection for the Upstream and Downstream systems from architectural perspective.

Write functional specification document for additional data load requirement on ad-hoc basis.

Create mapping documents and give columns description, data acquisition forms from meta data repository, Data Dictionary.

Technical Environment: Teradata, Hadoop, Impala, MS Office

M & T Bank, Buffalo, NY July 2016 to July 2017

Data Modeler/EDW

Description: The CLM (Credit Life Cycle Management) program is intended to significantly improve the banks’ ability to process, manage, monitor, and report, on the credit portfolio from origination through resolution.

The creation of an operational and analytic data environment that supports the nCino software application, robust portfolio analytics, and operational reporting is critical to the success of the program. These efforts result in increased efficiency, improved standardization, and reduced risk. There are three planned releases of nCino and the data environment will evolve in each release. This document is focused primarily on the first release requirements.

Responsibilities:

Responsible for data analysis and data mapping along with defining the Meta data repository for Enterprise Data Warehouse using Teradata from MDM/Enterprise party Hub

Worked and coordinated with Informatica team for data load design and ETL process.

Creating views to insert the data from the upstream databases based on the business logic.

Involved in writing the stored procedures with PLSQL programming to create views and load data.

Responsible to design the FSDM Data model/Dimensional model and nCino data model while working to develop Operational Data Store

Contributed for SAD (System Application Design) document for the implementation of ODS with BRD reviews and PDD reviews.

Analyze the design of the current systems with data flow/interactions to propose options for future systems to store design on the historical data for the retiring applications used in Credit Life Cycle Management

Analyze the existing accountings systems to populate data marts used for CCAR reporting and Loan /Case management reports.

Enhancement and maintenance of the existing data models for multiple projects

Worked with the Development team and QA Team to develop standards to ensure compliance with requirements.

Used MS Visio for designing flow-charts, process models and architectural design of the application.

Developed logical/ physical data models in Erwin based on the specifications and established referential integrity on the system.

Done Data mapping work between legacy and prospect systems using Excel extracts.

Carried out PL/SQL queries to validate the extraction and loading to data warehouse.

Design a high-level conceptual enterprise data model and data analysis using Erwin, MS Visio and Excel on current systems of the bank.

Technical Environment: MS SQL Server 2012/Oracle/Erwin

Microsoft Corporation, Redmond-WA Jan 2016 to June 2016

MSBI Developer

Description: Partner Incentives project offers various programs to the partners across the globe considering numerous factors on revenue, region, agreements and subscriptions to the incentives to the partners in the form of credit note.

Responsibilities:

Development of ETL packages to extract, transform and load into the staging database schemas External sources like Amazon others database servers.

Involved in writing the stored procedures with TSQL programming to load and correlate the data used for reporting services.

Involved in gathering the requirements to integrate new data sources to design column mapping document.

Involved in coding the test automation scripts for validation purposes.

Responsible for managing the data analysis, data sources, data mapping along with defining the meta data repository.

Migration of the existing Push packages to SQL 2014 and performance improvement.

Implement the data extract methodologies to extract and load the data from various sources.

Technical Environment: SSIS, SSRS, SQL Server 2012, Net Framework 4, C#,

Western Union Jun 2013 to Nov 2015

Sr. BI Developer

Project Description: Western Union involved in Money Transfer business activities. Western Union has various Agents spread across the globe geographically to maintain business activities. Western Union maintains various portals and application to do the transactions internally and externally. They monitor the business data using graphs which extract the data from various sources.

Responsibilities:

Migrated DTS packages from SQL Server 2005 to SQL Server 2008 as SSIS Packages.

Involved in designing, developing, debugging, and testing of reports in SQL Server Reporting Services (SSRS).

Experience in creating Parameterized reports and Linked reports with thorough knowledge of report serving architecture. (Table, chart and matrix report).

Used SSIS to unite data from existing system and performed transformations on MS SQL 2008.

Extract Transform Load (ETL) development Using SQL server 2008, SQL 2008 Integration Services (SSIS)

Generated drill down and drill through reports using SSRS.

Technical Environment: SSIS, SSRS, SSAS SQL Server 2012, Net Framework 4, C#

Microsoft Corporation (Offshore) June 2011 to May 2013

Sr. MSBI Developer

Description: Data Management Platform is a business intelligence solution that is enabled to maintain the data of the Microsoft Customers and Partners that is spread across the 13 regions worldwide.

Microsoft product sales are being made through various internal applications by creating a user account and based on the various business rules data gets updated in the External Systems and at the backend.

Responsibilities:

Created ETL metadata reports using SSRS, reports include like execution times for the SSIS packages, Failure reports with error description.

Develop Dimensions and Measures for the Sales of the Microsoft Corporation

Created database objects like tables, users, roles, stored procedures, triggers.

Used various formatting techniques like interactive sorting, grouping to created user friendly reports.

Participated in initial knowledge transfer activities of the application.

Support the existing SSAS Cubes and administration.

Capture day to day activities of the sales and marketing team (Calls and Subscriptions)

Involved in acquiring the requirements from the customer and consolidating them.

Developed, deployed, and monitored SSIS Packages using various transformation tasks.

Unit Testing the SSIS packages with respect to its flow and performance issues.

Designed SSIS Packages to extract, transfer, load (ETL) existing data into SQL Server from different environments for the SSAS cubes.

Involved in designing, developing, debugging, and testing of reports in SQL Server Reporting Services (SSRS).

Technical Environment: ASP.NET Technologies (MVC 3), SSRS, SSIS, SQL Server 2008, .Net Framework 4.0

GTA - Presto May 2009 to April 2011

Sr. Software Engineer

Description: The Greater Toronto Area Fare System (GTAFS) is a regional automated fare collection system that will establish a single fare-card based on contact-less smartcard technology that will allow cardholders to access any of the GTA’s nine transit service providers providing multiple modes of transit (e.g., trains, subways, buses).

Responsibilities:

Responsible for configuration and cube deployment in SSAS

Reprocess the partitions and cubes.

Responsible for Deployment Automation (Workflow 4) using .Net and Power shell scripting.

Developed tools in compliant with Microsoft SCCM

Responsible as the primary owner for the automation tools used for Build and Deployment in TFS

Involved in the defect fixing of SSIS packages in Insurance Domain

Involved in writing SQL Server Queries to meet business requirements.

Developed Calculations on business criteria in Varicent tool.

Technical Environment: Visual Studio 2010, .NET 3.5, Biztalk 2009, Dynamics, SharePoint, SSRS, SSAS, SSIS, WCF, Workflow, SQL Server 2005, Team Foundation Server

T-Mobile Jun 2007 to May 2009

Software Engineer

Description: Prithvi Domain Framework (PDF) is a generic framework that can be used as a base platform for all the business domain development activities using Microsoft.Net 3.5 or higher version for both Windows and Web applications.

Responsibilities:

As a Team Member I was responsible for Coding and Code review

Technical Environment: Visual Studio 2008, .NET 3.5, SQL Server 2005, VSS

SRR Institutions Jul 2003 to Apr 2007

Involved in training the Databases and Programming Languages

Coordinate and assign project work to students.

Develop internal projects for the institution and department.



Contact this candidate