S u m m a r y
*+ years of Experience as Data Analyst with functional and industry experience with accomplished process and project responsibility such as Data analysis, design, development, quality assurance, user acceptance & performance management.
Experience in Data Modeling and Architecture, Database Administration, Data Conversion Validation, Data Warehouse Development, Report Creation, Software Quality Assurance, User Acceptance Testing, Training and Support.
Experience in Data Analysis, Data Migration, Data Validation, Data Cleansing, Data Verification, identifying data mismatch, Data Import, and Data Export through the use of multiple ETL tools such as Informatica PowerCenter.
Have Good knowledge on Python Collections and Multi-Threading.
Good experience of using Python libraries like Panda, Numpy, SciPy, Scikit-learn, Matplotlib, Seaborn.
Designed and implemented a business-critical system using Object oriented design and Programming concepts using Python.
Experience in creating Visualization and dashboards in Tableau desktop using Oracle and Teradata as data source.
Data Analysis-Data collection, data transformation and data loading the data using different ETL systems like
SSIS and Informatica.
Experience in administrative activates comprising Performance Tuning, Query Optimization, Client/Server Connectivity and Updating Statistics.
Experienced in developing Custom Report and different types of Chat Reports, Reports, Tabular Reports, Matrix Reports and distributed reports in multiple formats using SQL Server Reporting Services (SSRS).
Proficiency in multiple databases like Teradata, PostgreSQL, MySQL, ORACLE and MS SQL Server.
Good understanding of Relational Database Design, Data Warehouse concepts and methodologies.
Involved in writing shell scripts on UNIX for Teradata ETL tool and data validation.
Tracked defects using Bug tracking tools like Jira and Git.
Excellent Communication, interpersonal, analytical skills and strong ability to perform in a team as well as individually.
S k i l l s
Programming Languages: Python, SQL, PHP, CSS3, HTML5, C++, R
Python Libraries: NumPy, Pandas, Sci-kit Learn, Mat plot, Seaborn
Data Visualization: Looker, Power BI, Tableau, Advanced Microsoft Excel
Database: PostgreSQL, Microsoft SQL Server(T-SQL), MySQL, PostgreSQL, PL/SQL, MS Access, Teradata
Project Management: MS Project, SDLC, Agile, SCRUM. MS Visio, Waterfall.
Analytic Skills: Data Modeling, Statistical Analysis, Predictive Analysis, Unsupervised Learning, Supervised Learning, Informatica, SSAS, SSRS, SSIS
Other: JIRA, Git, MS Office, Windows, Unix
E d u c a t i o n
Masters of Science in Information System Sep 2017 – May 2019
New Jersey Institute Technology, Newark
E x p e r i e n c e
Humana, NJ Mar 2019 – Current
Utilized agile software development methodology to manage and control software and product development.
Created Data dictionary, Data mapping for ETL and application support, DFD, ERD, mapping documents,
metadata, DDL and DML as required.
Developed database objects like PL/SQL packages, Oracle tables, stored procedures, triggers that are used for meeting the business objectives.
Creation of multiple complex reports in SSRS and Power BI that run on high volume of data, with response time less than a few secs and which pulls data from SQL Server as well as PostgreSQL.
Used Power BI Power Pivot to develop data analysis prototype, and used Power View and Power Map to visualize reports.
Implemented discretization and binning, data wrangling: cleaning, transforming, merging and reshaping data frames using Python.
Used MLlib, Sci-kit Learn library to build and evaluate different models.
Used clustering technique K-Means to identify outliers and to classify un-labelled data.
Performed data analysis for testing that included supporting the ETL systems that interact with both OLTP & Data Warehouse System.
Executing the ETL Test Cases manually with the help of SQL scripts and UNIX Scripts.
Understanding and querying the master data tables to develop the KPI’s to measure the actual data quality to increase the efficiency of customer and vendor master data.
Lamda, India Jan 2015 – July 2017
Developed client desktop database and reporting applications from concept to Implementation SDLC with focus on user's need, scalability, reliability, supportability and security.
Performed Data Analysis, Data Validation, Data Cleansing, Data Verification, and Data Mismatch Identification on the data stored in data warehouse.
Monitored Performance and Optimized SQL Queries for maximum efficiency.
Used Python programs automated the process of combining the large datasets and Data files and then converting as Teradata tables for Data Analysis.
Experience in writing shell scripts for automating the jobs. Automated reports by connecting Teradata from
MS Excel using ODBC.
Performed Performance tuning of the SQL queries to increase the speediness in running query and generate the data accordingly by imposing proper Indexes and Creating Subqueries.
Created UML diagrams like Activity diagram, State diagram, Class diagram and Sequence diagram to meet the business needs and interacted with business users to understand their data needs.
Created reports using SSRS (SQL Server Reporting Services), SSAS (SQL Server Analytical Services).
Used DAX (Data Analysis Expressions) functions for the creation of calculations and measures in the Power BI
Automated recurring reports using SQL and Python and visualized them on BI platform like Tableau.
Extensively used ETL to load data from Excel spreadsheet, flat files into the target databases.
Managed bugs using Defect Tracking Tool JIRA, reviewed each defect and assigned development team to resolve problems.
Enhance the Resourcing Reports using power pivots, slicers and sparklines.
Involved in data modelling meetings with the Business stakeholders and Technical team.