Post Job Free

Resume

Sign in

Data Engineer

Location:
Port Mercer, NJ, 08540
Posted:
March 15, 2021

Contact this candidate

Resume:

Varun Reddy

Title : Data Engineer

adkxq4@r.postjobfree.com

Phone # : 404-***-****

PROFESSIONAL SUMMARY:

A highly seasoned professional with 8 Years of experience in Data warehousing and the Analysis, Design, Development, Testing and Implementation of business application systems for insurance, financial, medical and e-commerce

Proficient in managing entire data science project life cycle and actively involved in all the phases of project life cycle including data acquisition, data cleaning, data engineering, features scaling, features engineering

Expertise in transforming business requirements into analytical models, designing algorithms, building models, developing data mining and reporting solutions that scale across a massive volume of structured and unstructured data.

Skilled in performing data parsing, data manipulation and data preparation with methods including describe data contents, compute descriptive statistics of data, regex, split and combine, Remap, merge, subset, reindex, melt and reshape.

Extensive experience in Text Analytics, generating data visualizations using Python and creating dashboards using tools like Tableau.

Experience with Data Analytics, Data Reporting, Ad-hoc Reporting, Graphs, Scales, PivotTables and OLAP reporting.

Good Knowledge in Proof of Concepts (PoC's), gap analysis and gathered necessary data for analysis from different sources, prepared data for data exploration using data munging.

Good industry knowledge, analytical &problem-solving skills and ability to work well within a team as well as an individual.

Highly skilled in using visualization tools like Tableau, ggplot2, dash for creating dashboards.

Worked and extracted data from various database sources like Oracle, SQL Server, DB2, regularly accessing JIRA tool and other internal issue trackers for the Project development.

Highly creative, innovative, committed, intellectually curious, business savvy with good communication and interpersonal skills.

Extensive experience in Data Visualization including producing tables, graphs, listings using various procedures and tools such as Tableau.’

Strong experience in the Analysis, design, development, testing and Implementation of Business Intelligence solutions using Data Warehouse/Data Mart Design, ETL, OLAP, BI, Client/Server applications

Strong Data Warehousing ETL experience in using Informatica 9.6.1/9.5/9.1/8.6/8.5/8.1/7.1/7.0 Power Center Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor and Server tools, Informatica Server, Repository Server manager

Expertise in Data Warehouse/Data mart, ODS, OLTP and OLAP implementations. Teamed with project scope, Analysis, requirements gathering, data modeling, Effort Estimation, ETL Design, development, System testing, Implementation and production support.

Experienced in Repository Configuration using Transformations, creating Informatica Mappings, Mapplets, Sessions, Worklets, Workflows, Processing tasks using Informatica Designer / Workflow Manager to move data from multiple source systems into targets.

Strong working experience in Informatica Data Quality (IDQ) 9.5 with Informatica Developer tool, Analyst tool

Extensively worked on Informatica Power Center Transformations such as Source Qualifier, Lookup, Filter, Expression, Router, Joiner, Update Strategy, Rank, Aggregator, Stored Procedure, Sorter, Sequence Generator, Normalizer, Union, and XML Source Qualifier.

Strong experience in Dimensional Modeling using Star and Snow Flake Schema, Identifying Facts and Dimensions, Physical and logical data modeling using ERwin and ER-Studio.

Extensively worked on data extraction, Transformation and loading data from various sources like Oracle, SQL Server, Teradata and files like Flat files.

WORK EXPERIENCE

TECHNICAL SKILLS

Databases: MS SQL Server, Oracle, MySQL, PostgreSQL, Snowflake SQL

Database Tools: MS SQL Server Management Studio (SSMS), Oracle SQL Developer Data Modeler, Erwin Data Modeler, MySQL Workbench, MS Visio, Teradata Studio

Data Warehousing: Snowflake, RedShift ETL Tools : Informatica

BI Tools: Tableau, MS SQL Server Reporting Services (SSRS)

Programming : C, C++

Scripting: Python, Perl, UNIX

Python frameworks: numpy, pandas, matplotlib, seaborn, plotly, iplot, sklearn, scipy, pytables

Supervised ML: Linear Regression, Logistic Regression, Classification modeling, Neural Network, KNN Classification, Decision Trees, Sentiment Analysis, Time Series Analysis

Unsupervised ML: K-means clustering, Anomaly Detection, Recommender Engine Data Visualization : Tableau, QlikView, Python (matplotlib, seaborn, plotly) Operating systems : Windows, UNIX, LINUX

Project management: Waterfall, Agile

WORK EXPERIENCE

Data Engineer– Wellmark, Des Moines IA, USA

Mar 2019 – Sep 2020

Provided data engineering solutions for Insurance Analytics & Credit and also automate some of the processes involved

Utilize analytical skills and work with technologies like Python, SQL, Snowflake to develop and maintain data driven solutions for customers.

Participated in all phases of data mining; data collection, data cleaning, developing models, validation, visualization and performed Gap analysis.

Extracted data from files and prepared data for exploratory analysis using data munging

Converted raw data to processed data by merging, finding outliers, errors, trends, missing values and distributions in the data.

Translated business propositions into quantitative queries and collect/clean the necessary data.

Did Data Profiling using Python, Data Modeling using Erwin Data Modeler.

Performed ETL & Load using Created modularized code for init, clean, enrich and load.

Also created python framework for daily load validations and automated checks to email out results to the team.

business requirements using Workflow Diagrams, Sequence Diagrams, Activity Diagrams and Use Case Modeling

Data Analyst – Davita Malvern, PA Apr 2018 – Feb 2019

Worked on analysis scripts for Kidney dialysis Reporting Team

Migrated business critical scripts responsible for Business Growth Monitoring

Used matplotlib & seaborn to generate reports by Business Segments. Did Code Validation using custom Python scripts.

Clean data and processed third party spending data into maneuverable deliverables within specific format with excel macros and python libraries such as NumPy

Used Pandas as API to put the data as time series and tabular format for manipulation and retrieval of data.

Generating various capacity planning reports (graphical) using Python packages like Numpy, matplotlib.

Created coherent Logical Data Models that helped guide important client business decisions

Worked at conceptual/logical/physical data model level using Oracle SQL Developer Data Modeler

Analyzing various logs that are been generating and predicting/forecasting next occurrence of event with various Python libraries.

Created and executed the test plans, test cases, unit and integration testing for all PLSQL.

Elicited, analyzed and documented requirements for implementation of the trading system in the Agile Environment

Created python scripts to communicate analytical results to senior management for strategic decision making

Data Analyst – Cinergytech,NJ Jan 2017 – Mar 2018

Performed Data mining, handled missing values, removed outliers using Proximity Distance and Density basedtechniques

Assisted users in creating/modifying worksheets and data visualization dashboards in Tableau.

Tuned and performed optimization techniques for improving report/dashboard performance.

Assisted report developers with writing required logic and achieve desired goals.

Parameterized the mappings and increased the re-usability.

Extracted data from oracle database and spreadsheets and staged into a single place and applied business logic to load them in the central oracle database.

Graduate Research Assistant at eLab – ASU Jan 2016 – Dec 2017

Developed IoT human-machine interaction (HMI) toolbox. (to be released as open-source https://goo.gl/gU3Fdb)

Created data acquisition framework, developed FSM based algorithms in Python.

Developed functionality for calculating the next velocity for the motion of the wheel chair.

Prototyped an HMI system consisting of gesture, BMI and speech based interaction to control the maneuver of an electric wheelchair.

ETL Developer – TCS,Hyderabad,India Apr 2012– Nov 2014

Migrated data from Excel, Flat file and XML to MS SQL Server by using SSIS. Followed 2NF/3NF DB Normalization standards to create database.

Analyzed the business requirements and functional specifications.

Extensively used shell scripting for filtering and cleaning various text files.

Performed day to day unix maintenance jobs, user creations, file system creations involving LVM’s password handling, server building

Used Informatica Power Center 8.6 for extraction, transformation and load (ETL) of data in the data warehouse.

Extensively used Transformations like Router, Aggregator, Normalizer, Joiner, Expression and Lookup, Update strategy and Sequence generator and Stored Procedure.

Developed complex mappings in Informatica to load the data from various sources.

EDUCATION

Master of Science, Arizona State University Dec 2016

Bachelor of Engineering, JNTU,Andhra Pradesh India Apr 2011



Contact this candidate