Post Job Free

Resume

Sign in

Data Engineer

Location:
West Windsor Township, NJ, 08550
Posted:
July 26, 2021

Contact this candidate

Resume:

Varun Reddy

Phone: 732-***-****

adnwnz@r.postjobfree.com

PROFESSIONAL SUMMARY:

* ***** ** ********** ** Data warehousing and the Analysis, Design, Development, Testing and Implementation of business application systems for financial, insurance, medical and e-commerce

Proficient in managing entire project life cycle and actively involved in all the phases of project life cycle including data acquisition, data cleaning, data engineering, features scaling, features engineering

Expertise in transforming business requirements into analytical models, designing algorithms, building models, developing

data mining and reporting solutions that scale across a massive volume of structured and unstructured data.

Skilled in performing data parsing, data manipulation and data preparation with methods including describe data contents, compute descriptive statistics of data, regex, split and combine, Remap, merge, subset, reindex, melt and reshape.

Extensive experience in Text Analytics, generating data visualizations using Python and creating dashboards using tools like Tableau.

Experience with Data Analytics, Data Reporting, Ad-hoc Reporting, Graphs, Scales, PivotTables and OLAP reporting.

Good Knowledge in Proof of Concepts ( PoC's ), gap analysis and gathered necessary data for analysis from different sources, prepared data for data exploration using data munging.

Good industry knowledge, analytical &problem-solving skills and ability to work well within a team as well as an individual.

Highly skilled in using visualization tools like Tableau, ggplot2, dash for creating dashboards.

Worked and extracted data from various database sources like Oracle, SQL Server, DB2, regularly accessing JIRA tool and other internal issue trackers for the Project development.

Highly creative, innovative, committed, intellectually curious, business savvy with good communication and interpersonal skills.

Extensive experience in Data Visualization including producing tables, graphs, listings using various procedures and tools such as Tableau.’

Strong experience in the Analysis, design, development, testing and Implementation of Business Intelligence solutions using Data Warehouse/Data Mart Design, ETL, OLAP, BI, Client/Server applications

Strong Data Warehousing ETL experience in using Informatica 9.6.1/9.5/9.1/8.6/8.5/8.1/7.1/7.0 Power Center Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor and Server tools, Informatica Server, Repository Server manager

Expertise in Data Warehouse/Data mart, ODS, OLTP and OLAP implementations. Teamed with project scope, Analysis, requirements gathering, data modeling, Effort Estimation, ETL Design, development, System testing, Implementation and production support.

Experienced in Repository Configuration using Transformations, creating Informatica Mappings, Mapplets, Sessions, Worklets, Workflows, Processing tasks using Informatica Designer / Workflow Manager to move data from multiple source systems into targets.

Strong working experience in Informatica Data Quality (IDQ) 9.5 with Informatica Developer tool, Analyst tool

Extensively worked on Informatica Power Center Transformations such as Source Qualifier, Lookup, Filter, Expression, Router, Joiner, Update Strategy, Rank, Aggregator, Stored Procedure, Sorter, Sequence Generator, Normalizer, Union, and XML Source Qualifier.

Strong experience in Dimensional Modeling using Star and Snow Flake Schema, Identifying Facts and Dimensions, Physical and logical data modeling using ERwin and ER-Studio.

Extensively worked on data extraction, Transformation and loading data from various sources like Oracle, SQL Server, Teradata and files like Flat files.

WORK EXPERIENCE

TECHNICAL SKILLS

Databases : MS SQL Server, Oracle, MySQL, PostgreSQL, Snowflake SQL

Database Tools : MS SQL Server Management Studio (SSMS), Oracle SQL Developer Data Modeler, Erwin Data Modeler, MySQL Workbench, MS Visio, Teradata Studio

Data Warehousing : Snowflake, Teradata

Cloud technology : AWS redshift,Ec2,S3,Azure data lake storage, Azure data factory ETL Tools : Informatica

BI Tools : Tableau, MS SQL Server Reporting Services (SSRS) Programming : C, C++

Scripting : Python, Perl, UNIX

Python frameworks : numpy, pandas, matplotlib, seaborn, plotly, iplot, sklearn, scipy, pytables Data Visualization : Tableau, Python (matplotlib, seaborn, plotly)

Operating systems : Windows, UNIX, LINUX Project management : Waterfall, Agile

WORK EXPERIENCE

Data Engineer– Wellmark, Des Moines IA, USA Mar 2019 – present

Provided data engineering solutions for health Insurance Analytics and also automate some of the processes involved

Worked with data governance team to provide and validate the necessary key codes for fields

Utilized analytical skills and worked with databases like SQL,DB2,Oracle and teradata to develop and maintain data driven solutions for customers; Built data pipelines using ETL tool Informatica and powerexchange

Translated business propositions into quantitative queries and collected/cleaned the necessary data.

Developed a data pipeline using Delta Lake that led to a process optimization and corresponding revenue increase.

Did Data Profiling using Python, Data Modeling using Erwin Data Modeler.

Performed data cleaning and analysis on mainframe files. Created modularized code for init, clean, enrich and load.

created daily load validations and automated checks to email out results to theteam using shell scripting

Used cloud based AWS to load data. Experienced with AWS redshift, S3 and EC2 components of AWS

Worked on migration of datawarehouse from Teradata to cloud based redshift to improve scalability and performance

Used aws S3 for cloud based storage and backup of archive files

Deployed code into the cloud using aws lambda service for ease of operations.

Data Analyst – Davita Malvern, PA Apr 2018 – Feb 2019

Worked on analysis scripts for Kidney dialysis Reporting Team

Migrated business critical scripts responsible for Business Growth Monitoring

Used matplotlib & seaborn to generate reports by Business Segments. Performed Code Validation using custom Python scripts.

Created coherent Logical Data Models that helped guide important client business decisions

Worked at conceptual/logical/physical data model level using Oracle SQL Developer .

Performed data integration by creating data pipelines for smooth transfer of HL7 files which had various message segments like MSH,PID,PV1,NK1 etc

Created and executed the test plans, test cases, unit and integration testing

Worked with Google Cloud platform to load data and increase the data storage efficiency

Elicited, analyzed and documented requirements for implementation of the trading system in the Agile Environment

Created python scripts to communicate analytical results to senior management for strategic decision making

Worked on Azure based cloud environment for data storage and retrieval.

Data Analyst – Cinergytech,NJ Jan 2018 – Mar 2018

Performed Data mining, handled missing values, removed outliers using Proximity Distance and Density based techniques

Assisted users in creating/modifying worksheets and data visualization dashboards in Tableau.

Tuned and performed optimization techniques for improving report/dashboard performance.

Assisted report developers with writing required logic and achieve desired goals.

Parameterized the mappings and increased the re-usability.

Extracted data from oracle database and spreadsheets and staged into a single place and applied business logic to load them in the central oracle database.

Graduate Research Assistant at eLab – ASU Jan 2017 – Dec 2017

Developed IoT human-machine interaction (HMI) toolbox. (to be released as open-source https://goo.gl/gU3Fdb)

Created data acquisition framework, developed FSM based algorithms in Python.

Developed functionality for calculating the next velocity for the motion of the wheel chair.

Prototyped an HMI system consisting of gesture, BMI and speech based interaction to control the maneuver of an electric wheelchair.

ETL Developer – TCS,Hyderabad,India Apr 2012– Nov 2014

Migrated data from Excel, Flat file and XML to MS SQL Server by using SSIS. Followed 2NF/3NF DB Normalization standards to create database.

Analyzed the business requirements and functional specifications.

Extensively used shell scripting for filtering and cleaning various text files.

Performed day to day unix maintenance jobs, user creations, file system creations involving LVM’s password handling, server building

Used Informatica Power Center 8.6 for extraction, transformation and load (ETL) of data in the data warehouse.

Extensively used Transformations like Router, Aggregator, Normalizer, Joiner, Expression and Lookup, Update strategy and Sequence generator and Stored Procedure.

Developed complex mappings in Informatica to load the data from various sources.

EDUCATION

Master of Science, Arizona State University Dec 2016

Bachelor of Engineering, JNTU,Andhra Pradesh India Apr 2011



Contact this candidate