Post Job Free

Resume

Sign in

Data Digital Marketing

Location:
Irvine, CA
Posted:
July 19, 2018

Contact this candidate

Resume:

HARSHAVARDHAN REDDY K

BI Developer

+1-949-***-****

ac6b2s@r.postjobfree.com

I am based out of Irvine, CA and I have 7+ years of experience in the area of Data Warehousing and Analytics using Tableau, Pentaho, Jaspersoft, and Talend. I’ve also extensively worked with Cloud-based platforms like AWS Glue, Athena, EMR, EC2 Quicksight, Redshift and S3. Very much hands on with the Big Data technologies like Hadoop, Map Reduce, Pig, Sqoop, Hive, Spark.

Professional Summary

Well versed in all phases of data warehouse development lifecycle, from gathering requirements to testing, implementation and support. Key strengths in using best Data Warehousing practices and Data Re-engineering with proven ability in the design, analysis and management of Enterprise scale BI/DW systems.

Proven ability in planning and implementing the overall analytics and business intelligence strategy, also ability to lead the design and development of BI projects designed to understand key business behaviors that drive customer acquisition, retention and engagement.

Rich working experience in Data Visualization with tools like Tableau, Tibco Spotfire and Amazon QuickSight. Have worked on integrating 3rd party charting libraries to devise custom visualizations as needed.

Being a Certified Pentaho BI Developer I’ve implementated a range of BI solutions right from Stand-alone reporting to Cloud based Multi-tenant solutions.

Also as a Certified Talend Data Integration Engineer I’ve vast experience in Data Integration and have sourced data from various sources including DBs, Flat files, 3rd party APIs, Python scripts, cloud-based data pipelines, Scrapper programs, etc.

Performing systems analysis & suggesting alternatives to improve and enhance the services & delivery standards.

Taken up initiatives in Performance benchmarking and improvements.

Consummate professional with excellent communication, planning, execution, monitoring and problem solving skills.

Strong Exposure on Lending and Finance domain, Energy IoT Analytics, Digital Marketing and Education and Career Services

Agile enthusiast and Agile Enthusiast and active participant in all Agile ceremonies

Strong interpersonal and communication skills with an ability to lead a team, collaboration with off-shore teams and coordination of release activities with DevOps team.

Technical Skills:

Big Data Stack

Map Reduce, Pig, Sqoop, Hive, Spark, AWS EMR, AWS S3, AWS Data pipeline etc.

BI Platforms

Tableau server, Tibco spotfire, Pentaho BI server, Jasper server, Mondrian OLAP server

Visualization and Reporting

Tableau, AWS Quicksight, Spotfire, Pentaho Dashboarding and CDE, Saiku Analytics, Jasper, BIRT, Fusion charts, Highcharts, D3 charts.

Data Integration tools

AWS Glue, Amazon Athena, Talend, SSIS, Pentaho PDI

Databases

Redshift, Teradata, Mongo, SQL Server, Postgres, MySQL, Oracle, SnowflakeDB

Operating Systems

Linux/Unix, Windows

Programming Languages

Java, J2EE, Linux Shell Scripting, HTML, CSS, JavaScript, XML

Domains

FinTech, Energy IoT, Digital Marketing, Education and Career Services, etc

Professional Training, Achievements and Certifications

Certified Talend Data Integration Developer.

Certified Pentaho BI Platform Consultant

Roles & Responsibility

Analyze business needs through discussions with customer, including business users, in order to understand current and future needs, both stated as well as unstated.

Design BI solution keeping application’s performance in mind.

Create appropriate Specification and Design documents, and conduct walk-throughs with the customer in order to bring clarity of specifications, business rules and finalize performance parameters.

Evaluate latest technologies with a view to bring scalability and adaptability of the final solution, and proposing the same to the customer.

Interaction with multiple third party API providers to enable integrations

Perform Self-Review of artifacts and superior review of team’s work products.

Education Details

Bachelors of Engineering in Electronics & Communication from Andhra Pradesh, India.

Work Experience

Sigma Infosolutions Inc., Irvine, CA, USA – From May 2017 to PRESENT

Sigma Infosolutions Ltd, Bangalore, India – From Aug 2015 to May 2017

TekSystems Global IT Services, Bangalore, India – From Nov 2014 to July 2015

Sigma Infosolutions Ltd, Bangalore, India – From June 2014 to Nov 2014

Saviance Technologies, Bangalore, India – From Feb 2014 to June 2014

MindWave Solutions, Technology Park, Malaysia – From Sept 2013 to Jan 2014

Sigma Infosolutions Ltd, Bangalore, India – From Oct 2010 to Sept 2013

Project Details

PROJECT: Loan Rates Comparative Analytics

Environment: AWS Glue, Athena, EMR, Hadoop Spark, Python, Pyspark, AWS S3

Domain: Finance

Client: Leading US based Personal Loan & Credit card Finance Aggregator

Description:

This client is mainly in to comparing rates and fees on loans and credit cards. We work on pulling the data from different lenders and have a single data mart for reports/dashboards.

Role:

Design Glue ETL scripts to transform the data to target S3 folders.

Athena tables to read the json data from s3.

Aws EMR cluster to process the huge volumes of data.

Design Spark scripts for data processing.

Pentaho ETL code to pull the data from other sources like Email, API, google analytics etc.

Data pipeline to schedule the EMR cluster to start and stop when needed.

PROJECT: Education and Discipline Analytics

Environment: Pentaho BI server 5.4, Pentaho CDE,CDF,CDA,CGG, Pentaho Report designer, Pentaho Data Integration, Mysql database

Domain: Education

Client: US based Education Specialist

Description:

HERO Education delivers an automated data dashboard and early warning system for school districts. These systems allow school personnel in easily extract and analyze district and school-wide discipline data necessary for identifying Universal (Tier 1) behavioral trends, and individual students who may require formal, tiered behavioral supports. These automated tools may be used by school administrators, Response to Intervention (RtI) Teams, Positive Behavioral Interventions and Supports (PBIS) Teams and other school-based teams (e.g., Teacher Support Teams, Student Assistance Teams) and other school personnel.

Role:

Design Pentaho CDE dashboards with filters, drilldowns and exports.

Implement ETL scripts to load the data from SIS to the MySQL database.

Automate ETL scripts to process the data everyday on scheduled time.

Design the jsp pages to use Pentaho Rest APIs for updating the user information.

Design security system in order to filter the reports data based on the role.

Made all the reports responsive, where the charts/tables resize according to the window size.

Added functionality that allows us to download the entire CDE dashboard in to pdf format.

Designed the UI pages where users can upload the data docs and ETLs will pick up them and load to the database.

Used the Pentaho APIs to add/edit users and their roles. Also added reset password functionality.

PROJECT: Loan Origination System

Environment: Tableau Desktop 9.0, Tableau Server 9.0, MS Sql server database

Domain: Fintech

Client: US based Financial services provider

Description:

The client is in to personal loans and mortgage loans. Sigma designed LOS (Loan Origination System) and our goal is to design the reports and dashboards based on the loan conversion rates for each stage (i.e. application submitted to fund). We did integrate Kafka and Storm with LOS. LOS sends data to Kafka queue through the event triggering. Storm pulls the data from Kafka and applies manipulations that are required. Finally we process this data to database. Also we are involved in designing the reports for operations & risk team.

Role:

Design sql scripts as per the requirement and use them in the tableau reports.

Design tableau workbooks with the sheets and dashboards and publish them to the tableau server.

Schedule the data refresh based on the daily and monthly reports.

Normalized the data by designing the sql scripts.

Designed the dynamic reports with the date between PST and UTC including the daylight saving.

Automate the reports schedule to mail the pdf format every day.

PROJECT: Lincoln Financial Group (LFG)

Environment: Hadoop, Pig, Spark, Sqoop, AWS EMR, EC2, Data pipeline, S3 etc., Talend ETL.

Domain: Finance

Client: Financial Service provider

Description:

In this project, we handle data from multiple source systems and process to database. As the data is very huge, we use AWS EMR clusters to handle the data. We use Talend jobs to process data via Hadoop EMR clusters.

Role:

Analyze different source systems and the data model.

Design pig scripts to generate the conformed files from the raw data.

Using Sqoop to process data to MySQL DB.

Talend jobs to use the conformed files and upload to the database.

Design data pipeline via AWS to trigger the pig scripts and Talend jobs.

Unit testing the data.

Designed the ant scripts that automate the entire process starting from copying the ETL scripts and configuration files to the remote server and running them automatically.

PROJECT: Data migration for a US Capital Loan provider

Environment: Pentaho Data Integration, MS-SQL DB

Domain: Finance

Client: US Based Capital Loan Provider

Description:

The client is one of the fastest growing Short-term Capital loan providers. We have built a Loan Origination System (LOS) and Loan Servicing System (LSS), and all the historical data migrated into these new systems including the financial data in QuickBooks, Credit history of the applicants, etc. The migration was done with minimal downtime and Pentaho ETL tool. There was lot of calculations to be done in order to get the data from historical dataset and placed into the target dataset being used by the new system..

Role:

Analyzing the source and target DB and taking calls with client regarding the business requirements.

Designing the ETL scripts for migrate the data from source to target.

Testing the target DB data in the application.

Performing data reconciliation by maintaining everything in the sheet.

Implemented the parallel processing of ETLs with multiple instances, which help in processing the data much faster than the usual load. (via the ant scripts). And capturing and reprocessing the failed data.

PROJECT: Customer care and billing

Environment: Talend open studio 5.4, MS SQL, DB2 and Oracle DB

Domain: Customer care

Client: IT Services Provider

Description:

This project is migrated the old CIS application to the Oracle CC&B application and migrating the data of all the customers using two stages i.e., Staging and production, and validating the data.

Role:

Designing the Talend jobs for each entity to extract data from MySQL server and DB2 database to the Oracle DB.

Implemented the reusable jobs so that no need to change the configurations in all the jobs if any changes required.

Implemented the SVN repo for the ETL scripts.

Scheduled the jobs in the Linux using crontab.

Implemented the error handling and context variables in all the jobs.

Implemented the error handling which triggers a mail with all the failed records.

PROJECT: Mindmart

Environment: Pentaho kettle, Pentaho CDE, CDF, Raphael, Google maps API

Domain: Analytics

Client: Mindmart

Team Size:1

Description:

This is a location based analytics where my job is to design the custom dashboards using the CDE, Raphael JS and google maps API etc.,

Role:

Designing the CDE dashboards by understanding the project requirements.

Deploying the dashboards to the production server.

Creating the user/roles and giving access to the reports according to the roles.

Designing the interactive google maps reports using the Raphael JS and google maps API.

Designed interactive google maps which show the number of fire hydrants in the covered location (with circle). Also included a slide bar using which we can increase the covered location and dynamically it displays the fire hydrants.

PROJECT: Numerify

Environment: Talend, MySQL DB.

Domain: Analytics

Client: US based Cloud analytics provider firm

Description:

In this project we mainly focused on the extracting and transforming the data from staging to the production DB using the error handling and updating the client through mails if there is any error in the ETL script running and finally deploying the ETL jobs in the integration server using the SVN and scheduling it.

Role:

Designing the ETL jobs using Talend 5.2.1.

Processing data from staging DB to the S3 bucket to speed up the process as the client is using the redshift DB as the target DB.

Designing jobs to load data from S3 bucket to the redshift DB.

Testing the data comparing the source and target tables etc.

Implemented error handling in such a way that whenever there is an error it aborts the whole job and in the next run it starts from the stage where the error has occurred.

PROJECT: Digital Marketing Analytics

Environment: Talend, Tableau Desktop and server

Domain: Business Intelligence

Client: Digital Marketing Consultancy

Description:

The client is a data focused Digital Marketing Consultancy specialized in Business Intelligence, Data warehousing and Predictive Analytics of Digital Marketing Data. We work together with client in delivering the digital data integration services. We design the Talend jobs to extract and transform the data from multiple data sources and load the data to Mongo DB/ Hadoop. The sources include the json files, google analytics etc.

Role:

Designing the ETL jobs using Talend 5.2.1.

Loading the transformed data to MongoDB on daily basis.

Scheduling the Talend jobs using windows task scheduler.

Designing the Error handling Talend jobs so that they will send a mail to the user whenever the error occurs in the transforming the data.

PROJECT: Flight Operation Analytics

Environment: Pentaho Kettle, Pentaho BI server, Pentaho Report designer, Lucid DB

Domain: Business Intelligence

Client: US Based Flight Operations Service Provider

Description:

Flight Office is a real-time, end-to-end solution to easily manage flight operations, giving you a complete overview of your business in real time. We have developed the report to provide end-to-end solution to business analyst. Our goal is to design the good looking reports using Pentaho report designer and publishing them to the Pentaho BI server. For this project we used the Lucid DB as well as Couch DB for fetching the flight data. .

Role:

Designing the reports and publishing them to the pentaho BI server.

Verifying the reports once publishing them to the server.

Implemented the internationalization in all the reports, but there was a bug in Pentaho which doesn’t allow to proceed with localization. I fixed it by modifying the java code in one of the jar file.

PROJECT: BI framework setup

Environment: Pentaho Kettle 4.1.0, Pentaho BI server 3.7.0, Pentaho Report designer 3.7.0, CDE (community dashboard editor), MySQL 5.0,UNIX.

Role: Developer

Domain: Energy Storage

Client: US based Energy Storage firm

Description:

The client is a leading Energy Storage firm in US and it operates a diverse and growing portfolio of generation and distribution businesses with the capacity to serve 100 million people worldwide. The proposed BI platform helps them to monitor the power grid stations and regulate them appropriately as and when required. The solution is completely built on Pentaho stack with lot of customizations to suit their needs. Standardized and timely information is provided to the business users, which will help them to make strategic decisions regarding to the business

Responsibilities:

Performance tuning of ETL scripts and DB stored procedures.

Data migration from one environment to another environment using kettle ETL tool.

Created ETL mappings using kettle 4.1.0 to move data from multiple sources like ISO links, URL's, csv files etc.

Query Optimization.

PROJECT: Fuel Quest

Environment: Pentaho BI server, Report designer, Design studio, Oracle 10g, Toad for Oracle.

Domain: Retail management

Client: Fuel Quest

Description:

This project is to provide different reports and supporting business queries for making intelligent invoice decision based on data available in various branches across the country collected over a period of time. Fuel Quest wants to have a tool, which will help it to analyze the business trends and to make predictions for future business trends.

Responsibilities:

Dashboard Design using open flash charts with good look and feel.

Flash charts integration with the reports that display the data in the form of tables below the flash charts.

Parameters passing to the flash charts and reports.

Testing & Bug Fixing.



Contact this candidate