HARSHAVARDHAN REDDY K
BI Developer
*********@*****.***
I am based out of Irvine, CA and I have 7+ years of experience in the area of Data Warehousing and Analytics using Tableau, Pentaho, Jaspersoft, and Talend. I’ve also extensively worked with Cloud-based platforms like AWS Glue, Athena, EMR, EC2 Quicksight, Redshift and S3. Very much hands on with the Big Data technologies like Hadoop, Map Reduce, Pig, Sqoop, Hive, Spark.
Professional Summary
Well versed in all phases of data warehouse development lifecycle, from gathering requirements to testing, implementation and support. Key strengths in using best Data Warehousing practices and Data Re-engineering with proven ability in the design, analysis and management of Enterprise scale BI/DW systems.
Proven ability in planning and implementing the overall analytics and business intelligence strategy, also ability to lead the design and development of BI projects designed to understand key business behaviors that drive customer acquisition, retention and engagement.
Rich working experience in Data Visualization with tools like Tableau, Tibco Spotfire and Amazon QuickSight. Have worked on integrating 3rd party charting libraries to devise custom visualizations as needed.
Being a Certified Pentaho BI Developer I’ve implementated a range of BI solutions right from Stand-alone reporting to Cloud based Multi-tenant solutions.
Also as a Certified Talend Data Integration Engineer I’ve vast experience in Data Integration and have sourced data from various sources including DBs, Flat files, 3rd party APIs, Python scripts, cloud-based data pipelines, Scrapper programs, etc.
Performing systems analysis & suggesting alternatives to improve and enhance the services & delivery standards.
Taken up initiatives in Performance benchmarking and improvements.
Consummate professional with excellent communication, planning, execution, monitoring and problem solving skills.
Strong Exposure on Lending and Finance domain, Energy IoT Analytics, Digital Marketing and Education and Career Services
Agile enthusiast and Agile Enthusiast and active participant in all Agile ceremonies
Strong interpersonal and communication skills with an ability to lead a team, collaboration with off-shore teams and coordination of release activities with DevOps team.
Technical Skills:
Big Data Stack
Map Reduce, Pig, Sqoop, Hive, Spark, AWS EMR, AWS S3, AWS Data pipeline etc.
BI Platforms
Tableau server, Tibco spotfire, Pentaho BI server, Jasper server, Mondrian OLAP server
Visualization and Reporting
Tableau, AWS Quicksight, Spotfire, Pentaho Dashboarding and CDE, Saiku Analytics, Jasper, BIRT, Fusion charts, Highcharts, D3 charts.
Data Integration tools
AWS Glue, Amazon Athena, Talend, SSIS, Pentaho PDI
Databases
Redshift, Teradata, Mongo, SQL Server, Postgres, MySQL, Oracle, SnowflakeDB
Operating Systems
Linux/Unix, Windows
Programming Languages
Java, J2EE, Linux Shell Scripting, HTML, CSS, JavaScript, XML
Domains
FinTech, Energy IoT, Digital Marketing, Education and Career Services, etc
Professional Training, Achievements and Certifications
Certified Talend Data Integration Developer.
Certified Pentaho BI Platform Consultant
Roles & Responsibility
Analyze business needs through discussions with customer, including business users, in order to understand current and future needs, both stated as well as unstated.
Design BI solution keeping application’s performance in mind.
Create appropriate Specification and Design documents, and conduct walk-throughs with the customer in order to bring clarity of specifications, business rules and finalize performance parameters.
Evaluate latest technologies with a view to bring scalability and adaptability of the final solution, and proposing the same to the customer.
Interaction with multiple third party API providers to enable integrations
Perform Self-Review of artifacts and superior review of team’s work products.
Education Details
Bachelors of Engineering in Electronics & Communication from Andhra Pradesh, India.
Work Experience
Sigma Infosolutions Inc., Irvine, CA, USA – From May 2017 to PRESENT
Sigma Infosolutions Ltd, Bangalore, India – From Aug 2015 to May 2017
TekSystems Global IT Services, Bangalore, India – From Nov 2014 to July 2015
Sigma Infosolutions Ltd, Bangalore, India – From June 2014 to Nov 2014
Saviance Technologies, Bangalore, India – From Feb 2014 to June 2014
MindWave Solutions, Technology Park, Malaysia – From Sept 2013 to Jan 2014
Sigma Infosolutions Ltd, Bangalore, India – From Oct 2010 to Sept 2013
Project Details
PROJECT: Loan Rates Comparative Analytics
Environment: AWS Glue, Athena, EMR, Hadoop Spark, Python, Pyspark, AWS S3
Domain: Finance
Client: Leading US based Personal Loan & Credit card Finance Aggregator
Description:
This client is mainly in to comparing rates and fees on loans and credit cards. We work on pulling the data from different lenders and have a single data mart for reports/dashboards.
Role:
Design Glue ETL scripts to transform the data to target S3 folders.
Athena tables to read the json data from s3.
Aws EMR cluster to process the huge volumes of data.
Design Spark scripts for data processing.
Pentaho ETL code to pull the data from other sources like Email, API, google analytics etc.
Data pipeline to schedule the EMR cluster to start and stop when needed.
PROJECT: Education and Discipline Analytics
Environment: Pentaho BI server 5.4, Pentaho CDE,CDF,CDA,CGG, Pentaho Report designer, Pentaho Data Integration, Mysql database
Domain: Education
Client: US based Education Specialist
Description:
HERO Education delivers an automated data dashboard and early warning system for school districts. These systems allow school personnel in easily extract and analyze district and school-wide discipline data necessary for identifying Universal (Tier 1) behavioral trends, and individual students who may require formal, tiered behavioral supports. These automated tools may be used by school administrators, Response to Intervention (RtI) Teams, Positive Behavioral Interventions and Supports (PBIS) Teams and other school-based teams (e.g., Teacher Support Teams, Student Assistance Teams) and other school personnel.
Role:
Design Pentaho CDE dashboards with filters, drilldowns and exports.
Implement ETL scripts to load the data from SIS to the MySQL database.
Automate ETL scripts to process the data everyday on scheduled time.
Design the jsp pages to use Pentaho Rest APIs for updating the user information.
Design security system in order to filter the reports data based on the role.
Made all the reports responsive, where the charts/tables resize according to the window size.
Added functionality that allows us to download the entire CDE dashboard in to pdf format.
Designed the UI pages where users can upload the data docs and ETLs will pick up them and load to the database.
Used the Pentaho APIs to add/edit users and their roles. Also added reset password functionality.
PROJECT: Loan Origination System
Environment: Tableau Desktop 9.0, Tableau Server 9.0, MS Sql server database
Domain: Fintech
Client: US based Financial services provider
Description:
The client is in to personal loans and mortgage loans. Sigma designed LOS (Loan Origination System) and our goal is to design the reports and dashboards based on the loan conversion rates for each stage (i.e. application submitted to fund). We did integrate Kafka and Storm with LOS. LOS sends data to Kafka queue through the event triggering. Storm pulls the data from Kafka and applies manipulations that are required. Finally we process this data to database. Also we are involved in designing the reports for operations & risk team.
Role:
Design sql scripts as per the requirement and use them in the tableau reports.
Design tableau workbooks with the sheets and dashboards and publish them to the tableau server.
Schedule the data refresh based on the daily and monthly reports.
Normalized the data by designing the sql scripts.
Designed the dynamic reports with the date between PST and UTC including the daylight saving.
Automate the reports schedule to mail the pdf format every day.
PROJECT: Lincoln Financial Group (LFG)
Environment: Hadoop, Pig, Spark, Sqoop, AWS EMR, EC2, Data pipeline, S3 etc., Talend ETL.
Domain: Finance
Client: Financial Service provider
Description:
In this project, we handle data from multiple source systems and process to database. As the data is very huge, we use AWS EMR clusters to handle the data. We use Talend jobs to process data via Hadoop EMR clusters.
Role:
Analyze different source systems and the data model.
Design pig scripts to generate the conformed files from the raw data.
Using Sqoop to process data to MySQL DB.
Talend jobs to use the conformed files and upload to the database.
Design data pipeline via AWS to trigger the pig scripts and Talend jobs.
Unit testing the data.
Designed the ant scripts that automate the entire process starting from copying the ETL scripts and configuration files to the remote server and running them automatically.
PROJECT: Data migration for a US Capital Loan provider
Environment: Pentaho Data Integration, MS-SQL DB
Domain: Finance
Client: US Based Capital Loan Provider
Description:
The client is one of the fastest growing Short-term Capital loan providers. We have built a Loan Origination System (LOS) and Loan Servicing System (LSS), and all the historical data migrated into these new systems including the financial data in QuickBooks, Credit history of the applicants, etc. The migration was done with minimal downtime and Pentaho ETL tool. There was lot of calculations to be done in order to get the data from historical dataset and placed into the target dataset being used by the new system..
Role:
Analyzing the source and target DB and taking calls with client regarding the business requirements.
Designing the ETL scripts for migrate the data from source to target.
Testing the target DB data in the application.
Performing data reconciliation by maintaining everything in the sheet.
Implemented the parallel processing of ETLs with multiple instances, which help in processing the data much faster than the usual load. (via the ant scripts). And capturing and reprocessing the failed data.
PROJECT: Customer care and billing
Environment: Talend open studio 5.4, MS SQL, DB2 and Oracle DB
Domain: Customer care
Client: IT Services Provider
Description:
This project is migrated the old CIS application to the Oracle CC&B application and migrating the data of all the customers using two stages i.e., Staging and production, and validating the data.
Role:
Designing the Talend jobs for each entity to extract data from MySQL server and DB2 database to the Oracle DB.
Implemented the reusable jobs so that no need to change the configurations in all the jobs if any changes required.
Implemented the SVN repo for the ETL scripts.
Scheduled the jobs in the Linux using crontab.
Implemented the error handling and context variables in all the jobs.
Implemented the error handling which triggers a mail with all the failed records.
PROJECT: Mindmart
Environment: Pentaho kettle, Pentaho CDE, CDF, Raphael, Google maps API
Domain: Analytics
Client: Mindmart
Team Size:1
Description:
This is a location based analytics where my job is to design the custom dashboards using the CDE, Raphael JS and google maps API etc.,
Role:
Designing the CDE dashboards by understanding the project requirements.
Deploying the dashboards to the production server.
Creating the user/roles and giving access to the reports according to the roles.
Designing the interactive google maps reports using the Raphael JS and google maps API.
Designed interactive google maps which show the number of fire hydrants in the covered location (with circle). Also included a slide bar using which we can increase the covered location and dynamically it displays the fire hydrants.
PROJECT: Numerify
Environment: Talend, MySQL DB.
Domain: Analytics
Client: US based Cloud analytics provider firm
Description:
In this project we mainly focused on the extracting and transforming the data from staging to the production DB using the error handling and updating the client through mails if there is any error in the ETL script running and finally deploying the ETL jobs in the integration server using the SVN and scheduling it.
Role:
Designing the ETL jobs using Talend 5.2.1.
Processing data from staging DB to the S3 bucket to speed up the process as the client is using the redshift DB as the target DB.
Designing jobs to load data from S3 bucket to the redshift DB.
Testing the data comparing the source and target tables etc.
Implemented error handling in such a way that whenever there is an error it aborts the whole job and in the next run it starts from the stage where the error has occurred.
PROJECT: Digital Marketing Analytics
Environment: Talend, Tableau Desktop and server
Domain: Business Intelligence
Client: Digital Marketing Consultancy
Description:
The client is a data focused Digital Marketing Consultancy specialized in Business Intelligence, Data warehousing and Predictive Analytics of Digital Marketing Data. We work together with client in delivering the digital data integration services. We design the Talend jobs to extract and transform the data from multiple data sources and load the data to Mongo DB/ Hadoop. The sources include the json files, google analytics etc.
Role:
Designing the ETL jobs using Talend 5.2.1.
Loading the transformed data to MongoDB on daily basis.
Scheduling the Talend jobs using windows task scheduler.
Designing the Error handling Talend jobs so that they will send a mail to the user whenever the error occurs in the transforming the data.
PROJECT: Flight Operation Analytics
Environment: Pentaho Kettle, Pentaho BI server, Pentaho Report designer, Lucid DB
Domain: Business Intelligence
Client: US Based Flight Operations Service Provider
Description:
Flight Office is a real-time, end-to-end solution to easily manage flight operations, giving you a complete overview of your business in real time. We have developed the report to provide end-to-end solution to business analyst. Our goal is to design the good looking reports using Pentaho report designer and publishing them to the Pentaho BI server. For this project we used the Lucid DB as well as Couch DB for fetching the flight data. .
Role:
Designing the reports and publishing them to the pentaho BI server.
Verifying the reports once publishing them to the server.
Implemented the internationalization in all the reports, but there was a bug in Pentaho which doesn’t allow to proceed with localization. I fixed it by modifying the java code in one of the jar file.
PROJECT: BI framework setup
Environment: Pentaho Kettle 4.1.0, Pentaho BI server 3.7.0, Pentaho Report designer 3.7.0, CDE (community dashboard editor), MySQL 5.0,UNIX.
Role: Developer
Domain: Energy Storage
Client: US based Energy Storage firm
Description:
The client is a leading Energy Storage firm in US and it operates a diverse and growing portfolio of generation and distribution businesses with the capacity to serve 100 million people worldwide. The proposed BI platform helps them to monitor the power grid stations and regulate them appropriately as and when required. The solution is completely built on Pentaho stack with lot of customizations to suit their needs. Standardized and timely information is provided to the business users, which will help them to make strategic decisions regarding to the business
Responsibilities:
Performance tuning of ETL scripts and DB stored procedures.
Data migration from one environment to another environment using kettle ETL tool.
Created ETL mappings using kettle 4.1.0 to move data from multiple sources like ISO links, URL's, csv files etc.
Query Optimization.
PROJECT: Fuel Quest
Environment: Pentaho BI server, Report designer, Design studio, Oracle 10g, Toad for Oracle.
Domain: Retail management
Client: Fuel Quest
Description:
This project is to provide different reports and supporting business queries for making intelligent invoice decision based on data available in various branches across the country collected over a period of time. Fuel Quest wants to have a tool, which will help it to analyze the business trends and to make predictions for future business trends.
Responsibilities:
Dashboard Design using open flash charts with good look and feel.
Flash charts integration with the reports that display the data in the form of tables below the flash charts.
Parameters passing to the flash charts and reports.
Testing & Bug Fixing.