Post Job Free

Resume

Sign in

Business Intelligence Lead

Location:
Santa Clarita, CA, 91351
Posted:
September 12, 2017

Contact this candidate

Resume:

Sergey Zavgorodni

***** *. ****** ***, ***** Clarita, CA, 91387 : 310-***-**** : ac19wr@r.postjobfree.com

https://www.linkedin.com/in/szavgorodni/

Accomplished Architect / Business Intelligence Lead with over 15 years of the deep industry experience in Big Data Business Intelligence, leveraging variety cutting edge technologies including Hadoop, Hive, Spark, Elastic Search, Vertica, Oracle, MSSQL, SAP Business Objects Data Services and WEBI, Tableau

Areas of expertise include data warehouse architecture, process engineering, data pipelines development, IT architecture, ETL, data modeling, data analysis, QA, and project management.

Qualification Summary

Lead and contributed to implementation of first enterprise data warehouse solution using MS SQL platform, which was adopted by entire company

Lead efforts to define the architecture, strategy and development standards within ETL / BI team

Lead and contributed to complete, petabyte scale, data warehouse refactoring using big data solutions (HP Vertica, Hadoop, Hive, Hbase, Elastic Search, MS SQL )

Key contributor, architect and technical lead of implementation high performance “in house” ETL tool using Python. This solution currently used for all ETL projects in BI department.

Supported implementation of variety end-user reporting applications with audience exceeding 5000 users, using Business Objects InfoView, Tableau and custom .Net and NodeJS applications

Currently leading refactoring of the enterprise data warehouse ETL process to adapt to migration from Hadoop map reduce to Kafka and Spark 2.0 data processing

Worked with other engineering teams within the company, providing the requirements for implementation of the data sources that can be effectively used by data warehouse process

managed 24x7 enterprise data warehouse operations, including monitoring, alerts, addressing priority zero productions issues

Introduced the Agile development methodology to BI team, which produced a significant increase in quality of all products and a more predictable product delivery.

Organized several HP Vertica, Hadoop Hive / HBase trainings for BI team lead by Cloudera and HP Vertica specialists

Established record of accomplishment of successfully managing multiple priorities, people and projects in a fast-paced and fluid environment.

Adaptive, resourceful, energetic leadership style that proven to encourage and foster innovation, productivity and quality.

My current position is highly technical, demanding and customer oriented, requiring solutions to be delivered to very tight deadlines. I have an open and inquiring mind and I am always looking for a professional challenge. Can you offer me that challenge?

Professional Objective:

To obtain a challenging position as a Architect / Business Intelligence team Lead in Data Warehouse / BI solution development utilizing my data base and analytical expertise, where I can contribute to the mission of organization.

Technical skills

Languages:

SQL (15+ years), Python (4+ years), Unix Shell Scripting (5+years), Scala(1+ year), C#, .NET (3+ years), HTML/DHTML (4 years), XML/XSLT (3 years)

Technologies:

SQL analytical functions, Apache Hadoop, Spark 2.0, HDFS, WebHDFS, Hive, Pig, HBase, Sqoop, HCataog, Solr, Elastic Search, Vertica User Defined functions, GPG/AES encrypting, SOAP/REST API Services, TDD, Unit Testing, AJAX

Databases:

HP Vertica, MS SQL, Oracle, MySQL, T-SQL, stored procedures, triggers, replication

Report Tools

Tableau, Business Objects Infoview

Developer Tools:

IntelliJ, PyCharm, Aqua Data Studio, Tableau, SAP WEBI / DS, Vagrant, Docker, Git/Stash, JFrog Artifactory, SaltStack, PyPi, Maven, SBT, Visio, Erwin, JIRA, Wiki, Salesforce.com, XMLSpy

Operating Systems:

Unix Centos, MS Windows

Graphic Tools

Adobe Photoshop, Adobe Illustrator

Work Status

US Citizen

Work Status

Full-Time, Permanent

Education:

Vertica Courses:

HP Vertica Analytics

HP Vertica Advanced Performance Tuning

Hadoop Courses:

2015

2015

Cloudera Data Analyst Training

Cloudera Training for Apache HBase

2014

2014

Agile Training Courses

2013

Microsoft Software Developer courses.

MCP (Microsoft Certified Professional) degree implementing a Database Design on Microsoft SQL Server 7.0

MCSD (Microsoft Certified Solution Developer) degree in Designing and Implementing Desktop and Distributed Applications with Microsoft Visual Basic 6.0. Analyzing Requirements and Defining Solution Architectures

1999

Certified Visual Basic developer, by Brain Bench (Tekmetrics)

1999

St. Petersburg State University, St. Petersburg (Leningrad), Russia

Master of Science in Biology

1982 - 1988

Affiliations:

Member of Los Angeles Area .NET Developers Group (LADOTNET)

Participant of TDWI and HP Vertica conferences, seminars, etc.

Author of several presentations and lectures on practical aspects of using cutting-edge programming methods

Professional Summary

Yellowpages.com

Architect / Business Intelligence Lead with May 2006 – Present

Lead and contributed to the development of variety of BI / data warehouse and reporting projects using full stack of technologies including HP Vertica, MSSQL, Oracle, Hadoop, Spark 2.0, Hive, SAP DS/WEBI and Tableau

BI team supported numerous analytical teams and reporting for thousands of users.

Established the processes, tools and best development practices for rapid and high quality projects delivery.

This includes:

Introduced standards for technical documentation, including requirements, workflow diagrams, code support, etc.

Agile, test driven project development process using Atlassian Jira and Stash/git

Custom highly scalable and performant python ETL application

JFrog Artifactory for python code packaging

Standard deployment process using Salt Stack application

Job scheduling using cron, Autosys or MS SQL Agent

Servers’ hardware monitoring using Munin Monitoring

Application status management and monitoring using custom Oracle/php based application

Alerts and notification framework using custom Python/unix shell application

Some of the major projects listed below

Projects:

Python ELT framework is a lightweight workflow management platform and set of the components necessary for implementations of typical ETL projects.

It designed with the following goals:

facilitate the rapid project development, deployment and maintenance

provide the platform that employs the most efficient extract, load and transformation techniques suited to work with Hadoop environment

ensure scalability and reusability

introduce standard design patterns to the data warehouse ETL projects

reduce the level of effort for implementing the typical routine tasks

ensure straightforward extensibility of the framework functionality

serve as guidance for the best development practices

Enterprise Data Warehouse (EDW)

Priority #1 project in the company aimed to address the scalability and performance issues of legacy Data warehouse solution

The based data collection implemented in Hadoop with petabyte data set stored on HDFS

The effective Vertica data import process runs in 15 minutes batches collecting and processing the base data.

Data stored in atomic and aggregated form. This allows users to drill down from high level report to the very granular details of traffic and billing data

The data warehouse solution collects and process over 1 billion events daily. This included Web/API traffic, Leads, Sales Activity, Billing, Listing and other business critical data sets

This solution required deep optimization of the Vertica cluster performance and leveraged the most effective highly parallel data transfer techniques

The optimization also required trough analysis and implementation of the best projection design practices that allowed majority of reports and analytical queries to be executed in sub second time period

EDW solution exposed via, Business Object Infoview, Tableau and custom web applications. It also allows power users to run the analysis using Hadoop map reduce application and Hive queries

Marketing Analytical Datamart.

The Marketing data mart provides a single repository for marketing team to run all marketing campaigns and track responses to the campaigns. The data mart sourced from both internal and external sources of data and serves as the only source of marketing lists. The data mart also keeps track of all the touch points in the marketing and sales funnel and help provide a view of the relative value of the leads generated through the various campaigns.

In addition, the data mart allows assignment of various scores to records that measure the propensity to buy or the value of the prospect. This allows for proper routing and handling of the leads generated through the various marketing campaigns.

Technical design:

Marketing data mart is consolidated and enriched data set sourced form YP Enterprise data warehouse, Sales Force application, and Dun and Bradstreet (D&B) business information.

D&B data imported to Vertica using highly efficient parallel ftp data transfer

The Vertica data sent to and received from Apache Solr cluster to perform the match between YP EDW data and D&B data using WEB Hdfs protocol.

Enriched data sent and received from SAS to calculated the propensity to buy metrics

The coordination of the process batches performed by custom pipeline status application

Users generate marketing campaigns data set, which sent by Marketing data mart to external YP application via custom API for further actions

Sales KPI reporting

Set of data marts, reports and applications that provide integrated view of sales activities tied to booking and revenue data

This project required the integration of over 30 various FTP, API, Hadoop, Oracle and MS SQL database sources involving implementation of non-trivial business logic.

This solution-required the implementation of robust coordination and signaling between involved processes and strict Service Level Agreement related to timing and accuracy of the delivered data

The reports exposed via multiple Tableau and Custom NodeJS web based applications developed by another team in our department.

The audience of the Sales KPI reporting solution exceeded 5000 sales representatives and business analysts

Contributed to creation of analytic platform on Hadoop, using Spark 2.0, Databricks spark notebooks and AWS platform

Maintained and supported multiple financial reports created in MS SQL SSRS, using MS SQL SSIS

on MS SQL Server 2014 platform

Cerner Corporation (http://www.cerner.com)

Software Architect, Manager, September 2001– May 2006,

Managed IT projects of large healthcare site contributing to organization’s revenue growth: $2.4 million-3.2 million in one year

Job responsibilities include managing a group of 5 programmers, developing applications using a cluster of Internet servers, and Microsoft SQL Server 2000 servers. Architected, designed, implemented, documented, managed, supported multiple complex large-scale web based solutions such as the user account management, etc. Built and maintained multi-level test/development environment.

List of managed projects:

“ Evidence Based Decision Support Center (EBDSC)“: a large-scale web solution, combining web-based (front-end), standalone (back-end) data-entry applications using MS SQL Server as database layer. The solution has about 10,000 registered users. The application provides sophisticated and flexible search and data meta-analysis tools, a variety of highly customizable reports, stores customer preferences, presents results of analysis as real time charts, reports results in dynamically created Excel, PDF files, etc. Performs an analysis of healthcare knowledge data, organization management, evaluates the impact on organization budget of switching from one strategy to another - both in terms of cost and effectiveness.

Languages and Technologies used: C#, ASP.NET, ADO.NET, SQL (T-SQL, stored procedures, triggers, DTS, etc.), Remoting, COM, CSS, DHTML, DOM, XML, JavaScript, VB, VBA (MS Word, MS Excel, MS Access), MSXML parser, custom and third-party components. OO design and project lifecycle management, using RUP, UML, Agile, TDD, NUnit, Reafactoring, etc.

“Prometheus”: Performs an analysis of healthcare organization management, evaluates the impact on organization budget of switching from one strategy to another - both in terms of cost and effectiveness. The model demonstrates cost savings or additional (incremental) costs and any incremental effectiveness as a result of the switch. Technologies used: VB, Custom and third-party ActiveX controls, Memory Management, References, WIN32 API, Callback procedures.

“Gastro esophageal Reflux Disease (GERD) Screener application”: This application read, store in database, analyze and output primary diagnosis, based on data from scanned paper forms with patient answers. VB application. Technologies used: Serial Port Data Acquisition, ADO, custom ActiveX controls, Data-aware classes, WIN32 API, Callback procedures

Zynx Health Incorporated (http://www.zynx.com/)

Senior Analyst Programmer, IT Team Technical Lead, May 2000 – September 2001, ($80K)

Lead design, development and support of the most projects.

“Pain Tracker”: Multi-tiered, interactive Web Application. Technologies used: JavaScript, HTML/DHTML, XML, ASP, ADO, custom and third-party ActiveX components.

“Pain Manager”: Interactive Web-based tool designed for use by the multidisciplinary clinical team, quality management specialists, and administrators. This tool was developed to help hospitals and health care organizations adhere to HealthCare standards. Statistical treatment of data entered by user. Real-time graphical display of performance measure. Technologies used: JavaScript, HTML/DHTML, XML, ASP, ADO, COM.

“Safety Optimizer”: Interactive web application. Error-reduction tool featuring guidelines for improving safety within the acute care setting and for tracking progress in achieving improvements. The program is designed for use by the multidisciplinary clinical team, as well as for experts in quality-improvement and risk-management. Real-time graphical display of performance measure. Technologies used: JavaScript, HTML/DHTML, XML, ASP, ADO, COM. Lead design, development and support.

“Product Manager”: The back-end web application, that allow manage, create, search and maintain customer accounts. The created design and conception was used by another team members for extend functionality of application. Technologies used: JavaScript, HTML/DHTML, ASP, ADO, COM. Lead design, development and support.



Contact this candidate