Sergey Zavgorodni
***** *. ****** ***, ***** Clarita, CA, 91387 : 310-***-**** : ac19wr@r.postjobfree.com
https://www.linkedin.com/in/szavgorodni/
Accomplished Architect / Business Intelligence Lead with over 15 years of the deep industry experience in Big Data Business Intelligence, leveraging variety cutting edge technologies including Hadoop, Hive, Spark, Elastic Search, Vertica, Oracle, MSSQL, SAP Business Objects Data Services and WEBI, Tableau
Areas of expertise include data warehouse architecture, process engineering, data pipelines development, IT architecture, ETL, data modeling, data analysis, QA, and project management.
Qualification Summary
Lead and contributed to implementation of first enterprise data warehouse solution using MS SQL platform, which was adopted by entire company
Lead efforts to define the architecture, strategy and development standards within ETL / BI team
Lead and contributed to complete, petabyte scale, data warehouse refactoring using big data solutions (HP Vertica, Hadoop, Hive, Hbase, Elastic Search, MS SQL )
Key contributor, architect and technical lead of implementation high performance “in house” ETL tool using Python. This solution currently used for all ETL projects in BI department.
Supported implementation of variety end-user reporting applications with audience exceeding 5000 users, using Business Objects InfoView, Tableau and custom .Net and NodeJS applications
Currently leading refactoring of the enterprise data warehouse ETL process to adapt to migration from Hadoop map reduce to Kafka and Spark 2.0 data processing
Worked with other engineering teams within the company, providing the requirements for implementation of the data sources that can be effectively used by data warehouse process
managed 24x7 enterprise data warehouse operations, including monitoring, alerts, addressing priority zero productions issues
Introduced the Agile development methodology to BI team, which produced a significant increase in quality of all products and a more predictable product delivery.
Organized several HP Vertica, Hadoop Hive / HBase trainings for BI team lead by Cloudera and HP Vertica specialists
Established record of accomplishment of successfully managing multiple priorities, people and projects in a fast-paced and fluid environment.
Adaptive, resourceful, energetic leadership style that proven to encourage and foster innovation, productivity and quality.
My current position is highly technical, demanding and customer oriented, requiring solutions to be delivered to very tight deadlines. I have an open and inquiring mind and I am always looking for a professional challenge. Can you offer me that challenge?
Professional Objective:
To obtain a challenging position as a Architect / Business Intelligence team Lead in Data Warehouse / BI solution development utilizing my data base and analytical expertise, where I can contribute to the mission of organization.
Technical skills
Languages:
SQL (15+ years), Python (4+ years), Unix Shell Scripting (5+years), Scala(1+ year), C#, .NET (3+ years), HTML/DHTML (4 years), XML/XSLT (3 years)
Technologies:
SQL analytical functions, Apache Hadoop, Spark 2.0, HDFS, WebHDFS, Hive, Pig, HBase, Sqoop, HCataog, Solr, Elastic Search, Vertica User Defined functions, GPG/AES encrypting, SOAP/REST API Services, TDD, Unit Testing, AJAX
Databases:
HP Vertica, MS SQL, Oracle, MySQL, T-SQL, stored procedures, triggers, replication
Report Tools
Tableau, Business Objects Infoview
Developer Tools:
IntelliJ, PyCharm, Aqua Data Studio, Tableau, SAP WEBI / DS, Vagrant, Docker, Git/Stash, JFrog Artifactory, SaltStack, PyPi, Maven, SBT, Visio, Erwin, JIRA, Wiki, Salesforce.com, XMLSpy
Operating Systems:
Unix Centos, MS Windows
Graphic Tools
Adobe Photoshop, Adobe Illustrator
Work Status
US Citizen
Work Status
Full-Time, Permanent
Education:
Vertica Courses:
HP Vertica Analytics
HP Vertica Advanced Performance Tuning
Hadoop Courses:
2015
2015
Cloudera Data Analyst Training
Cloudera Training for Apache HBase
2014
2014
Agile Training Courses
2013
Microsoft Software Developer courses.
MCP (Microsoft Certified Professional) degree implementing a Database Design on Microsoft SQL Server 7.0
MCSD (Microsoft Certified Solution Developer) degree in Designing and Implementing Desktop and Distributed Applications with Microsoft Visual Basic 6.0. Analyzing Requirements and Defining Solution Architectures
1999
Certified Visual Basic developer, by Brain Bench (Tekmetrics)
1999
St. Petersburg State University, St. Petersburg (Leningrad), Russia
Master of Science in Biology
1982 - 1988
Affiliations:
Member of Los Angeles Area .NET Developers Group (LADOTNET)
Participant of TDWI and HP Vertica conferences, seminars, etc.
Author of several presentations and lectures on practical aspects of using cutting-edge programming methods
Professional Summary
Yellowpages.com
Architect / Business Intelligence Lead with May 2006 – Present
Lead and contributed to the development of variety of BI / data warehouse and reporting projects using full stack of technologies including HP Vertica, MSSQL, Oracle, Hadoop, Spark 2.0, Hive, SAP DS/WEBI and Tableau
BI team supported numerous analytical teams and reporting for thousands of users.
Established the processes, tools and best development practices for rapid and high quality projects delivery.
This includes:
Introduced standards for technical documentation, including requirements, workflow diagrams, code support, etc.
Agile, test driven project development process using Atlassian Jira and Stash/git
Custom highly scalable and performant python ETL application
JFrog Artifactory for python code packaging
Standard deployment process using Salt Stack application
Job scheduling using cron, Autosys or MS SQL Agent
Servers’ hardware monitoring using Munin Monitoring
Application status management and monitoring using custom Oracle/php based application
Alerts and notification framework using custom Python/unix shell application
Some of the major projects listed below
Projects:
Python ELT framework is a lightweight workflow management platform and set of the components necessary for implementations of typical ETL projects.
It designed with the following goals:
facilitate the rapid project development, deployment and maintenance
provide the platform that employs the most efficient extract, load and transformation techniques suited to work with Hadoop environment
ensure scalability and reusability
introduce standard design patterns to the data warehouse ETL projects
reduce the level of effort for implementing the typical routine tasks
ensure straightforward extensibility of the framework functionality
serve as guidance for the best development practices
Enterprise Data Warehouse (EDW)
Priority #1 project in the company aimed to address the scalability and performance issues of legacy Data warehouse solution
The based data collection implemented in Hadoop with petabyte data set stored on HDFS
The effective Vertica data import process runs in 15 minutes batches collecting and processing the base data.
Data stored in atomic and aggregated form. This allows users to drill down from high level report to the very granular details of traffic and billing data
The data warehouse solution collects and process over 1 billion events daily. This included Web/API traffic, Leads, Sales Activity, Billing, Listing and other business critical data sets
This solution required deep optimization of the Vertica cluster performance and leveraged the most effective highly parallel data transfer techniques
The optimization also required trough analysis and implementation of the best projection design practices that allowed majority of reports and analytical queries to be executed in sub second time period
EDW solution exposed via, Business Object Infoview, Tableau and custom web applications. It also allows power users to run the analysis using Hadoop map reduce application and Hive queries
Marketing Analytical Datamart.
The Marketing data mart provides a single repository for marketing team to run all marketing campaigns and track responses to the campaigns. The data mart sourced from both internal and external sources of data and serves as the only source of marketing lists. The data mart also keeps track of all the touch points in the marketing and sales funnel and help provide a view of the relative value of the leads generated through the various campaigns.
In addition, the data mart allows assignment of various scores to records that measure the propensity to buy or the value of the prospect. This allows for proper routing and handling of the leads generated through the various marketing campaigns.
Technical design:
Marketing data mart is consolidated and enriched data set sourced form YP Enterprise data warehouse, Sales Force application, and Dun and Bradstreet (D&B) business information.
D&B data imported to Vertica using highly efficient parallel ftp data transfer
The Vertica data sent to and received from Apache Solr cluster to perform the match between YP EDW data and D&B data using WEB Hdfs protocol.
Enriched data sent and received from SAS to calculated the propensity to buy metrics
The coordination of the process batches performed by custom pipeline status application
Users generate marketing campaigns data set, which sent by Marketing data mart to external YP application via custom API for further actions
Sales KPI reporting
Set of data marts, reports and applications that provide integrated view of sales activities tied to booking and revenue data
This project required the integration of over 30 various FTP, API, Hadoop, Oracle and MS SQL database sources involving implementation of non-trivial business logic.
This solution-required the implementation of robust coordination and signaling between involved processes and strict Service Level Agreement related to timing and accuracy of the delivered data
The reports exposed via multiple Tableau and Custom NodeJS web based applications developed by another team in our department.
The audience of the Sales KPI reporting solution exceeded 5000 sales representatives and business analysts
Contributed to creation of analytic platform on Hadoop, using Spark 2.0, Databricks spark notebooks and AWS platform
Maintained and supported multiple financial reports created in MS SQL SSRS, using MS SQL SSIS
on MS SQL Server 2014 platform
Cerner Corporation (http://www.cerner.com)
Software Architect, Manager, September 2001– May 2006,
Managed IT projects of large healthcare site contributing to organization’s revenue growth: $2.4 million-3.2 million in one year
Job responsibilities include managing a group of 5 programmers, developing applications using a cluster of Internet servers, and Microsoft SQL Server 2000 servers. Architected, designed, implemented, documented, managed, supported multiple complex large-scale web based solutions such as the user account management, etc. Built and maintained multi-level test/development environment.
List of managed projects:
“ Evidence Based Decision Support Center (EBDSC)“: a large-scale web solution, combining web-based (front-end), standalone (back-end) data-entry applications using MS SQL Server as database layer. The solution has about 10,000 registered users. The application provides sophisticated and flexible search and data meta-analysis tools, a variety of highly customizable reports, stores customer preferences, presents results of analysis as real time charts, reports results in dynamically created Excel, PDF files, etc. Performs an analysis of healthcare knowledge data, organization management, evaluates the impact on organization budget of switching from one strategy to another - both in terms of cost and effectiveness.
Languages and Technologies used: C#, ASP.NET, ADO.NET, SQL (T-SQL, stored procedures, triggers, DTS, etc.), Remoting, COM, CSS, DHTML, DOM, XML, JavaScript, VB, VBA (MS Word, MS Excel, MS Access), MSXML parser, custom and third-party components. OO design and project lifecycle management, using RUP, UML, Agile, TDD, NUnit, Reafactoring, etc.
“Prometheus”: Performs an analysis of healthcare organization management, evaluates the impact on organization budget of switching from one strategy to another - both in terms of cost and effectiveness. The model demonstrates cost savings or additional (incremental) costs and any incremental effectiveness as a result of the switch. Technologies used: VB, Custom and third-party ActiveX controls, Memory Management, References, WIN32 API, Callback procedures.
“Gastro esophageal Reflux Disease (GERD) Screener application”: This application read, store in database, analyze and output primary diagnosis, based on data from scanned paper forms with patient answers. VB application. Technologies used: Serial Port Data Acquisition, ADO, custom ActiveX controls, Data-aware classes, WIN32 API, Callback procedures
Zynx Health Incorporated (http://www.zynx.com/)
Senior Analyst Programmer, IT Team Technical Lead, May 2000 – September 2001, ($80K)
Lead design, development and support of the most projects.
“Pain Tracker”: Multi-tiered, interactive Web Application. Technologies used: JavaScript, HTML/DHTML, XML, ASP, ADO, custom and third-party ActiveX components.
“Pain Manager”: Interactive Web-based tool designed for use by the multidisciplinary clinical team, quality management specialists, and administrators. This tool was developed to help hospitals and health care organizations adhere to HealthCare standards. Statistical treatment of data entered by user. Real-time graphical display of performance measure. Technologies used: JavaScript, HTML/DHTML, XML, ASP, ADO, COM.
“Safety Optimizer”: Interactive web application. Error-reduction tool featuring guidelines for improving safety within the acute care setting and for tracking progress in achieving improvements. The program is designed for use by the multidisciplinary clinical team, as well as for experts in quality-improvement and risk-management. Real-time graphical display of performance measure. Technologies used: JavaScript, HTML/DHTML, XML, ASP, ADO, COM. Lead design, development and support.
“Product Manager”: The back-end web application, that allow manage, create, search and maintain customer accounts. The created design and conception was used by another team members for extend functionality of application. Technologies used: JavaScript, HTML/DHTML, ASP, ADO, COM. Lead design, development and support.