Post Job Free

Resume

Sign in

Data Engineer Service Delivery

Location:
Menlo Park, CA
Salary:
80/hr w2
Posted:
December 05, 2022

Contact this candidate

Resume:

Robert P Wenzel

Menlo Park, CA ***** 650-***-****

adtvil@r.postjobfree.com www.linkedin.com/in/robert-p-wenzel

Data Wrangler

Data Warehousing Data/Talend Data Integration

Collaborative problem solver who uses process improvement, ingenuity, and training strengths to turn around or enhance overall efficiency. Expert in identifying problems and architecting and implementing solutions, generally thinking outside of the box

Experience includes:

Organization Design and Build Extensive Googling

Production Ready Databases Service Delivery

TECHNICAL SKILLS

Platforms: Amazon Web Services – EC2, S3, SMS, Lambda, Aurora, RDS, Azure VM, Data Lake and Storage, Docker Containers

Databases: Oracle, MySQL, MSSQL, DB2, Sybase, Informix, Hive, Postgres, Aurora RDS, Athena, Vertica, Impala, Hive, Redshift, Snowflake

Big Data Tools: Hadoop/HDFS, Pyspark, Airflow

Languages: Python, Perl, JavaScript, PHP, Unix Shell,/BASH, SAS, R, XML, HTML, CSS

Debugging: Java & C++

Software: Talend-Data Integrator, Pentaho-Kettle, MS-Office

Reporting: MicroStrategy, Crystal, Cognos

EXPERIENCE

JCC and Hearst Media Company 4/2021 – 8/2022

Data Engineer – Consultant, Remote, San Francisco, CA

•Designed and implemented automated processes primarily using Talend, Python and Airflow, to reduce downtime.

•Set up processing for the automating processing report information for 3 California Courts: Los Angeles, Yuba and Kings counties using Azure, with virtual machines, Loaded data into Snowflake from Azure Storage folders. Also assisted in fixing data with various other counties. Used Talend, Python, SQL in Snowflake, Oracle, and MSSQL. Wrote a program to load delimited files directly into MySQL using Python with Pandas which aided in diagnosing data problems along with a program list of failed Talend programs errors over the past “n” days.

•The following Talend components were used to create all of the JCC Talend Jobs: logStatus,

tAzureStorageConnection, tAzureStorageGet, tAzureStorageList, tAzureStoragePut, tChronomenterStart,

tChronomenterStop, tDBClose, tDBConnection(Oracle), tDBInput(Oracle), tDie, tFileDelete, tFileExist,

tFileInputDelimited, tFileList, tFileOutputDelimited, tFileTouch, tFilterRow, tGPGDecrypt, tJava,

tJavaFlex, tLogCatcher, tMap, tPivotToColumnsDelimited, tPostjob, tReplace, tRunJob, tSendMail,

tSortRow, tUniqRow and tUnite.

•Was given the task to figure how to cleanse and process Los Angeles Court. They are the largest county in California which consists of 1/3 of all court data. So I worked extensively with their many inputs and large amounts of data over the past year.

Course Hero 1/2020 - 6/2020

Senior Data Engineer, Remote Redwood City, Ca

•Functioned as a backup DBA for Course Hero’s Aurora RDS (MySQL)

•Automated generating PERCONA database changes.

•Created an ansible script to build a fresh OS and Application standalone (Bare Metal) system, in case of AWS failure lasting more than 30 minutes.

•Conceived and launched filtering, backing up and slack notifications for Linux and RDS errors

•Researched and prepared AWS RDS/Aurora parameters to move databases from Production Account to a new Development Account

Apixio 7/2018 – 11/2019

Data Integration Specialist, San Mateo, Ca

•Retrieved, cleansed, transformed and integrated our clients’ Medicare Advantaged data into Apixio Risk Assessment systems, primarily using Python and SQL extensively

•Researched, simplified, and automated some of our data processing of our client's data and reduce the time it took to process a customer’s data

•Worked Hive under Hadoop to query large databases containing Medicare clients

•Also worked with Linux and Bash commands/Scripts

Quicken 10/2016 – 11/2017

Data Warehouse Staff Engineer, Menlo Park Ca

•Coordinated and managed and extended Quicken’s off-shore team of 10 people for the implementation of Quicken’s Marketing Data Warehouse initiative.

•Designed and implemented Quickens Financial Data Warehouse conversion from excel and Access DB.

•Used AWS EC2, S3, SES, Lambda, RedShift, RDS, Talend and Python to accomplish these tasks.

•Contributed and managed Tableau reporting group for Quickens data to management.

•The following Talend components were used to create all of the Quicken Talend Jobs: tContextDump,

tContextLoad, tDie, tFileDelete, tFileExist, tFileInputDelimited, tFileOutputDelimited, tFileOutputExcel,

tFileOutputRaw, tFileProperties, tFileRowCount, tFilterColumns, tFilterRow, tFlowToIterate, tFlowTolerate,

tIterateToFlow, tJDBCClose, tJDBCCommit, tJDBCConnection, tJDBCInput, tJDBCOutput, tJDBCRollback,

tJDBCRow, tJava, tJavaFlex, tJavaRow, tLibraryLoad, tLogCatcher, tLogRow, tLoop, tMap, tMysqlClose,,

tMysqlCommit, tMysqlConnection, tMysqlInput, tMysqlOutput, tMysqlRow, tPostjob, tPrejob, tPrepjob,

tRedshiftBulkExec, tRedshiftClose, tRedshiftCommit, tRedshiftConnection, tRedshiftInput, tRedshiftOutput,

tRedshiftOutputBulk, tRedshiftOutputBulkExec, tRedshiftRollback, tRedshiftRow, tReplicate, tRunJob,

tS3BucketExist, tS3Close, tS3Connection, tS3Delete, tS3Get, tS3List, tS3Put, tSendMail, tSortRow

and tUniqRow.

MyBuys/Magnetic 1/2008 – 7/2016

Senior Data Engineer/Principal Professional Services Engineer, San Mateo, Ca

•Introduced Talend to our Professional Services group, replacing Web Methods which saw an immediate reduction of 75% in the time to implement a new client from 2 -3 months to 2 to 3 weeks.

•Within the first two years of using Talend we generated over 1600 Talend Jobs to implement our new clients. Each client required 8 jobs to clean and load their data into our SAAS system.

•At the beginning of the 3rd year of using Talend, I conceived a way to create generic 8 Talend program which used context variables from a file to clean and load all of a client’s data. This reduced the time to load a new client’s data from 1 to 2 days and our professional services staff from 12 to 4 people. Over the next 5 years very few modifications were needed to the 8 jobs. We implemented over 1500 clients using the generic Talend jobs. We came up with a unique way to handle xml input beside delimed files. It could have been easily extended to handle JSON files if we needed to.

•The following Talend components were used to create all of the Quicken Talend Jobs:

tBufferOutput, tContextDump, tContextLoad, tDie, tFTPConnection, tFTPDelete,

tFTPFileList, tFTPFileProperties, tFTPGet,

tFTPPut, tFTPRename, tFileCopy, tFileInputDelimited, tFileList, tFileOutputDelimited,

tFileProperties, tFixedFlowInput, tFlowToIterate, tGPGDecrypt, tIterateToFlow, tJava,

tJavaRow, tLogRow, tLoop, tMap, tNormalize, tOracleCommit, tOracleConnection,

tOracleInput, tOracleOutput, tOracleRow, tOracleSP, tReplicate, tRunJob, tSleep,

tSortRow,and tSystem.

•Migrated from Talend Enterprise edition to the free Talend version since the free version provided us everything we needed.

•Introduced Talend to our Professional Services group, replacing Web Methods which saw an immediate reduction of 75% in the time to implement a new client from 2 -3 months to 2 to 3 weeks

•Conceived and coordinated a replacement of our Talend ETL to use generic programs which further reduced the time to implement a new client from 2 to 3 weeks to 1 to 2 days.

•Replace Talend Enterprise edition with the free Talend version and over the years migrated from Talend 2.xx to 5.xx skipping only 1 major release.

•Over the 8.5 half years, helped to implement over 1700 clients using an average of 8 Talend programs per client. I with over 500 of them either by myself or help other members of our Team.

•Researched and processed several tens of terabytes of data using Hive SQL queries to successfully refute fraudulent billing of two of our subcontractors which returned hundreds of thousands of dollars

•Implemented the migration of data from our Vertica database to Impala to mitigate the expansion of Vertica expenses to the open-sourced database

•Enhanced and optimize SQL queries to reduce processing time and resources

•Mentored and trained all of our Professional Services personnel in Ann Arbor after my first year and served as the group’s technical consultant for our company.

Other miscellaneous positions from 12/1995 – 1/2008,

5 years SeeCommerce, 3.5 years at Blazent and 2 years at Merced Systems 1 year at Escend and 1.5 yeasr at Tradec

06/1992 – 11/1996 Consulting Technical Manager

Prism Solutions - Sunnyvale, CA

•I was employee at Bill Inmon’s Data Warehouse company. Bill is considered one of the grandfathers of Data Warehousing.

•In the beginning I was part of pre and post sales, along with IT, QA, Training and finally implementation/consulting.

•I provided analysis, Data Modeling and construction of several Data Warehousing projects for 45 companies in the US, Canada and Northern Europe eg: Scandinavian Airlines System, Maersk, Lego, Comerica, FHP, Zurich Insurance, Blue Cross of MO and Cardinal Health, Church of Latter Day Saints, University of California, Berkeley, Florida Power and Light, State of Florida and several other companies Using PRISM Products, I designed, programmed and debugged maps, user exits; change data capture programs and the overall flow of data to insure successful implementations of Data Warehouses.

I have other previous experience which has included management positions ranging from leading small teams to be the Director of the Data Center for 80 plus employees of a major Software Company.

EDUCATION

Master of Science: Operations Research, George Washington University, Wash. DC

Bachelor of Arts: Mathematics, University of Minnesota, Duluth MN

Machine Learning: Introduction to Machine Learning, Analytics Vidhya

Community Colleges: Programming classes – Java, C++, PHP, Pascal, Pl/1

Cousera Courses: Primarily Data Science & Machine Learning, over 25 completed courses

Experimentation for Improvement Introduction to Data Science

Data Science & Machine Learning Python Programming

Data Visualization Machine Learning

Using Databases with Python Pattern Discovery in Data Mining

Practical Machine Learning Using Python to Access Web Data

PostgreSQL R Programming

Set up Hadoop/Hive in development



Contact this candidate