Post Job Free

Resume

Sign in

Machine Learning Aws Cloud

Location:
Florissant, MO
Posted:
October 16, 2023

Contact this candidate

Resume:

JAMES DODSON

**** ****** **. **********, **. ****1

ad0euz@r.postjobfree.com

314-***-****

CAREER OBJECTIVE

I am currently seeking the role of Python Developer. I have formal training and experience in Statistical Modeling, Machine Learning, Deep Learning, Text Analytics, Data Pipeline Development, Data Analytics/Visualization and Software Design and Development. I have earned my Master’s degree in Supply Chain Management with a focus on Optimization Modeling and have earned a Certification in Data Analytics from Washington University in St. Louis, as well as a Certification in Data Science with Python from Cornell University. I have developed and implemented ETL solutions using AWS and Azure Cloud Platforms. I have Provisioned Databricks for statistical modeling using Python and Tableau to explain, predict and prescribe business outcomes. I have led development teams and deployed cost reduction applications including Data transformations and loading into Data Lake Systems using SQL, Python and Spark with AWS Cloud Formation. I desire this role because it will allow me opportunities to demonstrate technical and leadership expertise in managing business resources and improve process efficiencies. I have expertise in generating data structures and pipelines for analytics and dashboard visualizations prepared by programming languages and tools such as Python, R, SQL, Tableau, Azure, AWS, and Spark.

CAREER SUMMARY

Earned a Master’s degree in Supply Chain Management and a Data Analytics Certification at Washington University in St. Louis and earned a Certification in Python for Data Science at Cornell University.

Developed and Deployed Project Management Tracking Tool, application. A Python Web-based delivery management and schedule visibility system.

Published Coverity Tableau Dashboard to visualize Coverity application scans for compliance tracking.

TECHNICAL PROFILE

*This profile reflects my current skill set

Skill Years

Exp.

Proficiency

(Light, Medium, Strong, Expert)

Last used

Python, Scala, Spark, SQL, R, Perl, PHP, DHTML,

JavaScript, Bash, Tableau, Power BI, Excel

10 Expert Current

Databricks, Cloudera Hadoop, HDFS, Hive, Linux

Redhat, AIX and Spark Unix, GitHub, BitBucket

8 Strong Current

Azure, AWS, S3, EC2, EMR, Lambda, Glue, Redshift,

Machine Learning, Deep Learning, Snowflake, ECS,

Kubernetes, AzureDevOps, Jira

10 Strong Current

Oracle, SQL Server, MySQL, CosmosDB, MongoDB,

PostgreSQL

15 Expert Current

Data Engineering Data Science, Data Analytics 9 Strong Current MS Project 2019, Open Plan Professional Scheduling Tool Suite

12 Expert Current

Page 1 of 7 Resume of James Dodson

PROFESSIONAL EXPERIENCE

Insight Global – Pratt & Whitney – Raytheon Technologies 3/23 – current Senior Data Scientist/Python Developer

Being the first team member hired, I set up the Jira instance and the backlog of business features identified by the F100 Program leadership.

Worked closely with the business to capture requirements document them within the related user stories.

Worked with program leadership to prioritize features in the backlog and workout road mapping for expected delivery of features through the end of the year.

Generated PostgreSQL database schema and developed normalized database tables based on data requirements associated with business operations.

Developed Python ETL Pipeline for Work-In-Progress, WIP Shortages data to automate parts shortages identification and match-up possible “Swaps” and

“Transfers” parts to minimize downtime optimization models.

Prepared Gate 0-2, Gate 3-4, and DPAS Reports using “Openpyxl” Python library as well as SQLAlchmey and Excel to generate a multi-tab Excel Workbook to report on modules WIP as operations progress through each gate through completion.

Lead teams for Python development, Azure DevOps, and ML Model development.

Performed user story code reviews to ensure PEP8 standards.

Worked closely with Data Scientist to identify data requirements for a variety of ML Models for the purpose of predicting supplier performance by part and measuring probabilities of on-time parts delivery and turn-around-time, TAT for module deliveries.

eSolytics LLC – Federal Aviation Administration (FAA) – Remote 10/22 – 11/22 Senior Python Developer/Tableau Developer

Developed ETL automated process sourcing data provided by FAA using Alteryx to extract, scrub, and load data into a FAA Postgres database.

Prepared Data Modeling based on FAA provided data and FAA requirements including data normalization, primary and foreign key with referential integrity requirements.

Prepared on-demand Tableau dashboard visualizing data to provide performance insights of supplier contractual agreements.

Prepared Tableau burndown charts for conversion efforts from gas vehicles to electric vehicles to comply with sustainability KPI’s set by executive management. PSR Associates LLC – Price Waterhouse Coopers – Remote 01/22 – 09/22 ML Lead/Senior Python Developer

Prepared Data Pipelines loading MongoDb from various data sources and formats including CSV, Nested JSON, Parquet and Avro.

Scheduled and managed complex pipelines using Apache Airflow.

Prepared Databricks API’s using Postman and Notebooks to load generated Bearer Tokens for User Authentication Secrets and implement secrets in API’s or Notebooks within the Azure Workbench.

Performed special assignments on multiple Machine Learning, ML models using MLFlow with Tableau visualization within Azure Databricks.

Programmed Pipelines for MongoDb loads to support delivery of Wrangled and Cleaned data into Azure Data Lakes to support Spoke teams prepare ML models, Power BI and Tableau dashboards.

Page 2 of 7 Resume of James Dodson

Develop using Python, Spark, REST API and NiFi to automate Azure Databrick Provisioning and usage.

Work closely with Spoke teams to support automated pipelines implemented using Apache Airflow with various data frequencies for data consumption in Tableau and Power BI client dashboards.

Responsible for developing Databrick Notebooks, Jobs or Pipelines using Python, Scala and R for loading data from Azure Workbench into Azure Data Lake Systems, ADLS to Azure Blob File System Secure, abfss.

Work with the AAA scrum team on assigned work coordinated using Azure DevOps, Sprint Planning, with Daily Stand-ups and Sprint Reviews.

Responsible for entering Planned Features, User Stories and Tasks.

Responsible for entering Daily Work Accomplishments and related hours worked as well as reporting any Blockers daily on Planned Sprint Tasks Azure DevOps.

Responsible for delivering client solutions for assigned Azure DevOps Support tickets often within hours or no more than a couple of days.

Responsible for supporting Spoke teams, working assigned with Azure Databricks building Python, Scala and Spark workspaces.

Urbane Systems LLC – Neiman Marcus Group – Remote 10/21 – 01/22 Analytics Enablement Team Lead / Senior Python Developer

Work with directors and senior management to resolve Jira tickets submitted specifically for data needs to improve insights in business processes.

Developed and maintain Python applications to track Brand and Special Offers performance.

Used Kinesis Data Firehose to stream incoming data pointing to S3 buckets.

Created Lambda functions to invoke SageMaker Endpoint to scale Python scripts streaming incoming data.

Executed AWS Glue Crawler to process new incoming data and update Glue Data Catalog schema information.

Use SageMaker to build Predictive Machine Learning Regression Models for expected sales across specific NMG Specialty Beauty Products.

Used Snowflake and complex SQL to implement new data pipeline to Transform and Load data to identify repeat customers by brand based on criteria established by senior management.

Generated Analytics using Redshift and Snowflake with complex SQL to capture performance data points based on requirements aimed at providing visibility into specific delivery metrics for new “Click-to-Ship” fulfillment process.

Manage Off-Shore Analytics Team, responsible for monitoring, assisting and tracking off-shore Analysts for rapid Jira Tickets turnarounds.

Maintain Versioning Control using Git and BitBucket.

Prepared Tableau dashboard visualizing actual order fulfillment performance per major milestone compared to KPI’s establish by senior management.

Preparing process documentation using Confluence (Wiki) to establish standardized processes formalize strategic framework for our team members and the business. Page 3 of 7 Resume of James Dodson

McKinsey and Company – AT&T – Remote 10/20 – 9/21

Senior Data Analyst / Python Developer

Work in conjunction with teams within the business or the management team to establish business requirements.

Accountable for Extracting, Cleansing and Transforming data and developing pipelines in Palantir.

Accountable for training and testing data using K-fold and Cross-validation techniques for Machine Learning models within the Advanced Analytics Group at client sites.

Perform A/B testing on five markets using our largest market as our control market.

Prepared and presented Data Analytics presentations to communicate sales performance comparisons each to the control market based on different levers used during testing, then reject or not reject null hypothesis for each market comparison and presented findings to leadership and executive members.

Prepared complex SQL using Redshift and Athena within AWS as well as perform maintenance using S3 buckets and AWS Glue Data Catalog on data tables and views.

Optimized schemas on tables and views where the data is consumed for statistical modeling, data visualization reporting and adhoc support for Marketing and Advanced Data Analytics Group inquiries.

Developed Python ETL pipeline process for loading postpaid wireless and high- speed copper/fiber sales performance data into Palantir Cloud datasets for weekly Tableau dashboard reporting

Published interactive/dynamic Tableau Dashboard for user on-demand Customer Segmentation Visualization.

Designed and developed Python libraries and Tableau Dashboard for Predictive Customer Churn, Survival and Hazard Rates and visualized changes month over month.

Maintain Cross-Sell Tableau Dashboard to display product relationships based on product sales performance then applied Machine Learning Models to Analyze and Predict Cross-Selling opportunities within identified markets.

Identifying sales patterns to help drive formulation of marketing campaign offerings explanation along with recommendations to leadership.

Responsible for detecting sales trends within all regions then providing weekly cause/effect analysis to leadership.

Charged with researching and documenting Data Governance Best Practices regarding data storage and update processes, Pull Requests, Data Pipelines and Validations published on Confluence.

Worked with Advance Data Analytics group and Data Infrastructure group to document and provide process guidance for scheduling data updates and elevation processes to production.

Accountable for preparing the Data Update Readiness Checklist used to submit Data Pull Request’s, PRs for on-going data delivery.

Prepared and maintained Metadata Catalog within company Confluence.

Guided Brown-bag sessions to ensure teams understood systems capabilities and discussing important features necessary for elevating data jobs to production.

Responsible for code maintenance and versioning using GIT repository. Page 4 of 7 Resume of James Dodson

The Boeing Company – Hazelwood, MO 01/18 – 08/20

Senior Data Analyst and Python Developer

Partnered with directors and Senior management within all the organizations within Business and Project Management Office Group to identify and document major issues and opportunities statements.

Prepared the Business Case and Project Proposal for a new Management Decision Support System, DSS.

Secured project funding approval from the Directors and Senior Managers of Business and Supply Chain Systems organization to proceed to develop the new Decision Support System called Performance Tracking Tool, PTT.

Prepared analytical and statistical models using proven statistical methods to classify forecasted business change request and quantify and predict business resource needs.

Prepared OOP Design documents including Sequence and Class diagrams.

Prepared project plan based on Use Case Model Survey.

Worked with Senior Managers to develop prioritize KPI’s and acceptable performance metrics.

Prepared Markov Chain models using SAS scripts to identify viable management decision making alternatives and recommend actionable options with the most probable success rates to implement for performance improvement.

Lead effort with Senior managers to Prioritize Use Cases to Roadmap Iterations for Feature deliveries.

Developed SAS scripts to perform Monte Carlo models to measure risks and probability of success associated with several highly visible projects. The results were used to influence the decisions to reassign resources to other projects with higher success probabilities.

Performed SAS scripting to accomplish statistical analysis to establish baseline stats to benchmark results for the same statistical models prepared with Python using the same data to test and train with.

Responsible for provisioning and configuring SageMaker Pipelines using Kinesis Firehose Instantiating Dev/Test/Pre-Prod/Production Environments for deployment.

Implemented Data Models for the schemas stored in AWS Glue Data Catalog to deploy the Data Sets for PTT using AWS Cloud Formation.

Performed data analytics using Redshift and Athena from data stored in S3 buckets.

Wrote extensive Pyspark scripts using Data Frames, Data Sets & RDD's for transforming transactional Oracle database data and loading it into AWS tables and views.

Implemented AWS Glue Crawler to perform data cleansing including record deduplications.

Executed Glue Crawler to update Glue Data Catalog to maintain schema info for access from AWS Athena and Redshift.

Developed Dynamic Web based front-end using Dynamic HTML, JavaScript and AJAX.

Programmed ETL customized Data Extraction from S3 buckets using Restful API, Python, Numpy, and Pandas to automate Data Transformation and handle missing data. Programmed and Scheduled Data Loading into PTT’s MySQL database.

Programmed Server-side Python Class Objects to application processing of user requests, included but not limited to;

NLP classification to categorize Service Request Tickets from free form comments.

Page 5 of 7 Resume of James Dodson

Statistical analysis to Train and Test Data Sets for generating algorithms using methods such as Lasso, Ridge and Neural Networks for performance predictions and prescriptions.

Performed versioning control using GIT to push new production changes CICD using Kubernetes with Docker containers.

The Boeing Company – Hazelwood, MO 02/15 – 1/18

IT Scrum Master / IT Project Management Specialist

Developed and Published Tableau Dashboard to provided visibility and tracking of applications requiring Coverity scans for security compliance.

Designed, Developed and Deployed the Coverity Scans automated ETL process developed using Python with Pandas, MySQL then feed to Tableau.

Improved management visibility using a combination of Burndown charts with pop- ups and drilldowns within Tableau Server.

Performance Bar charts integrated with Smart Detailed Lists to display all remaining applications requiring scans categorized by organizations.

Managed migration of SQL Server 2008 R2 to SQL Server 2016 for Tech Debt across all organizations within Business and Supply Chain Systems, BSCS.

Coordinated and hosted scrum meetings for migrations of each application requiring migrations.

Performed weekly status on migration progress for all applications using SQL Server across BSCS.

Responsible for running the weekly Program Management Meetings (PMM) to report progress using schedule analysis techniques on the project schedule.

Prepare S-Curve analysis and reporting as well as generate Weekly Late reports.

Assisting in Agile implementation of Blockpoint 29 for 2018 including implementing AD&S agile processes and setting up Team Foundation Server, (TFS).

Prepared Velocity and Capacity planning and reporting, as well as TFS Tracking Dashboards.

Prepared and tracked Weekly S-Curve reporting presented in weekly CCB meetings for BMS leadership.

Prepared Tier I and II schedules using Milestone Professional.

Coordinated with internal product teams (IPT), suppliers The Boeing Company – Hazelwood, MO 02/12 – 02/15

Senior Python Application Developer/Data Scientist

Developed a multidimensional data warehouse using the star method for storing data cubes.

Lead project team the Integrated Product Data Management or IPDM System with Siemens Teamcenter.

Team Lead on various Application Development & Sustainment efforts.

Prepared Access Control Plan ACP, Disaster Recovery Plan DRP and various other Compliance documents for Boeing IT Compliance.

Performed vulnerability scans using Vera code and AppScan to identify software vulnerabilities then performed programming to remove vulnerabilities.

Maintained the MC2 Order Entry System for parts used by F18 and F15 production lines and field support.

Page 6 of 7 Resume of James Dodson

Responsible for interface for MC2 request for products and product information that fed the Technology Services Fulfillment Systems.

Researched, designed, programmed and managed the Technology Services Order Entry and Fulfillment systems.

Analyzed, designed and normalized data structures of the Technology Services Fulfillment System Database using Oracle Database to process order request for client accounts.

Analyzed, designed and normalized data for the Keystone Fulfillment System Database using Oracle Database and Oracle Forms to process order request for client accounts.

Used PL-SQL and SQL Plus to assist in the design as well as the maintenance and management of the Keystone Application.

Project Team Lead on the design, development and delivery of the Project Management Tracking Tool (PMTT) application.

Designed a MySQL database derived from data storage requirements identified within project use cases.

Developed PMTT, a web-based application, used to provide detailed delivery schedule visibility and management reporting tools across all programs within the Technical Publications.

Performed requirements gathering and documentation including preparing Use Cases with end users, stakeholders, middle, senior and executive management.

Maintained requirements, code designs and code versioning within GIT to maintain Release Tracking to promote new changes to production servers.

Designed and Developed User Interface, using Dynamic HTML, JavaScript and AJAX and deployed on Linux Redhat Servers running Apache Web Services.

Administered Linux Redhat servers, installing software including Python, MySQL and Apache.

Lead project team as well as developed PMTT using Python, Perl, backed with MySQL Database.

Through formal channels obtained the necessary hardware equipment to implement a Production Linux Web Server for PMTT to support approximately 500 end users on the Boeing network.

Conduct code review meetings to ensure first-time quality.

Scrutinize customer User Acceptance Testing to gauge how through the UA Testing was to ensure customer satisfaction.

Responsible for identifying and managing Program/Project risks of all types including supplier, cost, technical, and schedule risk.

Prepared, then presented quantitative financial information to explain the substantial savings of $250,000.00 annually achieved through the implementation of PMTT using financial graphs, charts, tables, etc. to mid-level and senior managers as well as Executive Change Board members to obtain funding for PMTT web server. The Boeing Company – Hazelwood, MO 08/07– 02/12

Master Integrated Planning & Scheduler

Accountable for analyzing project contracts and statements of work (SOW) to generate the Integrated Master Plan (IMP) and Integrated Master Schedule (IMS) and ensure customer requirements are identified and accounted for in the schedules.

Responsible for coordinating with various suppliers, internal product teams (IPT), and the C17 project team to prepare and track integrated master schedules using Open Plan Professional (OPP) scheduling application for upgrades and maintenance of the C17 aircraft and simulation trainer. Page 7 of 7 Resume of James Dodson

Created the C17 Bucket Schedule using Microsoft Project, which color coded tasks graphically by supplier, IPT and project team to coordinate efforts across projects to minimize trainer downtime.

Developed the automated Downtime Report, which is distributed to the customer to coordinate training schedules of the air crews and minimize downtime of the C17 Trainers at various Air Force Bases around the country.

Worked with the various suppliers around the country to schedule tasks, communicating dependencies between the suppliers and the project teams, as well as identifying risks and coordinating mitigation plans associated with those risks. A.G. EDWARDS – St. Louis, MO 10/06–07/07

IT Project Manager

Attended A.G. Edwards PM Boot Camp to learn about the A.G. Edwards Project Management Framework.

Prepared the Project Charter for the FC Payout Project based on the Project Feasibility Study.

Responsible for working with the business unit Fee Based Operations to determine the Project Scope and to get Sign-off on the Business Requirements Document

(BRD) and Use Cases.

Coordinate resources from other IT Departments to scheduled tasks for items such as: Database implementation, Builds and Deployment, Performance Testing, Operational Readiness Testing and Production Support.

Closely manage Project Cost by collecting Estimate to Completion (ETC) by task from each team member each week for the upcoming week. Continually challenging and scrutinizing each request to change scheduled dates or increase hours on any given task.

Constantly analyzing Estimates at Completion (EAC) to manage all Critical Path Tasks are on schedule to ensure on time delivery of the project’s final deliverable. POST GRADUATE PROJECT EXPERIENCE

R-Studio Cloud – Predicting Covid-19 infection rates on minorities living in the United States

Harvard University – Statistical Modeling with R

Project Objective was to provide evidence through statistical analysis to predict Covid-19 infection rates among minorities living in the United States.

Gathered data from Harvard on-line data sources, published CDC data, and Kaggle Data source.

Separated data into a Training set and Test set by randomly allocating 80% to Train set and 20% to Test set.

Removed multicollinearity by iteratively using automated Forward Selection, Backward Elimination and Stepwise Regression removing variables with high variance inflation factors, VIF.

Experimented with a variety of regression methods including Ridge Regression, Lasso Regression and Random Forests with regression trees.

Implemented a Neural Net with one hidden layer and skip layers. This non- parametric model captured a variety of interaction terms between the different features.

Amazon Web Services - Sentiment Analysis Project

Harvard University – Text Analytics

Page 8 of 7 Resume of James Dodson

Use Amazon Web Services API to retrieve customer reviews by product.

Performed Sentiment Analysis using Python 3, spaCy, and NLP to categorize reviews for the purpose of sorting according to relevance to support the reader’s purchase decision making.

Relevance was determined based on comparing means of word and phrase frequencies as tagged using parts of speech, POS allowing all the reviewers to impact which topics are relevant for a purchaser to know while making a purchasing decision.

Reviews are processed using WordNet and Google N-gram Corpus accessed by Python, spaCy and Natural Language Tool Kit (NLTK). Big Data Analytics – Class Projects

Harvard University – Big Data Analytics and Visualizations

Analytics Projects with hands-on experience using PySpark, PyTorch, Spacy, Scikit- learn, and TensorFlow. Worked with real-world Big Data using PySpark with Python, Keras API for TensorFlow for Data Visualization using Python Seaborn and Tableau.

Analyze Big Data Stored in AWS managed by EC2, S3, Redshift, Athena and complex SQL. Deep Learning Neural Network in Python to train and test models then visualize results using Matplotlib, Seaborn and Tableau.

Applied a Deep Learning Neural Network to predict the gender of a voice using acoustic properties of voice and speech. The model had 2 hidden layers, each layer with 100 nodes. The model was compiled and fitted, then quantified (scored) the model’s predictions compared to actual genders labels.

Prepared Python code to fit K-means model to a dataset, then visualized the results by creating a plot. Then determined how to get the centers of the K-means cluster and plot them.

Page 9 of 7 Resume of James Dodson

TECHNICAL CERTIFICATES & LICENSES

● Data Analytics Certificate – Washington University in St. Louis

● Python for Data Science – Cornell University

● Certificate in Data Science – Harvard University (in-progress 5/2024)

● AWS Certified Data Analytics – (Certificate AWS-DAS-C01 in-progress 6/2023)

● AWS Certified Cloud Practitioner – (Certificate AWS-CLF-C01 in-progress 6/2023)

● AWS Certified Machine Learning – (Certificate AWS-MLS-C01 in-progress 7/2023)

● AWS Certified Solutions Architect – (Certificate AWS-SAA-C02 in-progress 8/2023) EDUCATION

Bachelor of Science – Accounting

Southern Illinois University

Carbondale, IL

Master of Science in Supply Chain Management

Washington University at St. Louis

Olin Business School

St. Louis, MO

Master of Science in Data Science (in-progress 5/2024) Harvard University

Harvard Extension School

Cambridge, MA

REFERENCES

Francisco Rivadeneyra Senior Manager 719-***-**** Nathan Elmenhurst Senior Manager 425-***-****

Amy Yamamoto Manager 206-***-****

Tywana Young Manager 314-***-****

Peter Holder Senior Team Lead 610-***-****



Contact this candidate