Post Job Free
Sign in

Data Engineer Associate

Location:
Newport Beach, CA, 92663
Posted:
September 16, 2024

Contact this candidate

Resume:

CEDRIC BARNETT

North Hollywood, CA *****

*******@****.*** 818-***-****

AI EVANGELIST

Advise C-Suite to Establish IT Strategies across Organizations Translate Strategic Intent into Action; Deliver Business & Technology Changes; Build Holistic, Technology-Agnostic View of Strategy, Processes & Information Assets to Yield Best Return Approach problems hands-on when needed to help diagnose complex technical and business challenges and implement creative solutions. Employ entrepreneurial mindset and believe dead ends don’t exist when designing and deploying complex projects to solve organizational challenges (processes, information, cost reductions, and identifying and increasing IP assets value) using latest technologies. Champion IT risk reduction: analyze current architecture to pinpoint weaknesses, establish reusable frameworks, processes, and methodologies, and utilize technologies to create differentiation within 6 to 12-month time horizons. Certification: TOGAF Databricks Certified Data Engineer Associate Microsoft Certified Azure Data Engineer Associate GenAI IoT Cloud Enablement Integration Strategy & Re-Architecture AI Data Governance Strategy & Analytics PROFESSIONAL EXPERIENCE

Cognizant, College Station, TX April 2019 to Present Global Artificial Intelligence and Analytics (AIA) Principal Architect Architect and deliver creative and complex Modern Data solutions with GenAI capabilities on public cloud technologies like Amazon Web Services (AWS), Microsoft Azure, and Google Cloud Platform (GCP) as well as mentor and hire 20 + talented Architects

• Client: PG&E

• Architected and implemented a Palantir (Foundry) predictive model based on equipment health and geospatial data to send real-time outage notifications through Mulesoft.

• Client: Kore.ai XO Platform

• Created Virtual Assistant using Conversational Designer

• Integrate RAG & LLM with Search AI to analyze data in various document formats.

• Client: Cummins

• Architected and Implemented a Snowflake Chargeback Framework that calculated daily cost.

• Implemented Apache Iceberg to Snowflake POC

• Client: T-Mobile/Sprint

• Led Platform Engineering Team of 12+ in rearchitecting, refactoring and operationalizing $12M Azure Tier 1 solution that supports 150+ applications based on technologies such as Event Hub, Data Factory, Databricks (Unity Catalog, Overwatch), Snowflake, Synapse, Azure ML, Datadog, Yotascale, Trino, Dataiku, Kyvos, Gurobi, Thoughtspot, and Anodot and reduce FinOps by 10% Q/Q. Worked with T-Mobile Stakeholders to define and deliver quick win iterations and technology progression roadmap for cloud data-driven platform

• Client: Apple

• Authored and conducted 17-week Architecture Assessment and POC for Next-Gen ML Platform on GCP utilizing Dataflow, Dataproc, BigQuery, Databricks, and Snowflake

• Client: Canada Life

o Authored and conducted 15-week Architecture Assessment for Analytical/Data Science Platform driven by AutoML (SAS, Databricks, Dataiku, DataRobot, and H2O Driverless AI) on Azure to accelerate business insights led to a $1.5M delivery project.

• Client: Walgreens

o Architected and delivered Azure ML Environment and assisted Data Scientists in building the spaCy Model Development and Release Pipeline

• Client: Sanofi

o Architected and delivered Oracle on-premise to AWS RDS

• Client: Automobile Club of Southern California

o Architected and delivered $27M California Consumer Privacy Act (CCPA) solution based on technologies such as IIB, Collibra DGC, DAG MetaCenter, Cloudera, Teradata, and Workday in 10 months. o Architected and delivered SAP NS2 POC solution

• Client: Avista

• Authored and conducted 15-week Architecture Assessment for Modern Data and Analytics Platform driven by both Machine Learning and Salesforce on AWS using Cognizant Guiding Principles, leading to a $3.5M delivery project. CEDRIC BARNETT *******@****.*** PAGE TWO

THE HACKETT GROUP, Miami, FL Jan 2018 to April 2019 Principal Architect

Supported Cloud Strategy Sales Team in developing and closing sales by providing expertise in delivering revenue solutions on public cloud technologies like Amazon Web Services (AWS), Microsoft Azure, Oracle, and Google Cloud Platform as well as hired a talented team of Cloud Architects

• Client: Salesforce

o Assisted in upgrading a 96-node Splunk Cluster from v6.5.3 to v7.0.1 in 16 weeks on AWS

• Client: PVH

o Architected and delivered 10-node MapR Cluster on AWS to capture batch Marketing data for data discovery and building pricing models.

• Client: Fortune Brands

o Architected and delivered 10-node Hortonworks Cluster on Microsoft Azure DataLake, capturing, storing, and building Product Pricing Models using Tableau and QlikView

• Client: Outfront Media

o Architected and managed five resources in migrating DMP Explorer from Oracle Cloud (BDCS, EDQ, ODI, DBCS, OACS) to AWS S3 data lake (ElasticSearch, Lambda, EMR, Data Pipeline, Athena, Glue, API Gateway) in 8 weeks. TERADATA, San Diego, CA Nov 2015 to Jan 2018

Principal Architect

Provide top-down strategic direction for significant data architecture to support overall enterprise architecture initiative. Guide organization to produce comprehensive big data technology solution portfolio and ability to define important data solution patterns to create reusable assets within enterprise architecture. Interface across enterprise; act as visionary to proactively assist in defining direction for big-data technologies.

• Client: DSLab

o Architected and delivered first product release of Teradata Data Lake Management Product (Kylo) on AWS, evolving into www.kylo.io.

• Client: Horizon

o Architected and delivered Kerberos 10-node Teradata HDP 2.3.2 Cluster with Hadoop components, enabling capture and storage of batch and real-time medical insurance data for data discovery, analysis, and building predictive models.

• Client: Discover

o Increased performance and reliability by 50% for Nifi, Spark, Microsoft R, and H20 data pipelines by rearchitecting Kerberos 40-node fraud preventive Hortonworks cluster, including multiple big data technologies.

§ Enabled collection, storage, predictive model, and data discovery of credit card data by optimizing Capacity Scheduler with node labeling, reducing job queues, and using log4j to point pipelines toward separate HiveServer2.

§ Deployed Atlas to capture data lineage and metadata, Ataccamo to check data quality, and Attivo to catalog and classify data for Data Scientists.

• Client: HGST

o Improved job performance by 35% for Databricks and Impala by tuning Hive and Impala Queries, revamping Fair Scheduler, deploying Splunk to capture Hadoop logs and AppDynamcis for monitoring and optimizing on Kerberos 40-node Cloudera Cluster 5.8.2.

• Client: Ancestry.com

o Reduced implementation by one month for 100-node Hortonworks cluster by changing recommended 2-step upgrade process from HDP 2.1 to HDP 2.4.2 to new 1-step process, and becoming Hortonworks Best Practice for upgrading from HDP 2.1 to HDP 2.4.2.

§ Created ability to install and configure OpenTSDB, capturing Bosun monitoring agent information for HDP 2.4.2.

• Client: Major League Baseball

o Designed and deployed AWS POC for state-of-the-art, data-driven learning algorithms, solving business problems using latest technologies in neural networks, NLP, machine learning, statistical modeling, pattern recognition, and artificial intelligence; evolved into Statcast.

• Client: Loblaw

o Architected and delivered Kerberos 10-node Teradata HDP 2.3.3 Cluster, based on authoring, authenticating, auditing, and encrypting Customer Data with Ranger, Knox, Kerberos, and Protegrity.

• Client: Expedia

o Analyzed data and operational issues of Hortonworks 100-node ecosystem; performed Root Cause Analysis (RCA) and initiated corrective action based on ThinkBig Guiding Principles and Advocate Positions; delivered documents, addressed and prioritized issues; conducted Impact Analysis, and identified solutions.

§ Increased cluster performance and data trustworthiness by 50%. CEDRIC BARNETT *******@****.*** PAGE THREE

• Client: T-Mobile

o Conducted Security assessment of ‘as-is’ architecture of Hortonworks 100-node ecosystem, based on ThinkBig Guiding Principles and Advocate Positions, producing documentation and presentations/diagrams for dissemination to technical and business audiences based on Un-Carrier vision.

• Client: Zion Bank

o Conducted Disaster Recovery assessment of ‘as-is’ architecture of MapR 50-node ecosystem based on Teradata Unified Data Architect Guiding Principles and Advocate Positions, producing architectural roadmap for migrating existing Oracle and Teradata Data Warehouse and presenting to executives and management for approval. WIPRO, Schaumburg, IL Jan 2015 to Nov 2015

Solution Architect

Client: Zurich Financials

Managed 10 Hadoop Architects to identify key business and technology drivers; dovetailed into translating use cases into design patterns consistent with future platform strategy to solve different use cases for Data Profiling / Data Wrangling, building out GLM, Regression, Decision Trees, and Scoring models, and Metadata Management within Data Lakes. o Delivered ZNA Best Practices and Recommendations for data provisioning ZNA Data Lake, including building a Governance Catalog based on ZNA Data Governance Rules for business and technical data classification, translating into reference, conceptual, and detail-designed documents.

o Architected and delivered 10-node Palantir ecosystem (Raven & Gotham), allowing Data Scientists to use SAS data to profile, analyze, and build Regression, Decision Trees, and Scoring models. CEI America, Philadelphia, PA Jan 2013 to Jan 2015 DevOps Manager

Client: Comcast

Led and delivered game-changing data solutions for Comcast Xfinity digital business and built on latest technologies and patterns in data-driven environment; utilized data strategies, open source technologies and software engineering. Used Kanban Methodology in Rally to manage 15 Hadoop DevOps onshore and offshore resources daily. o Defined and reported Key Performance Indicators for 500-node Media Hortonworks (HDP) 2.3, reducing daily data errors/issues by 50%.

o Directed and delivered data, file, and storage conversion of MapR Hive and HBase to Hortonworks, evolving into Best Practices for Hortonworks.

INFOSYS, Bellevue, WA Jan 2011 to Jan 2013

Principal Architect

Client: Apple

Provided leadership in architecture, technical design, and development of large, scaled-out, batch, high-performing, multi-tenant Hadoop Infrastructure, allowing Apple to leverage data effectively for internal and external data ingestion between SAP, Hortonworks, Teradata, and Oracle.

o Architected and delivered first version of Hive with Hortonworks Professional Services. o Led Proof of Concept (POC) for data ingestion from Teradata and Oracle to HDP 1.0, evolving to enable collection, storage, modeling, analysis, and visualization of petabytes of data and 2.5k-node iRado Hortonworks cluster. o Architected and implemented SAP Hana

TECHNICAL EXPERTISE

GenAI: Palantir (Foundry), Lakera, Kore.AI, Dataiku, DataRobot,AzureML, SageMaker, H2O AI Data Governance Ataccama (Data Quality), Attivio (Data Catalog / Classification), Collibra Advance Analytics: Palantir (Contour), Kyvos, Gurobi, Thoughtspot, Databricks, Trificata, RapidMiner FinOps Yotascale, Anodot

Databases & NoSQL: Snowflake, Apache Iceberg, Synapse, Redshift, PostgreSQL, MySQL, Oracle, Teradata, Mariadb, Trino, Presto, Starburst, MongoDB,

DevOps Tools: Ansible, Terraform, DBT labs

Cloud Ecosystems: Databricks (Overwatch, Unity Catalog), Azure (OpenAI, Fabric,Event Hub, Data Factory),AWS

(IAM, EC2, S3, EMR, Data Pipeline, Glue, Athena, API Gateway, Lambda, Kinesis, Kinesis Analytics, DMS, VPC, Glacier, CloudWatch), GCP (Big Query, Dataflow, Dataproc), Oracle Cloud

(BDCS, EDQ, ODI, DBCS, OACS)

Monitoring: Splunk, Datadog

Programming Languages: Python, SQL, PL/SQL, Shell Scripting, R, SAS EDUCATION

Master of Business Administration (MBA), Finance, University of Pittsburgh, Pittsburgh, PA Master of Information Systems (MIS), University of Pittsburgh, Pittsburgh, PA Bachelor of Business Administration (BBA), Management, Minor: Mathematics and Computer Science, The College of William & Mary, Williamsburg, VA



Contact this candidate