Post Job Free
Sign in

Enterprise Data Solution Architect

Location:
Toronto, ON, Canada
Posted:
July 14, 2024

Contact this candidate

Resume:

Missagh Sadeghi

Toronto, Canada

514-***-**** *******.*******@*****.***

Profile

Experienced professional in architecture, design, modernization, development and implementation of large scale business applications/data frameworks

Migration of the existing data warehouses into the cloud (GCP, Azure, AWS)

Architecting, designing and advising large-scale enterprise data solutions and pipelines

Demonstrated working experience of Big Data (Hadoop, Schedulers, ETL, ELT)

Established Talend framework and center of excellence as extraction tool (ETL/ELT)

Solid knowledge/experience in RDBMS (SQL, NoSQL, TSQL, Programmability etc.)

Extensive experience in porting RDBMS architectures into cloud (log based CDC)

Possess strong expertise in code reviewing, optimizing, debugging and gatekeeping SKILLS

Cloud Technologies GCP (Composer/AirFlow, DBT, BigQuery, TerraForm, DataFlow, DataFusion, ML)

Azure (Synapse Analytics, Data Lake, Data Factory, Data Flow, Databricks, SQL)

Development environments Cloudera, Hortonworks, CLI, Talend, Hue, Eclipse, Different IDE Development modeling UML, Visio, ArchiMate, TOGAF, Gliffy, Miro, Borland Modeling Methodology Agile Scrum, Ace

Version controlling Sub-Version, Team Source, StarTeam, Git, Bitbucket Work Experiences

Scotia Bank, Toronto

www.scotiabank.com

2021– Present

Data Application Solution Architect – Team Lead

Canadian Banking Data Services in Scotia bank has more than 700 sources of data from internal/external providers that needs to be loaded, ingested, transformed and extracted from the Enterprise Data Lake. As the solution architect in the team, I deal with modernizing all of data pipelines for better solutions to enhance the ETL/ELT processes, data availability, data lineage, audit, logging requirement, data encryption and data quality.

Modernization and data solution for the existing on premise Enterprise Data Lake to the cloud with SOA

– Using Talend (Studio, Remote Engine, Management Console), Google Cloud Platform (BigQuery, Airflow, DBT, Universal Catalogue), Azure (Data Factory, Data Flow, Synapse)

Leading a POC on creating a IBM IMS on Mainframe data to be sent to RDBMS – Using IBM IMS, IBM CDC Agent

Run a POC on publishing log based RDBMS CDC into cloud – Using IBM IMS, GCP Dataflow, GCP BigQuery

POC on real-time data ingestion from on premise to GCP cloud – Using Kafka, GCP BigQuery, Sink Connector

Establishing the data integrity and quality for file transfer and RDBMS data pull by checksum, hash total

(Standard Deviation, Aggregation), threshold check and record count – using different upstream sources capabilities to meet the requirement

Upgrading the Hortonworks Hadoop 2 to Cloudera Hadoop 3 – using Cloudera, IBM Spectrum Scale, IBM ICSF

Upgrading the Hadoop data warehouse file formats to save space and enhance performance (resulted tax credit) – Using Parquet columnar storage format

Establishing new data Service Oriented Architecture pipeline with most recent standards and requirements for the security as well as the performance – Using Talend, Hadoop, Yarn, Kubernetes, Docker, MinIO, S3, JO/SFTP

Ground up a new SOA ETL eco-system for developing a new data applications – using IBM Linux AIX servers, IBM GPFS, Autosys scheduler

Design, document and lead a new CI/CD pipeline with threat and vulnerability online scanner – Using Maven, Azure cloud DevOps, JFrog/Nexus Artifact Repository, Git/Bitbucket Code Repository, Restful APIs, BlackDock Artifact Scanner

Design and lead configurable archive utility (archive, retrieve, purge along with data validity) for different File Systems – Using Dell Object Storage ECS, Spark, HDFS, Autosys

Integrating the Password Access Management (PAM) Vault to secure the credentials integrated with across the different eco-systems – using PAM APIs, GCP secrets APIs, Java

Implementation of Single Sign On (SSO) for different components to communicate seamlessly and securely within the bank and external integrations – using SAML, Token, Active Directory

Established Center of Practice to develop performant and secure ETL codes – using Talend, RE, TMC, Autosys, Bitbucket, CICD pipeline

Cloudera Hadoop, Talend, RE, TMC, Java, Autosys, Tidal, Bitbucket, Azure Devops, Maven, GCP, Dell ECS

CIBC Bank, Toronto

www.cibc.com

2019– 2021

Data Application Solution Architect – Team Lead

CIBC has one of the most enhanced AML and corporate security applications and it keeps renewing them to accommodate the latest standards of the matter. My objective was re-designing the legacy applications and ensure all the new requirements and regulations are met as well as enhancing the performance. This integrates with few sources that required data extraction, ingestion and cleansing accordingly.

Modernized the data solution/model along with Optimized the Hive queries according to the resource restriction, SLAs and business requirements – Using Hive, Ambari, Tez, YARN, DAG

Lead a team of 15 for CorpSec project – using NICE Actimize, Java, Oracle, Hive

Designed and led the team to create centralized data service – Using XML payload to optimized data request, SOA APIs for data sources

Re-designed and led a team of developers for FCLM project – using NICE Actimize, Java, MS-SQL, Hive

Distributed system (HDFS), Cloudera data management platform, Ambari, Hive, Autosys, CLI, Eclipse, GIT, Maven, Oracle, MS-SQL

TD Bank, Toronto

www.td.com

2017– 2019

Big Data Solution Designer/Architect – Team Lead

Considering having one of the biggest Canada’s bank data-lake, my objective was designing a solution for the different business sections requirements. We need to know the coming data structure, how to match them, where to find the required info, how to generate the needed info and hand them off to the desired downstream. The most exciting projects/challenges I accepted were from 29 different source applications’ data for all the North America customers for AML (ECRR) and VIP Flags. I designed, led and also helped on coding and delivered by me.

Lead a team of 12 for AML (ECRR) Hardening/Enhancement project – using Big Data, Java, Hive, Bash, Oozie

Designed and led a team of 8 developers for VIP project – using Big Data, Hive, Java, Bash, Oozie Workflow and coordinator

Designed, trained and led development of 3 segment delivery projects - using Bitbucket, Hive, Java, Bash, Python

Redesigned the development platform – based on the whole bank’s environment architecture

Recode and re-designed continuous integration – using GIT (Bitbucket), Nexus, Jenkins, Bash scripting

Code reviewing and Code optimization consulting

Distributed system (HDFS), Cloudera data management platform, Hue, Sqoop, Hive, Oozie, CLI, Eclipse, GIT, Maven, MS-SQL

Senior Big Data designer / developer

As a consulting company, we designed, implemented and delivered a Big Data project for one of our well- known clients, AMD. In the project, we collected data of computers across the world into AWS S3 bucket and put the valid data out of it into Hadoop to generate the desired report off.

Designed the ETL through working with the business analysis – based the SRS and investigating the client’s requirements

Implemented data sanity validation and ETL from S3 to HDFS through custom NiFi processors – using AWS, NiFi, Java

Designed and implemented data ETL and customized content validation in different zones in HDFS – using Hive, Spark, Java

Created the client’s requested reports off Big data platform – using Hive

Initiated, drove and developed a framework to “unit test” all the ETL – using Bash, Python, Hive, Spark

Automating the whole ETL and report generating – using Oozie, Bash, Python

Exporting data to SQL Server from HDFS – using Sqoop, HCatalog, Hive

Optimizing data storage and accelerating ETLs – using Snappy and ORC file format

Facilitating report generation and automation to be able to be easily developed by client for any further probable needs

Code reviewing and Code optimization consulting

Technical Environment: AWS, Nifi, Hadoop file distributed system (HDFS), Hortonworks data management platform, Ambari, Sqoop, Hive, Oozie, CLI, Eclipse, Maven, MS-SQL NOSA, Iran

www.nosa.com

2002 – 2015

Technical Team Lead / Scrum Master / Developer (2010 – 2015)

Analyzed, designed, and led the development of NOSA ERP solution (CRM, Accounting, Cash Flow, Inventory, Sales, Payroll software systems) with MS-Office integrated reporting - using SOA (COM+, DCOM, MIDAS), ADO, SQL Server, Object Pascal (Delphi)

Recruit trained and built a team for next generation of CRM tools

Integrated windows app with SOAP on HTTP layer over internet – Using IIS Technical Environment: Delphi IDE, Eclipse, Maven, Multi-tier Programming, ADO, REST, spring, Hibernate, MS-SQL, MTS, COM+, DCOM, MIDAS, SOAP, XML, Client Dataset Jonah Group (Collaborating with AMD), Toronto

www.JonahGroup.com

2016 – 2017



Contact this candidate