Missagh Sadeghi
Toronto, Canada
514-***-**** *******.*******@*****.***
Profile
Experienced professional in architecture, design, modernization, development and implementation of large scale business applications/data frameworks
Migration of the existing data warehouses into the cloud (GCP, Azure, AWS)
Architecting, designing and advising large-scale enterprise data solutions and pipelines
Demonstrated working experience of Big Data (Hadoop, Schedulers, ETL, ELT)
Established Talend framework and center of excellence as extraction tool (ETL/ELT)
Solid knowledge/experience in RDBMS (SQL, NoSQL, TSQL, Programmability etc.)
Extensive experience in porting RDBMS architectures into cloud (log based CDC)
Possess strong expertise in code reviewing, optimizing, debugging and gatekeeping SKILLS
Cloud Technologies GCP (Composer/AirFlow, DBT, BigQuery, TerraForm, DataFlow, DataFusion, ML)
Azure (Synapse Analytics, Data Lake, Data Factory, Data Flow, Databricks, SQL)
Development environments Cloudera, Hortonworks, CLI, Talend, Hue, Eclipse, Different IDE Development modeling UML, Visio, ArchiMate, TOGAF, Gliffy, Miro, Borland Modeling Methodology Agile Scrum, Ace
Version controlling Sub-Version, Team Source, StarTeam, Git, Bitbucket Work Experiences
Scotia Bank, Toronto
www.scotiabank.com
2021– Present
Data Application Solution Architect – Team Lead
Canadian Banking Data Services in Scotia bank has more than 700 sources of data from internal/external providers that needs to be loaded, ingested, transformed and extracted from the Enterprise Data Lake. As the solution architect in the team, I deal with modernizing all of data pipelines for better solutions to enhance the ETL/ELT processes, data availability, data lineage, audit, logging requirement, data encryption and data quality.
Modernization and data solution for the existing on premise Enterprise Data Lake to the cloud with SOA
– Using Talend (Studio, Remote Engine, Management Console), Google Cloud Platform (BigQuery, Airflow, DBT, Universal Catalogue), Azure (Data Factory, Data Flow, Synapse)
Leading a POC on creating a IBM IMS on Mainframe data to be sent to RDBMS – Using IBM IMS, IBM CDC Agent
Run a POC on publishing log based RDBMS CDC into cloud – Using IBM IMS, GCP Dataflow, GCP BigQuery
POC on real-time data ingestion from on premise to GCP cloud – Using Kafka, GCP BigQuery, Sink Connector
Establishing the data integrity and quality for file transfer and RDBMS data pull by checksum, hash total
(Standard Deviation, Aggregation), threshold check and record count – using different upstream sources capabilities to meet the requirement
Upgrading the Hortonworks Hadoop 2 to Cloudera Hadoop 3 – using Cloudera, IBM Spectrum Scale, IBM ICSF
Upgrading the Hadoop data warehouse file formats to save space and enhance performance (resulted tax credit) – Using Parquet columnar storage format
Establishing new data Service Oriented Architecture pipeline with most recent standards and requirements for the security as well as the performance – Using Talend, Hadoop, Yarn, Kubernetes, Docker, MinIO, S3, JO/SFTP
Ground up a new SOA ETL eco-system for developing a new data applications – using IBM Linux AIX servers, IBM GPFS, Autosys scheduler
Design, document and lead a new CI/CD pipeline with threat and vulnerability online scanner – Using Maven, Azure cloud DevOps, JFrog/Nexus Artifact Repository, Git/Bitbucket Code Repository, Restful APIs, BlackDock Artifact Scanner
Design and lead configurable archive utility (archive, retrieve, purge along with data validity) for different File Systems – Using Dell Object Storage ECS, Spark, HDFS, Autosys
Integrating the Password Access Management (PAM) Vault to secure the credentials integrated with across the different eco-systems – using PAM APIs, GCP secrets APIs, Java
Implementation of Single Sign On (SSO) for different components to communicate seamlessly and securely within the bank and external integrations – using SAML, Token, Active Directory
Established Center of Practice to develop performant and secure ETL codes – using Talend, RE, TMC, Autosys, Bitbucket, CICD pipeline
Cloudera Hadoop, Talend, RE, TMC, Java, Autosys, Tidal, Bitbucket, Azure Devops, Maven, GCP, Dell ECS
CIBC Bank, Toronto
www.cibc.com
2019– 2021
Data Application Solution Architect – Team Lead
CIBC has one of the most enhanced AML and corporate security applications and it keeps renewing them to accommodate the latest standards of the matter. My objective was re-designing the legacy applications and ensure all the new requirements and regulations are met as well as enhancing the performance. This integrates with few sources that required data extraction, ingestion and cleansing accordingly.
Modernized the data solution/model along with Optimized the Hive queries according to the resource restriction, SLAs and business requirements – Using Hive, Ambari, Tez, YARN, DAG
Lead a team of 15 for CorpSec project – using NICE Actimize, Java, Oracle, Hive
Designed and led the team to create centralized data service – Using XML payload to optimized data request, SOA APIs for data sources
Re-designed and led a team of developers for FCLM project – using NICE Actimize, Java, MS-SQL, Hive
Distributed system (HDFS), Cloudera data management platform, Ambari, Hive, Autosys, CLI, Eclipse, GIT, Maven, Oracle, MS-SQL
TD Bank, Toronto
www.td.com
2017– 2019
Big Data Solution Designer/Architect – Team Lead
Considering having one of the biggest Canada’s bank data-lake, my objective was designing a solution for the different business sections requirements. We need to know the coming data structure, how to match them, where to find the required info, how to generate the needed info and hand them off to the desired downstream. The most exciting projects/challenges I accepted were from 29 different source applications’ data for all the North America customers for AML (ECRR) and VIP Flags. I designed, led and also helped on coding and delivered by me.
Lead a team of 12 for AML (ECRR) Hardening/Enhancement project – using Big Data, Java, Hive, Bash, Oozie
Designed and led a team of 8 developers for VIP project – using Big Data, Hive, Java, Bash, Oozie Workflow and coordinator
Designed, trained and led development of 3 segment delivery projects - using Bitbucket, Hive, Java, Bash, Python
Redesigned the development platform – based on the whole bank’s environment architecture
Recode and re-designed continuous integration – using GIT (Bitbucket), Nexus, Jenkins, Bash scripting
Code reviewing and Code optimization consulting
Distributed system (HDFS), Cloudera data management platform, Hue, Sqoop, Hive, Oozie, CLI, Eclipse, GIT, Maven, MS-SQL
Senior Big Data designer / developer
As a consulting company, we designed, implemented and delivered a Big Data project for one of our well- known clients, AMD. In the project, we collected data of computers across the world into AWS S3 bucket and put the valid data out of it into Hadoop to generate the desired report off.
Designed the ETL through working with the business analysis – based the SRS and investigating the client’s requirements
Implemented data sanity validation and ETL from S3 to HDFS through custom NiFi processors – using AWS, NiFi, Java
Designed and implemented data ETL and customized content validation in different zones in HDFS – using Hive, Spark, Java
Created the client’s requested reports off Big data platform – using Hive
Initiated, drove and developed a framework to “unit test” all the ETL – using Bash, Python, Hive, Spark
Automating the whole ETL and report generating – using Oozie, Bash, Python
Exporting data to SQL Server from HDFS – using Sqoop, HCatalog, Hive
Optimizing data storage and accelerating ETLs – using Snappy and ORC file format
Facilitating report generation and automation to be able to be easily developed by client for any further probable needs
Code reviewing and Code optimization consulting
Technical Environment: AWS, Nifi, Hadoop file distributed system (HDFS), Hortonworks data management platform, Ambari, Sqoop, Hive, Oozie, CLI, Eclipse, Maven, MS-SQL NOSA, Iran
www.nosa.com
2002 – 2015
Technical Team Lead / Scrum Master / Developer (2010 – 2015)
Analyzed, designed, and led the development of NOSA ERP solution (CRM, Accounting, Cash Flow, Inventory, Sales, Payroll software systems) with MS-Office integrated reporting - using SOA (COM+, DCOM, MIDAS), ADO, SQL Server, Object Pascal (Delphi)
Recruit trained and built a team for next generation of CRM tools
Integrated windows app with SOAP on HTTP layer over internet – Using IIS Technical Environment: Delphi IDE, Eclipse, Maven, Multi-tier Programming, ADO, REST, spring, Hibernate, MS-SQL, MTS, COM+, DCOM, MIDAS, SOAP, XML, Client Dataset Jonah Group (Collaborating with AMD), Toronto
www.JonahGroup.com
2016 – 2017