Adam Ghaith
Cell:646-***-****
**********@*****.***
US CITIZEN
Professional Summary:
●Around 5 years of IT industry experience in Data-Engineer from gathering requirements to Developing, Testing, Implementation and support in Data Extraction, Transforming and Loading (ETL) and Data migration
●Extensive experience on major components in Hadoop Ecosystem like Hadoop, Map Reduce, HDFS, Spark, HIVE, Python, Parquet and Avro
●1+ years of experience in Databricks and Snowflake
●Worked with large datasets using Partitions, Spark in Memory capabilities, Broadcasts in Spark, Efficient Joins
and Transformations.
●Implemented Partitioning, Dynamic Partition, Buckets, Cubes, Windowing & Analytics functions in HIVE
●Experience in Benchmarking Cluster & SQL optimizations and setting the right Spark configuration parameters to the job and obtaining the best performance out of it.
●Extensively used AWS EC2, EMR, S3 AND Athena
●Extensive experience with SQL extensively by using Joins, Sub Queries, Set Operations and Advanced OLAP functions.
●Good knowledge in Data Warehousing concepts and Dimensional modeling like Star Schema and Snowflake schema.
●Having good communication skills, Interpersonal relations, hardworking and result- oriented as an Individual and in Team.
Education
BA Degree of business management 2011
Technical Skills:
TECHNOLOGY
TOOL
Database
HDFS, Hive, Spark, MYSQL
Operating Systems
Linux, Windows,,Git,VS,VSC
Programming
Languages
python, SQL,
AWS Stack
EMR, S3, Athena,ETL(GLUE)
Lambda Function, Terraform
Professional Experience:
●Data Engineer, AIG Jersey City NJ
●June 2021–May 2022
Responsibilities
●Responsible for building scalable distributed data solutions with Apache Spark using Python/Scala.
●Developed aggregated & bi tables in AWS EMR by using hive, spark and load them into Data Lake and Athena.
●Developing and scheduling the ETL pipelines by using Databricks.
●Implemented Partitioning, Dynamic Partition, Buckets in HIVE.
●Load and transform large sets of structured, semi-structured data.
●Developed Airflow workflows for orchestration
●Developed the Python automation framework & scripts for Data integrity and validations
●Strong problem-solving experience on test, identify, address, debug and then resolve technical issues that affect the integrity of the application.
Deutsche Bank, Manhattan, New York May 2020 to May 2021
Responsibilities:
●Responsible for building scalable distributed data solutions with Apache Spark using Python.
●Developed aggregated & bi tables in AWS EMR by using hive, spark and load them into Data Lake and Athena.
●Developed Airflow workflows for orchestration
●Developing and scheduling the ETL pipelines by using Databricks, snowflake. ·
●Implemented snowflake cache design for building real-time aggregates in snowflake
●Load and transform large sets of structured, semi-structured data.
●Developed the Python automation framework & scripts for Data integrity and validations
●Strong problem-solving experience on test, identify, address, debug and then resolve technical issues that affect the integrity of the application.
Data Analyst
Credit Agricole, New York, New York February 2015 to March 2016
Responsibilities:
Created customized Table Calculations, Conditions and Filters for various analytical reports and dashboards.
Created and modified Interactive Dashboards and Creating guided navigation links within Interactive Dashboards.
Created navigation page which had all the links to various dashboards.
Scheduled frequency and time to refresh data for the when sources are published or extracted to server.
Used SQL tools like Toad, Pl/SQL to run SQL queries and validate the data in the warehouse and data marts.
Used Performance recording on Desktop and on Tableau Server and Tuned the Performance of Dashboards.
Followed Best practices while developing Dashboards
Provided UAT assistance and post implementation production support.
Used Dual Axes Charts-Shown multiple measures with dollar numbers on left Y axis and Percentages on the right Y axis
Created Tableau Server backups using Tableau admin utility
Set up Users, Groups, Permissions, Schedules on Tableau Server
Done unit testing of Dashboards for data quality and done Interface testing like Quick filters, parameters, Actions.