Dinh Duc Nguyen Vu - Data Engineer
Phone: +84-905****** / Email: *********@*****.***
LinkedIn:linkedin.com/in/nguyenvu202
SUMMARY
I am currently a final year student at HCM University of Technology Education. My major is Data Engineering, and I am currently seeking a position as a Data Engineer. With the knowledge i gained from university about data, big data and various Apache frameworks and skills in designing databases and data warehouses I believe that I can help you in designing and managing database and data warehouse.Additionally, after completing an internship at AWS,where I gained experience with cloud services, I am eager to apply this expertise to help scale systems effectively. TECHNICAL SKILLS
Programming Language: Python, C#
Framework: Apache Hadoop, Apache Spark, Airflow
Database management Systems: MSSQL, PostgreSQL,Cassandra, MongoDB Foreign Language: TOEIC 800+
Others:
Good understanding of Cloud Computing Platform AWS Experience in using Docker
Knowledge of Real-time Data Streaming with Kafka
EXPERIENCE
Amazon Web Service(AWS) Vietnam - Cloud Engineer (Internship) June 2024 - September 2024 (4 months)
Gained comprehensive knowledge of AWS cloud services and their applications.
Participated in hands-on workshops guided by experienced mentors, working with a variety of AWS tools and technologies. Conducted workshops and practical projects using AWS services. Link: https://vund202.github.io/
PERSONAL PROJECT
Building Data Warehouse for E-Commerce: [Link]
Designed and implemented a data warehouse and database architecture to handle CSV datasets containing over 100K rows.
Do ETL (Extract, Transform, Load) processes to cleanse and load data into the warehouse efficiently.
Created insightful data visualizations using PowerBI to support decision-making.
Tech stack: MSSQL, SSIS, PowerBI
EDUCATION
Ho Chi Minh University of Technology Education (September 2020 - September 2025) Bachelor of Engineering in Data Engineering (GPA 3.0/4.0) Log Analytíc with Spark Streaming and Kafka: [Link] Analyzed log data from website URLs to ensure scalability and efficiency in handling large volumes of streaming data. From the given website, data is downloaded containing the NASA access log data in csv format, containing different components of a web server log
The extraction process is done using NiFi and Kafka, by streaming data from log file using NiFi and creating topics, publishing logs using Kafka.
In the transformation and load process, the schema is defined and data is read from Kafka as a streaming Dataframe, storing in Cassandra Finally, data is visualized using different plots in a Realtime, Hourly and Daily manner using Plotly and Dash.
Tech stack: Docker, Ubuntu, Jupyter Lab, Spark Streaming, Nifi, Kafka,Cassandra,Plotly, Dash
CERTIFICATE
IBM Data Engineer
AWS Academy Data Engineering
SIC Big Data