Nguyễn Văn Dũng
Data Engineer Intern
****************@*****.*** https://github.com/NguyenVanDung1806 159 Xa Lộ Ha Noi, Thao Dien, Quan 2, TP. Thu Duc
SUMMARY
I am sekking a data engineering internship to lauch my journey in the word of data with enthusiasm and hunger to learn . Eager to soak up new experiences and contribute positively to projects, all while building a foundation for a bright data career ahead . Beside the technological knowledge at school, I am also confident in my soft skill when taking on the management for a coffee house and my our small business
Master basic knowledge of computer science, data engineer (data warehouse, big data, data pipeline, ETL, ELT)
Want to become Intern / Fresher, after 5 years reach Senior level EXPERIENCE
Coffee Shop Manager
Nha Ban Coffee House
2023 - Present 790 My Phuoc Tan Van, Phu My,Thu Dau Mot, Binh Duong Coffee in Thu Dau Mot, Binh Duong is very Architecture for weekends and weekdays, cool space, super delicious drinks, national prices Manage daily operations of your coffee shop
Build reports and charts to get an overview of your business situation Manage staff and optimize work situation
Hire and onboard new waiters and bartenders
Coordinate with suppliers and order supplies as needed Train baristas to prepare drinks and use coffee brewing equipment properly EDUCATION
Computer Science
INDUSTRIAL UNIVERSITY OF HO CHI MINH CITY
2021 - Present 12 Nguyen Van Bao, Ward 4, Go Vap, Ho Chi Minh SKILLS
Python SQL C++ Pandas Numpy PySpark
MySQL MongoDB PostgreSQL Cassandra
Airflow Docker PySpark OpenCV
AWS Glue Amazon S3 Snowflake Amazon Redshift
LANGUAGES
English
Native
STRENGTHS
Analytical skills
Strong in problem analysis, independent work,
able to withstand high pressure
Comunication Skill
Good communication skills thanks to experience
in human resource management, always ready to
change to suit the environment
Problem-solving
Problem solving ability, always willing to handle
difficult problems to increase work experience
PROJECTS
Real Time Data UserAPI
2024 - Present Thao Dien, Quan 2
Building an end-to-end data engineering pipeline.
Covers each stage from data ingestion to processing and finally to storage, utilizing a robust tech stack that includes Apache Airflow, Python, Apache Kafka, Apache Zookeeper, Apache Spark, and Cassandra
Data Source: We use randomuser.me API to generate
random user data for our pipeline.
Apache Airflow: Responsible for orchestrating the
pipeline and storing fetched data in a PostgreSQL
database.
Apache Kafka and Zookeeper: Used for streaming
data from PostgreSQL to the processing engine.
Control Center and Schema Registry: Helps in
monitoring and schema management of our Kafka
streams.
Apache Spark: For data processing with its master
and worker nodes.
Cassandra: Where the processed data will be stored. Data Pipeline ( Airflow, Spark )
2024 - Present Thao Dien, Quan 2
In this work, user want to have understand behaviour of user went they integrate with application. Application export big data with stream data.
Building stream application to ingress and egress data of both user and application data.
Postgres: stores Airflow metadata and Testing data. Airflow Webserver: Airflow webserver and Scheduler. Spark: Spark Master.
Spark-worker-N: Spark workers.
Jupyter-Spark: has pyspark for interactive
development.
www.enhancv.com Powered by
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
•
q
CERTIFICATION
IBM Data Engineer
Certified in Data Data Engineering
from IBM focusing on advanced SQL
concepts and data manipulation.
www.enhancv.com Powered by