Post Job Free
Sign in

Data Engineer (Databricks)

Company:
Coherentsolutions
Location:
Moldova
Posted:
February 26, 2026
Apply

Description:

Project Description

We are looking for an experienced Data Engineer with deep expertise in Databricks to join our advanced analytics and data engineering team. The ideal candidate will play a key role in designing, building, and optimizing large-scale data solutions on the Databricks platform, supporting business intelligence, advanced analytics, and machine learning initiatives. You will collaborate with cross-functional teams to deliver robust, scalable, and high-performance data pipelines and architectures.

Technologies

Databricks (including Spark, Delta Lake, MLflow)

Python/Scala

SQL

ETL concepts

Distributed data processing

Data warehousing

Cloud Platforms & Storage

What You'll Do

Lead the design, development, and deployment of scalable data pipelines and ETL processes using Databricks (Spark, Delta Lake, MLflow);

Architect and implement data lakehouse solutions, ensuring data quality, governance, and security;

Optimize data workflows for performance and cost efficiency on Databricks and cloud platforms (Azure, AWS, or GCP);

Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver actionable insights;

Mentor and guide junior engineers, promoting best practices in data engineering and Databricks usage;

Develop and maintain documentation, data models, and technical standards;

Monitor, troubleshoot, and resolve issues in production data pipelines and environments;

Stay current with emerging trends and technologies in data engineering and Databricks ecosystem;

Job Requirements

Bachelor’s or Master’s degree in Computer Science, Engineering, Mathematics, or a related field;

5+ years of experience in data engineering, with at least 2 years of hands-on experience with Databricks (including Spark, Delta Lake, and MLflow);

Strong proficiency in Python and/or Scala for data processing;

Deep understanding of distributed data processing, data warehousing, and ETL concepts;

Experience with cloud data platforms (Azure Data Lake, AWS S3, or Google Cloud Storage);

Solid knowledge of SQL and experience with large-scale relational and NoSQL databases;

Familiarity with CI/CD, DevOps, and infrastructure-as-code practices for data engineering;

Experience with data governance, security, and compliance in cloud environments;

Excellent problem-solving, communication, and leadership skills;

English: Upper Intermediate level or higher;

What Do We Offer

The global benefits package includes:

Technical and non-technical training for professional and personal growth;

Internal conferences and meetups to learn from industry experts;

Support and mentorship from an experienced employee to help you professional grow and development;

Internal startup incubator;

Health insurance;

English courses;

Sports activities to promote a healthy lifestyle;

Flexible work options, including remote and hybrid opportunities;

Referral program for bringing in new talent;

Work anniversary program and additional vacation days.

Sep 25, 2025 12:50

Apply