Senior Data Engineer

Job openings     Senior Data Engineer

Senior Data Engineer

At DAC.digital, we are constantly growing our business. As part of our growth strategy, we are strengthening our strategic partnership with a company specializing in crafting data models, dashboards, and data activation..

Key information:

Salary: 28 000 - 31 000 PLN net/month - pure B2B contract

24 500 – 27 000  PLN net/month – B2B contract (days off included)

Requirements


It is vital that you have:

  • 4+ years hands-on data engineering / platform experience;

  • strong experience with Spark/PySpark, Kafka or other streaming platforms, cloud-native data lakes or lakehouses (Iceberg, Delta Lake, Hudi, with Iceberg as a plus), dbt or similar transformation frameworks, and advanced SQL including performance tuning;

  • solid Python skills for data pipelines and integration work;

  • familiarity with NLP / LLM-driven text processing workflows (embedding models, classification, enrichment);

  • experience with BI / query engines such as Dremio, Trino, Presto, or BigQuery plus a viz tool (Metabase, Looker, Power BI, etc.);

  • experience deploying data platforms in cloud and/or on-prem environments - Docker, Kubernetes (or equivalent orchestration), CI/CD for data pipelines;

  • understanding of security, networking basics and how to operate inside restricted enterprise environments (no direct internet access, firewall rules, etc.);

  • knowledge of English (min. B2);

  • high communication skills;

  • eager to learn and share knowledge.


Nice to have:

  • experience with marketing / growth analytics, CDP, or telco data;

  • background in ML/AI beyond pipelines (e.g. recommendation systems, uplift modelling);

  • prior experience in consulting / client-facing roles;

  • some front-end exposure (Metabase customisation, simple React dashboards) if we want one person to bridge.


Technology stack:

  • Python

  • SQL

  • Apache Spark / PySpark

  • Kafka

  • Iceberg

  • Nessie

  • MinIO

  • Kubernetes

  • Helm Charts

Responsibilities


You will be responsible for supporting our team in:

  • designing, implementing, and maintaining streaming and batch data pipelines using Kafka, Spark/PySpark, Apache Iceberg, and dbt;

  • managing schema evolution, partitioning, performance, and data quality in the Iceberg lakehouse;

  • building and maintaining NLP and enrichment flows;

  • integrating with LangChain and LangGraph-based components to resolve senders to companies, enrich data from public internet sources, and classify companies into brand categories and verticals such as F&B, footwear, and online brokers;

  • supporting experimentation with ML and NLP components, including monitoring and logging;

  • working with Dremio and Metabase to ensure performant queries and dashboards and expose data in a usable ways;

  • handling deployment and operations in constrained environments, including on-prem or behind-firewall deployments at MNOs, containerisation with Docker and Kubernetes, and logging and lightweight monitoring for services and pipelines;

  • ensuring stability, observability, and reproducibility of the platform.

Offer


What do we offer:

  • possibility to work 100% remotely or on-site at our office in Gdańsk;

  • b2b contract with included 30 days off;

  • private Medical care;

  • group insurance;

  • pre-paid card or Sport Card;

  • referral program;

  • real Agile practices;

  • employee well-being online platform;

  • cafeteria benefits.



Share this Job

   
Recruitment powered by tomHRM