Aplikuj teraz

Data Engineer (Praca zdalna)

DataArt Poland

Kraków (+4 innych)
20000-23000 PLN / miesiąc (b2b)
B2B
apache kafka
apache spark
☁️ aws
📊 databricks
machine learning
pyspark
sql
💼 B2B
☁️ azure
📊 bigdata
ci/cd
📊 data modeling
etl
sql server
apache flink
java
🐍 python

Position overview

We are looking for a Data Engineer to help build the next generation of our cloud-based data platform using AWS and Databricks. In this role, you will design and operate scalable, resilient, high-quality data pipelines and services that empower analytics, real-time streaming, and machine learning use cases across the organisation.

Responsibilities

  • Design, build, and operate robust, scalable, secure data pipelines across batch, streaming, and real-time workloads.
  • Transform raw data into high-quality, reusable datasets and data products that power analytics and ML.
  • Work hands-on with AWS, Databricks, PySpark/Spark SQL, and modern data tooling.
  • Develop ETL/ELT processes, ingestion patterns, and streaming integrations using services such as Kafka, Kinesis, Glue, Lambda, EMR, DynamoDB, and Athena.
  • Ensure data reliability and observability through monitoring, alerting, testing, and CI/CD best practices.
  • Drive engineering best practices in performance tuning, cost optimization, security, metadata management, and data quality.
  • Partner with Data Product Owners, ML teams, and business stakeholders to translate requirements into technical solutions.
  • Lead technical design discussions, influence data platform decisions, and mentor other engineers.
  • Operate services in production with a focus on uptime, data availability, and continuous improvement.

Requirements

  • 4+ years of experience building data pipelines and large-scale ETL/ELT workflows.
  • Strong hands-on experience with AWS cloud data services and Databricks ecosystem.
  • Deep proficiency in Python, PySpark/Spark SQL, SQL optimization and performance tuning.
  • Experience with streaming architectures: Kafka, Kinesis, or similar.
  • Familiarity with CI/CD, infrastructure-as-code, automation, and DevOps practices.
  • Experience with data warehousing, structured and semi-structured data, and performance-optimized storage formats (Parquet/Delta).
  • Knowledge of Agile development and modern engineering practices.
  • Collaborative communicator who works well with engineering, product, and business teams.
  • An enabler mindset and a drive to support peers and function as one-team.
  • Able to translate business needs into scalable technical solutions.
  • A mindset of ownership, accountability, and continuous improvement.
  • Passion for data craftsmanship, clarity, and quality.

Nice to have

  • Experience with Machine Learning data pipelines, feature stores, or MLOps.
  • Familiarity with data governance, data cataloging, lineage, and metadata tools.
  • Experience with containerization and orchestration (Docker, ECS, Kubernetes, Airflow, Step Functions).
  • Knowledge of scalable data warehousing technologies.
  • Contributions to engineering communities, open-source, or internal tech groups.
Wyświetlenia: 4
Opublikowana15 minut temu
Wygasaza 27 dni
Rodzaj umowyB2B
Źródło
Logo

Podobne oferty, które mogą Cię zainteresować

Na podstawie "Data Engineer"