Aplikuj teraz

Ingestion Data Engineer

Antal

Kraków
25200 - 33600 PLN
B2B
Hadoop
💼 B2B

Must have

  • Data engineering

  • Hadoop

  • Hive

  • HDFS

  • Spark

  • Scala

  • SQL

  • Airflow

  • Jenkins

  • GCP services

  • BigQuery

  • Dataflow

  • DataProc

  • Cloud Storage

  • Git

  • GitHub

  • English

Nice to have

  • Google cloud

  • DevOps

  • Ansible

  • Jira

Requirements description

  • 5+ years of experience in data engineering or a related field.

  • Strong hands-on experience with:

o Hadoop, Hive, HDFS, Apache Spark, Scala

o SQL and distributed data processing

o Airflow and Jenkins for workflow orchestration and CI/CD

o GCP services (BigQuery, Dataflow, DataProc, Cloud Storage, Composer)

  • Experience with data ingestion, migration, and transformation in large-scale environments.

  • Familiarity with version control tools (Git, GitHub) and Agile methodologies.

  • Strong problem-solving and debugging skills.

  • Excellent communication and collaboration abilities in global teams.

  • Nice to have: Google Cloud certification or experience with DevOps tooling (Ansible, JIRA, etc.).

Offer description

Location: Kraków, Poland (Hybrid – 2 days per week in office)

Employment type: Full-time, B2B Contract

Rate: 190–200 PLN per hour

Industry: Financial Services

At Antal, we connect top tech talent with exceptional career opportunities. For our client – a global financial institution and technology leader – we are currently looking for an Ingestion Data Engineer to join an innovative Environmental, Social & Governance (ESG) data initiative within the Data & Analytics Office.

The engineering team builds and maintains large-scale data ingestion and processing pipelines that power the bank’s ESG analytics platforms. You will work in a multidisciplinary environment, collaborating with data analysts, architects, and engineers to design robust, scalable, and secure data solutions using Apache Spark (Scala) and the Google Cloud Platform.

We Offer

• The opportunity to work on a strategic ESG data platform used globally across the organization.

• Collaboration with experienced engineers and architects in a dynamic, international environment.

• Long-term engagement with potential for professional growth and upskilling in cloud and data technologies.

• Benefits package including Multisport card and private medical care.

Your responsibilities

  1. Design, develop, and optimize data ingestion and processing pipelines using Spark (Scala) and Hadoop ecosystem tools.
  2. Orchestrate and automate workflows using Airflow and Jenkins within a CI/CD environment.
  3. Migrate and process data using Google Cloud services such as BigQuery, Dataflow, DataProc, and Composer.
  4. Collaborate with cross-functional teams to translate business logic into scalable data solutions.
  5. Ensure data quality, reliability, and performance across distributed environments.
  6. Contribute to architecture design and continuous improvement initiatives within the data platform.

show all (7)

Wyświetlenia: 4
Opublikowanadzień temu
Wygasaza 29 dni
Rodzaj umowyB2B
Źródło
Logo

Podobne oferty, które mogą Cię zainteresować

Na podstawie "Ingestion Data Engineer"