JustJoin.IT Praca zdalna Senior New

5656 2 x Mid+/Senior Flink Data Engineer

emagine Polska

⚲ Warsaw

Wymagania

  • Apache Kafka
  • Cloud
  • CI/CD
  • Governance
  • Artificial Intelligence (AI)
  • Cloud Storage
  • Google Cloud
  • Agile
  • API (Application Programming Interface)
  • Microsoft Platform

Opis stanowiska

Remote from Poland Budget: 165PLN/h + VAT Our company is a global digital platform operating at massive scale, serving hundreds of millions of users worldwide. We focus on building consumer-facing subscription products that connect users with content creators through personalized, data-driven experiences. Our culture emphasizes collaboration, experimentation, and engineering excellence in a highly distributed, cloud-native environment. Project Overview The project centers on the continuous development and modernization of a large-scale streaming data platform. The main objective is to evolve existing Apache Flink-based streaming pipelines while executing complex data migrations and platform upgrades. Key responsibilities include: • Developing and enhancing real-time streaming pipelines using Apache Flink • Migrating existing Flink jobs using the DataStream API and adapting them to newer platform standards • Leading and executing the upgrade of the Flink platform to version 2.0 • Designing, optimizing, and maintaining high-throughput, fault-tolerant streaming architectures • Migrating large-scale datasets from BigQuery (BQ) to Data Cloud Storage (DCS) • Scaling and automating ongoing data migration processes to support growing data volumes • Converting datasets from Avro to Parquet format with emphasis on performance, schema evolution, and storage optimization • Leveraging AI-powered tools to accelerate migration, validation, and transformation workflows • Ensuring data quality, integrity, and minimal downtime during migrations • Collaborating with cross-functional teams and effectively communicating technical concepts to non-technical stakeholders Key Requirements • Strong hands-on experience with Apache Flink, especially development using the DataStream API • Proven experience maintaining and upgrading Flink environments, ideally with exposure to Flink 2.0 • Deep understanding of streaming pipeline architecture, performance tuning, state management, and fault tolerance • Experience migrating large-scale datasets from BigQuery (BQ) to Data Cloud Storage (DCS) • Strong proficiency in data format conversion, particularly Avro to Parquet • Ability to design, scale, and automate migration workflows while ensuring data integrity and minimal service disruption • Solid knowledge of Google Cloud Platform (GCP) and its data services • Good understanding of distributed systems, schema evolution, and storage optimization strategies • Ability to break down complex migration and platform challenges into clear, actionable steps • Proactive mindset with strong ownership of solutions and risk identification • Clear and effective communication skills, especially in explaining technical topics to non-technical stakeholders Nice to Have • Interest in and familiarity with emerging AI-driven practices, with a willingness to explore and experiment beyond standard approaches. • Experience working on high-scale, consumer-facing data platforms • Background in long-running migration programs involving multiple data sources and formats • Familiarity with observability, monitoring, and alerting for streaming systems • Practical experience using AI-powered assistants to improve productivity, quality, or decision-making in software delivery.