Head of Data Engineering
Autopay S.A.
⚲ Sopot
Wymagania
- PySpark
- Databricks
- Google Cloud
- Spark
- AI
Opis stanowiska
About the company Autopay Global is the newest member of the Autopay family, aiming to expand the reach of the group’s state-of-the-art payment integration and payment data technologies to the international market, providing seamless integration with local PSPs, support for multiple currencies and compliance with local frameworks. We have a very forward-looking approach to our products, we value creativity, passion and drive to leverage the newest achievements in technology to our advantage. To support our dynamic expansion, we are looking for a new Head of Data Engineering for a full-time, hybrid work in Warsaw or Gdańsk. About the role The Head of Data Engineering owns the end-to-end data architecture and execution, with hands-on depth in PySpark and Databricks and strong experience building AI-ready data foundations on Google Cloud Storage (GCS) and Google Vertex AI. You will be responsible for delivering a secure, scalable, and low-latency data lakehouse and feature platform that enables Autopay’s AI core (agents, RAG, ranking, decisioning) and activation systems to run on reliable, high-quality, well-governed data across batch and streaming. You will also be responsible for hiring, leading and mentoring a team of high-performing data engineering proffessionals. • Define the lakehouse reference architecture on GCS with Databricks/Delta Lake, • build and operate PySpark pipelines in Databricks for both streaming and batch workloads, • implement streaming ingestion, • own the Customer 360 / CDP layer: unify events, transactions, and user identifiers, • deliver a real-time feature layer (feature store) that publishes segments, scores, and vectors, • create and maintain embeddings and retrieval indexes to power RAG in Autopay AI Core (chunking strategies, metadata, refresh policies, and retrieval evaluation, • establish data governance with Dataplex/Data Catalog and/or Unity Catalog, • own data observability for pipelines: freshness, completeness, schema drift, anomaly detection, and automated remediation workflows. What tools will you be working with? • Technology: PySpark, Databricks, Google Cloud Storage, Google Vertex AI, Delta Lake • Nice to have: Experience with identity resolution inputs, experience building near-real-time segmentation, CLV, and propensity scoring pipelines, familiarity with vector databases and multi-cloud data movement patterns. Requirements and skills we are looking for in a person hired for this role: • 10+ years in data engineering and still hands-on to build ground up forming a team; 3-5+ years leading data platform teams with ownership of production data SLAs, • deep hands-on expertise with PySpark and Spark performance tuning (shuffle optimization, partitioning, checkpointing, incremental loads), • strong experience with Databricks (jobs/workflows, Delta Lake, governance) and building lakehouse architectures on GCS, • proven delivery of streaming + batch data platforms that power real-time product experiences (not just analytics), • experience building feature stores and ML-ready datasets with point-in-time correctness and strong governance, • strong grasp of privacy and compliance in data systems: PII handling, consent, and auditability, • Google Vertex AI experience: building data pipelines that feed training, evaluation, and inference workflows; understanding of dataset/version management, • hands-on experience supporting RAG systems: document ingestion, chunking, embedding generation, retrieval evaluation, and index refresh strategies, • experience with retrieval-aware training approaches (e.g., retrieval augmented fine-tuning / RAFT) and producing high-quality supervised datasets with provenance, • ability to collaborate with AI Engineers on MCP-based tools and agent workflows (tool schemas, rate limits, caching, and audit logs). What we offer • a lidership role in fast-growing, global fintech company, • possibility to work with cutting-edge tools and technologies, • independence in decision-making, • friendly working environment, team support, no dress code. Join us and let's head together where no one has gone before!