Senior/Principle Data Engineer with strong MLOps expertise
Intellias
⚲ Poland (Remote)
Wymagania
- Azure
- Machine Learning
- Python
Opis stanowiska
Drivers of change, it’s your time to pave new ways. Intellias, a leading software provider in the automotive industry, invites you to shape the future of driving. Join the team and co-create digital products for the world’s top-tier brands. Our customer is a Dutch multinational developer and creator of location technology and consumer electronics headquartered in Amsterdam. They are looking for a skilled Data Engineer with strong MLOps experience to lead the design and automation of production-grade data and ML pipelines on Azure Databricks. You’ll work at the intersection of data engineering, infrastructure automation, and machine learning — transforming prototype workflows into robust, observable, and cost-efficient pipelines. Responsibilities • Design and maintain end-to-end data and ML pipelines using Databricks Workflows, Delta Lake, and Unity Catalog (bronze–silver–gold layers, schema evolution, access policies). • Build reproducible training and deployment workflows integrated with tools for experiment tracking, model registry, and artifact management. • Implement data quality frameworks and observability metrics aligned with industry best practices. • Build and monitor dashboards (e.g. past experience with Lakeview, Grafana, or similar) for data quality, model performance, and operational metrics. • Automate data ingestion and feature generation jobs, leveraging PySpark, SQL Warehouses, and Databricks Asset Bundles (DAB) under CI/CD (GitHub Actions or Azure DevOps). • Manage access and security, ensuring compliance and reliability. • Optimize compute performance and cost (spot/autoscaling, cluster tuning, caching, partitioning). Qualifications • Strong proficiency in Python is essential, along with experience in shell scripting and potentially other languages like Java. • Hands-on experience with at least one major cloud service provider (Azure is preferable) • Experience in Azure Databricks Workflows, Delta Lake, and Unity Catalog • Experience in PySpark, SQL Warehouses, and Databricks Asset Bundles (DAB) under CI/CD (GitHub Actions or Azure DevOps). • Strong software engineering practices, including testing, code optimization, and design patterns. • Excellent communication and collaboration skills to bridge the gap between technical and non-technical teams. Perks and Benefits: • Flexible work schedule. • Fixed financial bonus issued upfront on a quarterly basis, covering the average market price of private medical care and sport card - B2B contract. • Present on the occasion of birthday, wedding, child birth. • E-learning accounts for Coursera, O'Relly, Udemy. • Corporate language school.