Junior Data Engineer
PROEXE
⚲ Warszawa
9 000 - 15 500 PLN (B2B)
Wymagania
- SQL
- BigQuery
- dbt
- Python
- Airflow
- Cloud Composer
- Git
- GCP
- Terraform (nice to have)
- IaC (nice to have)
- GCP (nice to have)
Opis stanowiska
O projekcie: For one of our projects in the car rental industry, we are looking for an experienced GCP Data Engineer with a strong focus on BigQuery. In this role, you will be responsible for building and optimizing data pipelines within the Google Cloud ecosystem as part of an international team. We offer: - Flexible working hours - Paid holidays - Benefit package including a sports card, private health insurance, and private healthcare. - B2B contract - 100% remote - International team Wymagania: - 1+ years of experience in a data engineering or analytics engineering role - Strong SQL skills - Hands-on experience with BigQuery or a comparable cloud data warehouse - Experience with dbt (Core or Cloud) - building models, writing tests, using macros, incremental materializations - Working knowledge of Python for data engineering tasks (scripting, API integrations, light transformations) - Experience with at least one orchestration/scheduling tool (Airflow, Cloud Composer, Dagster, Prefect, or similar) - Familiarity with data ingestion patterns - batch loading, CDC, API extraction - Comfortable with Git workflows (branching, PRs, code review) - Awareness of PII handling and basic GDPR compliance practices - Fluent in Polish and English Nice to have: - Experience with Terraform or other IaC tools for managing GCP resources - Background in the mobility, automotive, or car rental domain Codzienne zadania: - Build and maintain ETL/ELT pipelines that ingest data from various sources into BigQuery - Develop and maintain dbt models - Monitor pipeline health, troubleshoot failures, and implement alerting - Collaborate with backend engineers to design data contracts and integrate new data sources - Write and maintain data quality tests and validation checks - Optimize BigQuery performance (partitioning, clustering, query tuning, cost control) - Maintain documentation: data dictionaries, pipeline runbooks, model descriptions - Participate in data modeling discussions and propose improvements to existing structures - Support analysts and business stakeholders with data access and ad-hoc queries