Platform Engineer
ITMAGINATION
⚲ Warsaw
24 360 - 26 040 PLN (B2B)
Wymagania
- Kubeflow
- Kubernetes
- Amazon EKS
- AKS
- Autoscaling
- RBAC
- Networking
- Unix
- Linux
- AWS
- AWS EC2
- AWS S3
- IAM
- CloudWatch
- Amazon RDS
- Terraform
- Git
- GitHub Actions
- Jenkins
- GitLab CI
- Python
- Shell
- Prometheus
- Grafana
- Splunk
- New Relic
- MLflow (nice to have)
- Argo (nice to have)
- Airflow (nice to have)
- Storage (nice to have)
- SQL (nice to have)
- NoSQL (nice to have)
- VPC (nice to have)
- Ingress (nice to have)
- GXP (nice to have)
- Jira (nice to have)
- JavaScript (nice to have)
- TypeScript (nice to have)
Opis stanowiska
O projekcie: We are looking for experienced team players to fill the position of Platform Engineer and participate in our up-and-coming project for our client from pharmaceutical industry. Wymagania: - M.Sc. in Computer Science/Engineering (or equivalent) or comparable industry experience. - Practical, production experience operating Kubeflow Pipelines for reproducible ML workflows at scale. - Proven experience deploying and operating workloads on Kubernetes (EKS/GKE/AKS), including upgrades, autoscaling, RBAC, networking, and reliability; strong Unix/Linux fundamentals. - Hands-on experience with AWS services (EKS, EC2, S3, IAM, CloudWatch; RDS a plus) and the ability to design secure, cost-aware architectures. - Strong Terraform skills and Git-based workflows for repeatable infrastructure provisioning and configuration management. - Practical experience with CI/CD platforms (GitHub Actions/Jenkins/GitLab CI), including artifact management, environment promotion, and progressive delivery. - Solid Python and/or shell scripting for platform automation and toil reduction. - Experience implementing logging, metrics, and tracing with SLOs, alerts, and runbooks (e.g., Prometheus, Grafana, CloudWatch, Splunk/New Relic) and a security-first mindset. - Ability to lead technical initiatives, communicate trade-offs clearly, and collaborate effectively with engineering and science teams. Nice to have: - Experience with MLflow, Feast, Argo, Airflow, Ray, and model versioning/monitoring. - Familiarity with S3/object storage, artifact registries, and handling large image datasets; basic SQL/NoSQL exposure. - Experience with digital pathology or large-scale image processing (e.g., whole-slide images) and tools like OpenSlide, scikit-image, or OpenCV. - Experience tuning high-throughput pipelines, concurrency, memory usage, and integrating GPUs/accelerators. - Experience with VPC design, ingress/egress, service meshes, secrets management, IAM, and policy as code. - Experience in regulated environments (e.g., GxP), including data governance, privacy, and building software under regulated processes. - Experience with Jira/Zendesk and with JavaScript/TypeScript for internal tools or dashboards.