NoFluffJobs Praca zdalna Senior New

Deep Learning Senior Engineer

ITMAGINATION

⚲ Remote

32 760 - 36 120 PLN (B2B)

Wymagania

  • Deep learning
  • AI
  • NLP
  • MLOps
  • Docker
  • Kubernetes
  • Machine learning
  • Python
  • NumPy
  • pandas
  • PyTorch
  • TensorFlow
  • GCP
  • Cloud
  • Storage
  • REST API
  • GraphQL
  • Microservices
  • GDPR
  • Jira
  • Degree

Opis stanowiska

O projekcie: At Virtusa ( Former ITMAGINATION), every innovator has the potential to transform and lead in a digital world—but unlocking that potential takes more than technology; it takes a trusted partner who combines engineering excellence, creativity, and an AI-first mindset. Together, we co-create solutions that help businesses grow faster, operate smarter, and make experiences better with technology. We are seeking a highly technical Deep Learning Senior Engineer (T3) to join our AI delivery hub in Poland. In this role, you will be a primary architect and builder of advanced Generative AI solutions, moving past basic wrappers to design sophisticated Deep Learning architectures. You will specialize in Transformer-based models, RAG systems, and LLM orchestration, specifically leveraging the Google Gemini ecosystem on Vertex AI. This is a high-impact role requiring a blend of scientific rigor and production-grade engineering to deliver state-of-the-art AI applications for our global enterprise clients. Wymagania: - 6–8 years of experience in Software Engineering or Machine Learning, with a minimum of 3 years focused on Deep Learning and NLP.   - Expert-level Python skills, including deep proficiency with scientific and AI libraries (NumPy, Pandas, PyTorch, or TensorFlow).   - Strong theoretical and practical understanding of Transformers, attention mechanisms, and semantic embeddings.   - Proven track record of building production-ready applications with LangChain, LlamaIndex, and LLM APIs (OpenAI, Anthropic, or Vertex AI).   - Hands-on experience with FAISS, Pinecone, or Weaviate, including indexing strategies, metadata filtering, and hybrid search optimization.   - Advanced experience with GCP, specifically Vertex AI (Model Garden, Pipelines, and Notebooks) and Cloud Storage.   - Deep understanding of NLP concepts such as tokenization, named entity recognition (NER), and semantic search logic.   - Experience using RAGAS or similar tools to quantify model performance (precision, recall, faithfulness).   - Practical knowledge of Docker and Kubernetes; familiarity with CI/CD for ML models and automated deployment workflows.   - Experience building scalable REST/GraphQL APIs and microservices for AI-driven applications.   - Understanding of Responsible AI practices, including data privacy (GDPR), PII masking, and bias detection in LLM outputs.   - Strong analytical problem-solving skills and the ability to work in an Agile (Jira) environment.   - Professional English (C1) is mandatory for collaboration with our international technical leadership.   - Bachelor’s or Master’s degree in Computer Science, AI, Mathematics, or a related quantitative field. Codzienne zadania: - Model Architecture & Design: Design and implement high-performance Generative AI applications utilizing Transformers, Diffusion models, and advanced NLP techniques. - LLM Orchestration: Build and manage complex, agent-based workflows using frameworks like LangChain and LlamaIndex to automate multi-step reasoning tasks. - Advanced RAG Systems: Architect end-to-end Retrieval-Augmented Generation (RAG) pipelines, integrating enterprise data with Vector Databases (Pinecone, FAISS, Weaviate) while ensuring high semantic relevance. - Google GenAI Mastery: Lead the implementation of Google Gemini models within the Vertex AI platform, optimizing for latency, throughput, and cost. - Fine-tuning & Optimization: Perform model fine-tuning, quantization, and embedding optimization to tailor LLMs to specific domain requirements and enterprise datasets. - Prompt Engineering & Evaluation: Design sophisticated prompt strategies and implement rigorous evaluation frameworks (e.g., RAGAS) to track model accuracy, hallucination rates, and drift. - Deployment & Scaling: Collaborate with MLOps teams to deploy models into production environments using Docker and Kubernetes, ensuring scalability and fault tolerance.