Descrição do trabalho

We seek an experienced Data Scientist to join a cutting-edge AI team working on scalable, production-ready solutions.

Location: Lisbon or Porto (60% remote / 40% on-site)

Languages required: English (minimum B2)

  • Responsibilities:
  • Develop, fine-tune, and deploy GenAI models using AWS Bedrock, SageMaker , and Lambda
  • Work with LLMs , embeddings , transformers , and diffusion models for NLP and image generation
  • Optimize prompt engineering , fine-tuning , and RLHF (Reinforcement Learning from Human Feedback) techniques
  • Build scalable MLOps pipelines using SageMaker, ECS, and Kubernetes
  • Manage large-scale datasets with AWS Glue, Athena , and Redshift
  • Implement vector databases (Pinecone, Weaviate, FAISS, Amazon OpenSearch) for RAG (Retrieval-Augmented Generation) systems
  • Design and optimize ETL pipelines for AI/ML workflows
  • Collaborate with DevOps, Software Engineers , and product teams to deploy AI models into APIs and applications
  • Ensure data privacy, compliance , and model security
  • Monitor model performance and retraining needs using CloudWatch, MLFlow , and observability tools
  • Requirements:
  • Proven background in Data Science , Machine Learning , and Generative AI
  • Strong skills in Python , SQL , and ML frameworks such as TensorFlow , PyTorch , and Hugging Face Transformers
  • Expertise with AWS AI/ML stack : SageMaker, Bedrock, Lambda, Comprehend
  • Experience with LLMs, embeddings, transformers, and diffusion models
  • Familiarity with RAG , vector databases, and knowledge graphs
  • Experience in MLOps , Docker , Kubernetes , and CI/CD pipelines
  • Understanding of cloud optimization , distributed computing , and AI model scaling
  • Data engineering experience with Glue , Athena , or Spark
  • Knowledge of NLP , image generation , or multimodal AI

Interested candidates should send their CV and rate to or apply directly.