Descrição do trabalho
We seek an experienced Data Scientist to join a cutting-edge AI team working on scalable, production-ready solutions.
Location: Lisbon or Porto (60% remote / 40% on-site)
Languages required: English (minimum B2)
- Responsibilities:
- Develop, fine-tune, and deploy GenAI models using AWS Bedrock, SageMaker , and Lambda
- Work with LLMs , embeddings , transformers , and diffusion models for NLP and image generation
- Optimize prompt engineering , fine-tuning , and RLHF (Reinforcement Learning from Human Feedback) techniques
- Build scalable MLOps pipelines using SageMaker, ECS, and Kubernetes
- Manage large-scale datasets with AWS Glue, Athena , and Redshift
- Implement vector databases (Pinecone, Weaviate, FAISS, Amazon OpenSearch) for RAG (Retrieval-Augmented Generation) systems
- Design and optimize ETL pipelines for AI/ML workflows
- Collaborate with DevOps, Software Engineers , and product teams to deploy AI models into APIs and applications
- Ensure data privacy, compliance , and model security
- Monitor model performance and retraining needs using CloudWatch, MLFlow , and observability tools
- Requirements:
- Proven background in Data Science , Machine Learning , and Generative AI
- Strong skills in Python , SQL , and ML frameworks such as TensorFlow , PyTorch , and Hugging Face Transformers
- Expertise with AWS AI/ML stack : SageMaker, Bedrock, Lambda, Comprehend
- Experience with LLMs, embeddings, transformers, and diffusion models
- Familiarity with RAG , vector databases, and knowledge graphs
- Experience in MLOps , Docker , Kubernetes , and CI/CD pipelines
- Understanding of cloud optimization , distributed computing , and AI model scaling
- Data engineering experience with Glue , Athena , or Spark
- Knowledge of NLP , image generation , or multimodal AI
Interested candidates should send their CV and rate to or apply directly.