We seek an experienced Data Scientist to join a cutting-edge AI team working on scalable, production-ready solutions.
Location : Lisbon or Porto (60% remote / 40% on-site)
Languages required : English (minimum B2)
Responsibilities :
- Develop, fine-tune, and deploy GenAI models using AWS Bedrock, SageMaker, and Lambda
- Work with LLMs, embeddings, transformers, and diffusion models for NLP and image generation
- Optimize prompt engineering, fine-tuning, and RLHF (Reinforcement Learning from Human Feedback) techniques
- Build scalable MLOps pipelines using SageMaker, ECS, and Kubernetes
- Manage large-scale datasets with AWS Glue, Athena, and Redshift
- Implement vector databases (Pinecone, Weaviate, FAISS, Amazon OpenSearch) for RAG (Retrieval-Augmented Generation) systems
- Design and optimize ETL pipelines for AI / ML workflows
- Collaborate with DevOps, Software Engineers, and product teams to deploy AI models into APIs and applications
- Ensure data privacy, compliance, and model security
- Monitor model performance and retraining needs using CloudWatch, MLFlow, and observability tools
Requirements :
Proven background in Data Science, Machine Learning, and Generative AIStrong skills in Python, SQL, and ML frameworks such as TensorFlow, PyTorch, and Hugging Face TransformersExpertise with AWS AI / ML stack : SageMaker, Bedrock, Lambda, ComprehendExperience with LLMs, embeddings, transformers, and diffusion modelsFamiliarity with RAG, vector databases, and knowledge graphsExperience in MLOps, Docker, Kubernetes, and CI / CD pipelinesUnderstanding of cloud optimization, distributed computing, and AI model scalingData engineering experience with Glue, Athena, or SparkKnowledge of NLP, image generation, or multimodal AIInterested candidates should send their CV and rate to or apply directly.