R&D Engineer
Free Pro • Work-study program
Apr 2023 - Present•Marseille, France
- Design and deploy LLM workflows that automate customer care, cutting workload by 5–10%.
- Build agentic AI pipelines with tool-calling, retrieval, and safety enforcement with langchain and langgraph.
- Deploy and optimize open-source LLMs on specialized GPU clusters (NVIDIA H100, L40S) with vLLM and LlamaCPP.
- Develop full-stack platforms (React, FastAPI, microservices) and manage data infrastructure (Postgres, MongoDB, Qdrant).
- Operate CI/CD pipelines (Docker, GitLab) and administer Linux/HPC environments.
- Pioneered deployment of AI workloads on Scaleway’s first H100 GPU cluster in collaboration with Nvidia.
- Engineered Raspberry Pi–based OCR and object detection systems with DepthAI and YOLOv8 for live screen monitoring.