Nicolas Docolas

AI EngineerAgentic Systems & LLM Infrastructure

I build production-grade AI systems — agentic pipelines, RAG architectures, and LLM-powered backends that ship and scale.

Selected Work

Things I've built

Experience

Where I've worked

Mar 2026 — Present

TELUS Digital (Poatek)

Porto Alegre, Brazil

AI EngineerCurrent

  • Building AI microservices with Python, FastAPI, LangChain, and RAG pipelines for enterprise NLP applications.
  • Applying prompt engineering and LLM evaluation techniques to improve response accuracy and reliability in production.

Jan 2024 — Feb 2026

Hewlett-Packard Inc.

Porto Alegre, Brazil

Machine Learning Engineer

  • Worked on the design and development of large-scale enterprise AI systems, focusing on agentic architectures, multi-agent orchestration, and production-grade LLM applications. Built and maintained scalable backend services using Python, FastAPI, LangChain, Docker, and CI/CD pipelines, emphasising clean architecture, modularity, and maintainability. Designed secure access control through Role-Based Access Control (RBAC) integrated at the application's data layer.
  • Developed orchestration layers for AI systems capable of dynamically routing tasks across specialised agents. Designed execution frameworks that decompose requests into Directed Acyclic Graphs (DAGs) of interdependent tasks, supporting parallel execution, dependency resolution, and pause-and-resume capabilities. Integrated Redis for persistent state management, improving robustness, reducing latency, and optimising token usage.
  • Applied advanced LLM engineering techniques including prompt engineering, Retrieval-Augmented Generation (RAG), and guardrails such as hallucination mitigation and controlled retrieval. Built pipelines enabling context propagation across tools and agents, ensuring reliable outputs in enterprise environments.
  • Contributed to conversational AI systems with dynamic agent routing, tool-based execution, and observability through logging and metrics. Developed capabilities such as translation, summarisation, document generation, and image generation using structured prompts and API integrations.
  • Engineered on-premises AI infrastructure by migrating cloud-based systems to multi-node, multi-GPU environments. Containerised applications with Docker, orchestrated with Kubernetes. Benchmarked and optimised open-source models (Qwen, Llama, DeepSeek) using quantisation techniques to balance performance, accuracy, and efficiency while improving privacy and reducing external dependencies.

Skills

What I work with

Core AI / LLM

LangChainLangGraphRAGAgentic SystemsTool CallingPrompt EngineeringFine-TuningMulti-agent OrchestrationEvaluation PipelinesGoogle GeminiAzure OpenAIOllama

Backend

PythonFastAPISQLPydanticPyTestLocustTypeScriptReact

Data & Infra

PostgreSQLpgvectorRedisMongoDBFAISSDockerKubernetesTerraformGCPAzureAWSCI/CD

Achievements

Recognition

December 2024

HP International Case Competition — 1st Place

Won HP's inaugural international hackathon by designing and implementing a new AI feature for HP AI Companion.

June 2024

AGES Featured Project

University award from PUCRS's Experimental Software Engineering Agency for an outstanding project delivery.