0
Initializing neural network...
Machine Learning Engineer · Bogotá, Colombia

MANUEL
ALEJANDRO
DIAZ RUBIANO

Building end-to-end AI systems — from LLMs and RAG architectures to MLOps pipelines and production deployment.

Pipeline ready · inference Deployed
Latency p95
LLM Accuracy
Faithfulness
Throughput
CPU Usage
Uptime
99.8%
Hallucination
Scroll
About me
Manuel Alejandro Diaz Rubiano
Manuel Alejandro Diaz Rubiano
AI Engineer & ML Specialist
Background

How I got here

I started my journey in statistics and mathematics, but quickly fell in love with the intersection of data and artificial intelligence. During my studies at Universidad Santo Tomás, I developed my thesis on topic analysis using the LDA model applied to the Colombian case.

I've worked across finance, customer service, and enterprise technology — moving from dashboards and KPI analysis to building production AI systems with LLMs, RAG, and multi-agent orchestration. Each role gave me a different perspective on how AI can create real impact.

Today, as an AI Engineer at H&Co Latam, I design end-to-end GenAIOps pipelines: from architecture to deployment, observability, and continuous improvement. My goal is to build reliable AI systems that don't just work in demos — they work in production.

Expertise & Stacks

My expertise

Full-cycle AI engineering — from raw data and model training to production deployment, observability, and continuous improvement. I choose tools based on the problem, not the trend.

Stack
LLMs & GenAI

Working daily with frontier models — from prompt design and structured outputs to fine-tuning and guardrails for production safety.

Claude OpenAI Mistral Llama Bedrock Guardrails Prompt Engineering Structured Outputs Fine-tuning · LoRA Pydantic Function Calling
Stack
Agents & Orchestration

Building multi-agent systems with conditional routing, parallel tool execution, and stateful workflows — from prototype to production.

LangGraph LangChain LlamaIndex MCP AWS Step Functions Bedrock Agents FastAPI
Stack
Vector & Retrieval

Designing hybrid RAG pipelines with semantic and lexical fusion — pgvector, dense embeddings, and reranking at scale.

Pinecone OpenSearch pgvector · HNSW Qdrant Chroma FAISS Redis BM25 · RRF Cohere Rerank sentence-transformers
Stack
Observability & Eval

Tracing every agent step, evaluating output quality, and closing the feedback loop — AI systems must be measurable to be trustworthy.

LangSmith Langfuse Arize Phoenix MLflow W&B CloudWatch
Stack
Cloud

Multi-cloud fluency across AWS, Azure, and GCP — with deep hands-on experience deploying managed AI services and serverless pipelines.

AWS Azure GCP AWS Bedrock GCP Vertex AI Azure AI SageMaker Lambda CDK GitHub Actions vLLM
Stack
Containers & IaC

Infrastructure as code from day one — containerized workloads, Kubernetes orchestration, and reproducible deployments across environments.

Docker Kubernetes Terraform CloudFormation Canary Releases CI/CD
Stack
ML & Deep Learning

Strong foundations in classical ML, NLP, and computer vision — from statistical modeling to training custom neural architectures.

PyTorch HuggingFace BERT · T5 XGBoost Scikit-learn U-Net Prophet ARIMA · LSTM TorchServe
Stack
Languages & Data

Python-first, but fluent across the data layer — SQL, R for analysis, TypeScript for APIs, and modern data tools for clean pipelines.

Python SQL R TypeScript Supabase PostgreSQL dbt
Values

What drives me

  • 01
    Impact-driven AI

    Every architecture I design has production reliability and measurable business impact as its north star — not hype.

  • 02
    Model quality & alignment

    My RLHF and evaluation background gives me a quality lens that goes beyond metrics — systems must behave predictably in the real world.

  • 03
    Continuous improvement

    From Bayesian statistics to autonomous agents, I constantly evolve my knowledge and apply it where it matters most.

Now

What I'm building

  • 001
    AI Engineer at H&Co Latam

    Leading end-to-end GenAIOps pipelines with AWS Bedrock, LangGraph, and OpenAI API for production LLM apps with multi-agent orchestration.

  • 002
    Advanced RAG architectures

    Designing RAGOps with vector DBs (OpenSearch, Pinecone), semantic reranking, Phoenix Arize evaluation, and Kubernetes deployments.

  • 003
    Multi-agent systems

    Exploring LangGraph orchestration patterns: conditional routing, parallel tool execution, and state management for complex AI workflows.

Selected projects

WHAT I'VE
BUILT

001
The making of the
LLM Agent with RAG
AI / NLP — AWS Bedrock
Production agentic system routing multi-channel messages through a LangGraph state machine — Pinecone vector search, Redis memory, OpenAI moderation guardrails, and Blue/Green Kubernetes deployment.
LangGraph Pinecone LangChain AWS Bedrock
Click to explore
002
The making of the
ML Transaction Classifier
ML / Fintech — AWS SageMaker
Multi-tenant ML pipeline classifying financial transactions to General Ledger accounts using XGBoost, BERT embeddings, and AWS Step Functions — with per-tenant model specialization via feedback loops.
XGBoost SageMaker Step Functions AWS Lambda
Click to explore
003
The making of the
Travel Metadata Processor
LLM / Serverless — AWS Bedrock
Serverless LLM pipeline extracting structured metadata from raw travel content using AWS Bedrock and Claude — transforming unstructured data into enriched, queryable records at scale.
AWS Bedrock Claude AI AWS Lambda
Click to explore
004
The making of the
Serverless Cloud AWS Chatbot
AI / Cloud — AWS Serverless
100% AWS-native conversational AI platform on Bedrock Agent — with Bedrock Knowledge Base, Bedrock Guardrails, WebSocket API Gateway, Cognito auth, and CDK nested CloudFormation stacks. Zero external dependencies.
Bedrock Agent Guardrails Lambda CDK
Click to explore
005
The making of the
Hybrid Vector Search Engine
NLP / Search — Open-Source
Master's thesis — multimodal search over Glovo's food catalog using pgvector HNSW + BM25 tsvector fused via RRF. Search by text or image: upload a burger photo and find burger restaurants. 100% open-source.
pgvector · HNSW Jina CLIP v2 Supabase FastAPI
Click to explore
006
The making of the
NLP Sentiment Analysis
NLP / ML — HuggingFace
Fine-tuned BERT model for multi-class sentiment classification, served via FastAPI with HuggingFace Transformers.
BERT HuggingFace FastAPI
Click to explore
007
The making of the
Recommendation System
ML / RecSys — TensorFlow
Collaborative + content-based recommendation engine using TensorFlow embeddings and Redis for real-time serving.
TensorFlow Embeddings Redis
Click to explore
008
The making of the
U-Net Semantic Segmentation
CV / Medical — PyTorch
Medical image segmentation using U-Net architecture in PyTorch — trained for pixel-level classification of anatomical structures.
U-Net Medical PyTorch
Click to explore
5 / 8 projects
Contact
LET'S TALK
ABOUT YOUR
PROJECT