0
Initializing neural network...
Machine Learning Engineer · Bogotá, Colombia

MANUEL
ALEJANDRO
DIAZ RUBIANO

Building end-to-end AI systems — from LLMs and RAG architectures to MLOps pipelines and production deployment.

Pipeline ready · inference Deployed
Latency p95
LLM Accuracy
Faithfulness
Throughput
CPU Usage
Uptime
99.8%
Hallucination
Scroll
About me
Manuel Alejandro Diaz Rubiano
Manuel Alejandro Diaz Rubiano
AI Engineer & ML Specialist

I'm an AI Engineer and ML Specialist with a background in Statistics and Mathematics, specialized in building end-to-end AI systems — from model design and fine-tuning to production deployment, observability, and continuous optimization.

My experience spans RAG and agentic architectures end-to-end: data ingestion and preprocessing, chunking and embeddings, vector storage, semantic retrieval, reranking, prompt orchestration, multi-agent workflows, evaluation, monitoring, and scalable deployment. I've built these systems using LangChain and LangGraph as core orchestration layers, with OpenSearch, Pinecone, Chroma, and Redis for retrieval and state management, and with AWS, GCP, and Azure to support cloud-native AI infrastructure.

I place strong emphasis on production reliability and measurable impact, using tools such as MLflow, Phoenix Arize, CloudWatch, Docker, Kubernetes, FastAPI, and GitHub Actions to track performance, evaluate quality, monitor failures, and ship maintainable AI systems with proper CI/CD and observability.

Beyond LLMs, I bring deep experience in traditional machine learning — forecasting, probabilistic scoring, NLP, fine-tuning, and predictive analytics with Prophet, ARIMA, LSTM, BERT, T5, and GPT-based architectures. That gives me the judgment to choose the right approach for each problem, whether generative, predictive, or a hybrid system that combines both.

My background in RLHF, annotation, and model evaluation helps me approach AI from a quality and alignment perspective, not just a systems perspective. I can help organizations design, build, deploy, and improve AI products end-to-end — from early architecture decisions to reliable production systems that support automation, decision-making, and user-facing intelligence.

LLMs & GenAI RAG & Retrieval Multi-Agent Systems MLOps / GenAIOps NLP & Fine-tuning ML Predictivo AWS / GCP / Azure RLHF & Evaluation FastAPI · Docker · K8s LangChain · LangGraph
Selected projects

WHAT I'VE
BUILT

001 LLM Agent with RAG
LangChainOpenSearchLangGraphRAG
002 LLM Transaction Classifier
AWS BedrockPydanticLangChainAWS Lambda
003 NLP Sentiment Analysis
BERTHuggingFaceFastAPI
004 Recommendation System
TensorFlowEmbeddingsRedis
005 VAE Anomaly Detection
KerasTime SeriesVAE
006 U-Net Semantic Segmentation
U-NetMedicalPyTorch
Technologies

MY
STACK

LLMs & Structured AI
GPT · Claude · OpenAI · Anthropic · Structured Outputs · JSON Schema · Function Calling · Pydantic · Instructor
RAG & Retrieval Systems
OpenSearch · Pinecone · ChromaDB · Redis · FAISS · pgvector · Semantic search / reranking · BGE · E5 · BM25 + vector · cross-encoders · Cohere Rerank
Embeddings & Context Pipelines
text-embedding-large · text-embedding-small · BAAI BGE · E5 · Instructor embeddings · sentence-transformers · chunking · metadata filtering
Agentic Orchestration
LangChain · LangGraph · Multi-Agent Workflows · Stateful Execution · Conditional Routing · Tool Calling · Parallel Execution
Serving & Inference
FastAPI · vLLM · TorchServe · AWS Lambda · API Gateway · Batch / async serving · micro-batching · async queues · event-driven inference · Streaming inference
Observability & Evals
Phoenix Arize · LangSmith · MLflow · CloudWatch · OpenTelemetry · OpenInference · tracing · evals · hallucination detection
MLOps / GenAIOps
Docker · Kubernetes · GitHub Actions · Terraform · AWS CDK · SageMaker · CI/CD · model versioning · canary releases
ML & Forecasting
PyTorch · TensorFlow · Scikit-learn · Prophet · ARIMA · LSTM · BERT · T5 · predictive analytics
Contact
LET'S TALK
ABOUT YOUR
PROJECT