- Analyzed and documented the agent code and its flow
- Examined the utility profit API
- Developed hypothetical flow for data acquisition
- Executed various curl requests to test endpoints
- Created automation script for initial flow
- Combined actual data with mocked data to produce sample JSON in required format
Choosing between RAG (Retrieval-Augmented Generation) and fine-tuning depends on your use case, data type, infrastructure, and latency/accuracy needs. This guide helps you understand when to use each technique.
You change the model itself.
This document outlines practical applications where either Retrieval-Augmented Generation (RAG), fine-tuning, or both are best suited. It also includes a rough cost breakdown to help with planning.
- Why RAG: Fetch live policy, FAQ, or documentation.
Fine-tuning can work wonders — but only within the bounds of a model's inherent intelligence. This guide explores when a model is too small to meet your task’s quality requirements and how to tell if you've hit that limit.
LangSmith (by LangChain) and LangFuse are powerful observability and analytics tools for LLM workflows. You can use them to track, analyze, and improve model performance — and even synthesize better training data using real user interactions.
This guide covers:
- How to connect LangSmith or LangFuse
- How to log successful Q&A pairs
- How to use logged data for fine-tuning or RAG
- Sample integration script
Cleaning and validating your fine-tuning dataset is one of the most important steps to ensure effective model training. Poorly formatted or inconsistent data leads to degraded model quality, wasted GPU hours, and hard-to-debug behaviors.
This doc covers:
- Common formatting pitfalls
- Best practices
- A sample validation script