AI Horizon Forecast

AI Horizon Forecast

Part 2: Making Documents Searchable with Vector Stores and Retrieval

Building RAG pipelines with Chroma and LangChain

Nikos Kafritsas's avatar
Nikos Kafritsas
Feb 09, 2026
∙ Paid
Created with DALLE*3

In this 3-part series, we walk through how to build a document intelligence app.

In Part 1, we explored Docling, a state-of-the-art tool for parsing and structuring documents—the first step in a RAG (Retrieval-Augmented Generation) pipeline.

In this part, we focus on a RAG workflow.

While there are many RAG tutorials available, this series shares an approach that has worked well for me, with a particular emphasis on content-rich documents such as images, tables, and text.

Let’s get started!

✅ Find the notebook for this article here: Language Models (Project 3)



Why Retrieval Augmented Analysis?

Why should we combine an LLM with an external vector store?

Top-tier models can already analyze papers, financial reports, or news articles. They aren’t limited by their training cutoff anymore because they call web search agents that can pull up-to-date info from Google. You can even upload a few PDFs and chat with them directly. This works well for small tasks, but there are 2 main issues:

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2026 Nikos Kafritsas · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture