This repository is a Retrieval-Augmented Generation (RAG) example using Apache Pinot, LangChain, and OpenAI. The use case is to load documentation and allow an LLM to answer questions provided by a user. This approach enables you to generate AI responses that are fresh and in real time. This RAG example uses LangChain’sDocumentation Index
Fetch the complete documentation index at: https://docs.startree.ai/llms.txt
Use this file to discover all available pages before exploring further.
RecursiveUrlLoader. It accepts a URL, recursively loads pages, and converts them into documents. These documents are converted into embeddings, submitted to a Kafka topic, and consumed by Apache Pinot.
To understand how this recipe processes data, examine the commands executed in the Makefile.
| Pinot Version | 1.1.0 |
|---|---|
| Code | startreedata/pinot-recipes/genai |
Prerequisites
To follow the code examples in this guide, do the following:- Install Docker locally.
- Download recipes
Docker
This repo builds the Apache Pinot project. You may get an errorNo space left on device when building the container. Execute the command below to free resources before building.
Building the Pinot image will take about 25 minutes to finish.

