DealForge autonomously sources, scores, and writes investment memos on venture deals. Stop manually hunting.
1,180+ deals tracked · 22 AI investment memos · Updated daily
Show HN: OpenFable – Open-source RAG engine using tree-structured indexes
Hi HN, I built OpenFable, an open-source retrieval engine that implements the FABLE algorithm (<a href="https://arxiv.org/abs/2601.18116" rel="nofollow">https://arxiv.org/abs/2601.18116</a>) for RAG pipelines. I'm using it in another project and thought that others might benefit.<p><pre><code> Most RAG systems chunk documents into flat segments and retrieve by vector similarity. This works for simple lookups but breaks when answers span multiple sections, when relevant content is buried in a subsection, or when you need to control how many tokens you're sending to an LLM. OpenFable takes a different approach: when you ingest a document, it uses an LLM to identify discourse boundaries (not fixed-size windows), then builds a hierarchical tree, root, sections, subsections, leaf chunks, with embeddings at every level. Retrieval combines two paths: 1. LLM-guided path: the LLM reasons about which documents and subtrees are relevant from summaries 2. Vector path: similarity search with structure-aware score propagation through the tree Results from both paths are fused, deduplicated, and trimmed to fit a token budget you specify. You get the most relevant chunks, in document order, within budget. From the FABLE paper: the algorithm matches full-context inference (517K tokens) using only 31K tokens, 94% reduction, while hitting 92% completeness vs. Gemini-2.5-Pro at 91% with the full document. Retrieval only; OpenFable returns ranked chunks, not generated answers. Bring your own LLM for generation. It runs as a Docker stack (FastAPI + PostgreSQL/pgvector) and exposes both a REST API and an MCP server, so LLM agents like Claude Desktop or Cursor can use it directly. Trade-offs I want to be upfront about: - Ingestion is expensive; every document requires multiple LLM calls for chunking and tree construction - Retrieval isn't sub-second, the LLM-guided paths add round-trips - No built-in auth; designed to sit behind a reverse proxy - v0.1.0 — works end to end but the roadmap includes async ingestion, document deletion, and metadata filtering Stack: Python 3.12, FastAPI, SQLAlchemy, pgvector, LiteLLM, fastMCP. Apache 2.0. Happy to answer questions about the algorithm, implementation choices, or benchmarks.</code></pre>
OpenFable addresses a high-value pain point in the RAG market—retrieval accuracy and context window efficiency—using a sophisticated hierarchical tree-indexing approach. While the technical implementation of the FABLE algorithm is impressive and timely, the project is currently a very early-stage open-source tool with significant competition from established frameworks and high operational costs.