Loading patterns...
If this never updates, open the site over HTTP(s) so data can load.
A field guide to LLM graph extraction gone wrong — and the occasional redemption arc.
Common mistakes that show up in production pipelines. If you recognize your setup, we're not judging — we're judging the graph.
If this never updates, open the site over HTTP(s) so data can load.
One highlighted example, rotated by calendar day (UTC).
Loading...
Examples of extraction gone wrong.
Loading...
Nothing glamorous, which is how you know it's real engineering.
Parse documents into logical units (sections, tables, lists) before asking an LLM to assert relations. Text locality matters — fake locality produces fake edges.
Canonical IDs, blocking, pairwise scoring, and human review for long tails. Lowercasing strings is not a merge strategy.
Version your ontology, constrain predicates, and test extraction against gold sets. If the model drifts types every deploy, you don't have a graph.
Edge-level precision/recall, constraint violations, and retrieval quality beats "the graph looks big." Big graphs full of slop are just high-entropy compost.
Store source spans, document IDs, and extraction version. When someone asks why an edge exists, the answer should not be vibes.
Rules and dictionaries for stable relations; LLMs for fuzzy bits — with calibration and abstention when confidence is low.