Your own literature review pipeline
Run a literature review on a topic you actually care about
- 1 Go to Downloads (curriculum.32dots.de/share) and re-import 'Session 8 — Multi-stage literature pipeline' if it's not already in your n8n.
- 2 Pick a narrow scientific question from your own work — e.g. 'CRISPR screens in primary T cells', 'gut microbiome and Parkinson's', 'single-cell RNA-seq of glioblastoma'. Specific beats broad.
- 3 Open the first node and change the PubMed query to your question. Keep retmax at 8 for the first run.
- 4 Execute the workflow. Watch the structured extraction column-by-column: method, sample size, key finding, limitation.
- 5 Read three of the rows critically. Find at least one place where the AI got the method wrong, missed a limitation, or invented a number.
You can name one paper where the AI extraction is good, one where it's wrong, and one where it's plausibly wrong but you'd need to read the full paper to tell.
Why this pipeline beats a single prompt — and where it still fails
In a science pipeline, the AI is one stage of five, not the whole thing. Search and fetch are deterministic so PMIDs are real. Filter and format are deterministic so the output is reproducible. The LLM only does the one thing a deterministic step cannot — read prose and pull out structure. This is the pattern you will see in every serious research-AI system.
- ?If you remove the Filter node, what kind of garbage shows up in the table?
- ?Why is the extraction step the only AI step? What goes wrong if you let the AI also choose which papers to include?
- ?Where would you add a second LLM call — and what would justify the extra cost?
Tighten the pipeline for your domain
Choose ONE concrete extension and implement it end-to-end. Keep the change small — one node added or modified — and test that it works on at least three papers.
- 1 Option A — Add a domain filter. In the Filter node, additionally drop rows where method does not contain one of your domain's keywords (e.g. 'qPCR', 'flow cytometry', 'scRNA-seq', 'mass spec'). Aim for precision over recall.
- 2 Option B — Add a 'sample size' threshold. Drop rows where sample size is below a number that makes sense for your field (n<10 for cell-line work, n<50 for clinical observational, etc.). Decide the cutoff before you look at the results.
- 3 Option C — Add a second extraction field. Pick one: 'cell line / organism', 'statistical test', 'effect size', 'control group description'. Update the JSON schema in the AI Extract node and the table header in the Format node.
- 4 Option D — Replace the AI Extract prompt with a stricter version that refuses to answer when the abstract does not contain the requested field — and verify it actually refuses on a few abstracts where the field is genuinely absent.
- 5 Run the modified pipeline on the same query you used in the USE phase. Compare the before/after tables side by side.
Export the modified workflow JSON. Share the before/after comparison table and one sentence on what you would still not trust this pipeline to do for your real research.
✎This pipeline reads abstracts, not papers. For your field, when is an abstract enough — and when does relying on abstracts actively mislead a literature review?