In an age of infinite content, curation isn't about finding more—it's about finding better.
At Knoww, we process thousands of books. But not everything makes it into NodeCore. We curate aggressively, filtering signal from noise with editorial rigor.
This is how we decide what knowledge deserves to exist in the graph.
The Quality Threshold
Most knowledge platforms optimize for volume. More books, more articles, more content. The assumption: let users decide what's valuable.
We take the opposite approach: selective curation at the platform level.
Our north star question: "Would a domain expert recommend this insight to someone learning the field?"
If not, it doesn't make the cut—regardless of how popular the source book is.
The Five Filters
Every insight passes through five editorial filters before entering NodeCore:
1. Verifiability
Can this claim be validated?
We prioritize insights backed by:
- Peer-reviewed research
- Empirical data or experiments
- Expert consensus in the field
- Reproducible methods
Personal opinions, unfounded speculation, and anecdotal claims get filtered out unless they're explicitly framed as perspectives rather than facts.
Exception: philosophical or ethical arguments don't require empirical proof, but they must be logically coherent and intellectually rigorous.
2. Generalizability
Does this insight apply beyond its immediate context?
Hyper-specific tactics ("how to rank on Google in 2019") age poorly. Principles ("search engines reward authoritative, well-linked content") remain relevant.
We favor timeless insights over time-bound tactics, universal patterns over edge cases, transferable concepts over context-dependent hacks.
3. Clarity
Can someone understand this without extensive prerequisites?
Insights should be accessible to motivated learners, not just domain experts. We rewrite jargon-heavy passages for clarity while preserving technical precision.
If an insight requires three paragraphs of background, it's either split into parent-child nodes (prerequisites + main concept) or reframed for clarity.
4. Novelty vs. Consensus
This is where it gets interesting. We want both:
- Established knowledge — proven concepts everyone should know
- Frontier insights — new research, contrarian perspectives, emerging patterns
But we're wary of:
- Stale conventional wisdom — outdated ideas presented as current
- Contrarianism for its own sake — provocative but unsubstantiated claims
The balance: include both canonical knowledge and well-reasoned challenges to it, clearly labeled.
5. Actionability
Does this insight enable action or understanding?
Not all knowledge is prescriptive, but it should be useful. Either:
- You can apply it (frameworks, mental models, techniques)
- It changes how you think (conceptual breakthroughs, paradigm shifts)
- It connects ideas (reveals relationships between concepts)
Purely descriptive facts without explanatory power rarely make the cut.
Source Selection: Not All Books Are Equal
Before we even process a book, we evaluate the source:
Author Credibility
- Domain expertise (credentials, experience, recognized authority)
- Track record (previous work, peer reputation)
- Transparency about methods and sources
Editorial Rigor
- Publisher reputation (academic presses, reputable trade publishers)
- Citation quality (original research vs. recycled summaries)
- Fact-checking evidence (are claims verified or assumptions?)
Relevance
- Does this cover new ground or rehash existing material?
- Is the author's perspective unique or derivative?
- Are insights backed by primary sources or tertiary summaries?
A bestseller isn't automatically included. A niche academic work might be. Popularity ≠ quality.
Curation vs. Automation
Our pipeline uses machine learning to suggest candidate insights. But humans make the final call.
Why not fully automate?
Because algorithms optimize for patterns in data, not for truth or utility. An ML model can identify that a passage "looks like" an insight (definitional language, causal structure, prescriptive phrasing). But it can't judge:
- Is this claim actually true?
- Is this insight meaningful or trivial?
- Does this contradict better evidence elsewhere?
- Will users find this useful in practice?
Those require human expertise.
The Deletion Decision
Curation isn't just about what to include—it's about what to exclude.
We delete insights that:
- Duplicate better versions: If five books explain compound interest, we keep the clearest explanation and link the others as "related perspectives"
- Become outdated: Technology, policy, and science evolve; we prune obsolete claims
- Fail quality thresholds: Even from good books, not every insight is graph-worthy
Deletion is controversial. Users sometimes disagree: "But that book is a classic!"
We respond: "The book is a classic. That doesn't mean every sentence in it is essential knowledge."
Bias Awareness
All curation involves judgment. Judgment involves biases.
We're transparent about ours:
- Western-centric sources: Most books in our corpus are in English, from Western publishers. We're actively expanding to include global perspectives.
- Recency bias: Newer books get processed faster; we're systematically backfilling classics.
- STEM over humanities: Early corpus leaned toward business, psychology, science. We're balancing with philosophy, history, literature.
- Practical over theoretical: We favor actionable insights; pure theory sometimes gets under-represented.
Awareness doesn't eliminate bias, but it helps us course-correct.
Community Feedback Loop
Curation isn't top-down. Users can:
- Flag issues: "This insight is inaccurate / outdated / unclear"
- Suggest additions: "This book is missing a key concept"
- Propose connections: "These insights should link"
Expert users become trusted curators, gaining editorial privileges to refine the graph.
The Curator's Dilemma
We face a constant tension:
- Inclusive: More sources, more perspectives, more comprehensive coverage
- Selective: Higher standards, tighter curation, less noise
We lean toward selective.
Why? Because the internet already offers infinite, uncurated content. Google gives you 10 million results. We give you the 10 that matter.
That's the value of curation.
The Human Touch
Automation scales. Curation elevates.
Explore NodeCore and experience the difference: every insight vetted by domain experts, every connection validated for relevance, every node crafted for clarity.
In a world drowning in information, curation is an act of kindness.