By the way – I have no idea if any of them are true.
Conjecture 1. That the distance between any two wikipedia pages, randomly chosen, as measured by wikilinks , is on average 6. Conjecture 2. That wikipedia is sufficiently formal and complete that you could build a useful general purpose AI knowledge base using it. Conjecture 3. That wikipedia has low information entropy . Conjecture 4. That the development of a wikipedia article over time occurs in a manner consistent to the biological evolution of a species . Conjecture 5. That the relationship between the amount of material in wikipedia and the number of article views is exponential . Conjecture 6. That wikipedia is, on average, factually accurate .
- Social networks conform to the "six degrees of separation" principle. If wikipedia does, what does that say about its social roots / the way it's constructed?
- See Cyc and others. Is there enough formally coded information in wikipedia? What about the semantic relationship between the source sentence containing a link and the summary of the linked article?
- What does "low entropy" mean anyway? More structured? Simpler? More redundant? More readable? What about the entropy across wikilinks?
- Does an article behave like DNA?
- Can we "prove" Reed's law? How do you measure the size of the content?
- Is it accurate more than average? Can you predict the accuracy of an article?