We study an access trace containing a sample of Wikipediaâ€™s trafï¬c over a 107-day period aiming to identify appropriate replication and distribution strategies in a fully decentralized hosting environment. We perform a global analysis of the whole trace, and a detailed
analysis of the requests directed to the English edition of Wikipedia. In our study, we classify client requests and examine aspects such as the number of read and save operations, signiï¬cant load variations and requests for non existing pages. We also review proposed decentralized wiki architectures and discuss how they would handle Wikipediaâ€™s workload. We conclude that decentralized architectures must focus on applying techniques to efï¬ciently handle read operations while maintaining consistency and dealing with typical
issues on decentralized systems such as churn, unbalanced loads and malicious participating nodes.
Download the paper from here.