- 1 What are your open access policies?
- 2 Where can I download images and media files?
- 3 Where can I get RDF data?
- 4 Where can I get page view data?
- 5 Where can I get unique visitors data?
- 6 Where can I get editor data?
- 7 Where can I get data about the fundraiser / banners?
- 8 Is it OK to spider the website?
What are your open access policies?
WMF encourages open access and ties technical, financial and data collection support for research to openness requirements as an incentive for open-access projects. See Wikimedia Foundation support for details.
Where can I download images and media files?
For more information about downloading images and media files see Where are images and uploaded files.
Where can I get RDF data?
Wikipedia data is avaiable in RDF format on DBpedia.
Where can I get page view data?
The best source of page view data are the Pageview statistics collected from Squid logs. WikiStats offers some information about page views based on the same source. Note that these are not unique hits.
Do not use the API to get page view stats as many requests only make it to the Squid cache.
Where can I get unique visitors data?
Browse raw data from the report card here.
Where can I get editor data?
- WikiChecker uses the API to get editor activity statistics
- X!'s Edit Counter is hosted on Toolserver and gives a summary of a user's editing history
- Number of Wikipedians by number of edits
The fundraising team is starting to make sanitized data fundraising data available at samarium.wikimedia.org.
Is it OK to spider the website?
It's a very bad idea to spider Wikipedia projects' websites. It puts load on the servers and will take you longer than downloading the project's dump. Use the dumps when possible. If you have to spider the website, spread your requests in time and use low-traffic hours.