Thread: Evopedia
View Single Post
Posts: 58 | Thanked: 42 times | Joined on Jan 2010
#79
Originally Posted by combiochem View Post
Sometimes, when I search a certain term, the result page says, "ERROR - Page not found". It means that the term is already indexed but it does not have an actual article for the term.
For example, "Machine Learning" term in "wikipedia_en_2010-01-16" dump

And the other question is about the dump.
I found that there is an original dump from wikipedia ("http://dumps.wikimedia.org/") and the dump for evopedia is different from the original wikipedia dump (also different from Aard dictionary?). In that case, is there any way to make a dump from other dumps (a kind of own wikipedia?)
The problem you mention is caused by the dump creation process, it simply contains some errors, sorry for that. I hope that the next dump will not have these problems.

The "database backup dumps" from http://dumps.wikimedia.org are actually used to generate the evopedia dumps. They have to be (mainly) converted from Wikipedia markup to HTML, indexed and recompressed. How to run this process yourself is explained on the wiki page http://wiki.maemo.org/Evopedia, although this information is a bit outdated. You get better results if you download the distributed dump process client from http://dumpathome.evopedia.info/client and run the individual jobs in scripts/dumpathome.sh: First both import jobs, then createdump and finally package. If you want, I can tell you more about it.

Unfortunately, I don't know which format the aard dictionary uses.