Menu

Main Menu
Talk Get Daily Search

Member's Online

    User Name
    Password

    In development: Full Wikipedia reader for N900

    Reply
    t3h | # 1 | 2010-04-19, 11:24 | Report

    There are currently several existing Wikipedia readers for the N900. However, many of them don't contain full articles, the full database, or otherwise have significant shortcomings.

    I'm currently working on something based around this project: http://users.softlab.ece.ntua.gr/~tt...iaOffline.html.

    There is a disadvantage with this method though. The articles.xml file that this uses is ~6GB, and the index that it builds is another 3. However, in the modern times we live in, the N900 has 32GB of storage, and my 16GB card cost about $50 AUD.

    An advantage, though, will be that the end user is more easily able to create their own updated dumps.

    I plan on replacing the PHP Mediawiki parser in that project with a faster Python one, as well as some Maemo specific tweaking.

    It might seem overboard to have all of Wikipedia in your pocket - but so was the idea of having a Linux machine in your pocket when the 770 first came out

    Any comments/suggestions/questions?

    Edit | Forward | Quote | Quick Reply | Thanks
    The Following 7 Users Say Thank You to t3h For This Useful Post:
    biggzy, Haus3r, mikec, omeriko9, sanki, tissot, zimon

     
    pantera1989 | # 2 | 2010-04-19, 11:35 | Report

    I don't know..6 GB is quite a lot. Between movies, PSX games mp3 and maybe sygic maps, 32GB becomes 5GB.

    And how will this be updated if the page is updated? Will it look through all the pages before updating? The N900 is an internet device. Offline viewing does kinda destroy the point.

    I would suggest that you create an app that is able to download the articles one wants for offline viewing. This way there won't be a lot of space taken up.

    Edit | Forward | Quote | Quick Reply | Thanks
    The Following 2 Users Say Thank You to pantera1989 For This Useful Post:
    biggzy, sanki

     
    tredlie | # 3 | 2010-04-19, 11:48 | Report

    offline wikipedia is indeed a killer application even on internet devices like the N900. What do you do when you are abroad without a data plan?

    A couple of questions: I guess images are not really part of the installation?
    what would be the realy advantage compared to evopedia, the LaTeX math presentation?
    Or being able to produce own dumps more quickly?

    tredlie

    Edit | Forward | Quote | Quick Reply | Thanks

     
    BigLebowski | # 4 | 2010-04-19, 11:52 | Report

    Originally Posted by pantera1989 View Post
    And how will this be updated if the page is updated? Will it look through all the pages before updating? The N900 is an internet device. Offline viewing does kinda destroy the point.
    Hm - there are quite some use cases where an offline Wikipedia might be quite handy ... areas with no 3G coverage, for instance ... or going abroad (data roaming is still expensive these days).

    I used to look up terms quite often in my Mobipocket Wikipedia - but then again, that was on a Palm Tungsten with no Internet connection at all

    Edit | Forward | Quote | Quick Reply | Thanks

     
    zimon | # 5 | 2010-04-19, 12:03 | Report

    Is the Wikipedia db compressed (bzip2 -9) when in offline locally?

    Edit | Forward | Quote | Quick Reply | Thanks

     
    t3h | # 6 | 2010-04-19, 12:12 | Report

    Originally Posted by tredlie View Post
    A couple of questions: I guess images are not really part of the installation?
    what would be the realy advantage compared to evopedia, the LaTeX math presentation?
    Or being able to produce own dumps more quickly?
    Images are not part of the installation a) because that would just be insanely large, and b) because Wikipedia doesn't provide dumps of them. http://dumps.wikimedia.org/enwiki/20100312/ is where they are from.

    LaTeX math presentation will be present as it is in the current desktop version. I am currently working to replace the PHP based parser with the Python-based mwlib. Then I just have to recode the bzip2 file extractor in Perl with some Python. After that, it's just replacing the Django webservice with a simple HTTP server, and then it shouldn't require anything special - apart from a few compiled bits.

    Edit | Forward | Quote | Quick Reply | Thanks

     
    t3h | # 7 | 2010-04-19, 12:17 | Report

    There is also this: https://launchpad.net/wikipediadumpreader/ which I will have a look at - I haven't seen a Maemo package for it yet, but if it works decently well, I may port that instead, considering how Maemo is getting QT'd.

    Edit | Forward | Quote | Quick Reply | Thanks
    The Following User Says Thank You to t3h For This Useful Post:
    tissot

     
vBulletin® Version 3.8.8
Normal Logout