> pko...@gmail.com schrieb am Freitag, 20. November 2020 um 12:28:16 UTC+1:
>> Hello, I am an Aard2 user with a little python knowledge. I heard that making a .slob involves downloading all pages via the MediaWiki API and an intermediate CouchDB storage, and that this takes a lot of time. Why not just download the .zim package of the official HTML dumps  and repack into a .slob with libzim ? This works on my Debian 10:
> Klings interestingly!
> I had a quick look to the equivalent Wikipedia german (full articles), which is quite actual from 11-2020, this has as zim-file 13 GB. As slob it has only about 5.4 GB!
> So did you try to convert a whole zim-Wikipedia (without pictures obviously) already? Is it as small as the direct produced slob?
slob creation and compression depend only on the slob python module,
unrelated to zim or other article source, as the code above
decompresses the articles in RAM anyway. I did not try, but, if the
output slob is not as small, then that is an issue/bug of the slob
> I tried zim reader and files in the past, but at this time the zim-files were mostly old, so I spent days to download my own wiki as zim (with a tool called MWoffliner) and in the end the zim-reader was far not that good as aard2, so I stopped that.
> But if these zim-files are now produced in always new versions, it might be a faster way to get actual slobs.
They are generated monthly since a few years. Wikimedia (the
non-profit behind Wikipedia) uses them to distribute Wikipedia to the