Am I the only one who is surprised how small the storage size is? I understand it’s a lot of text but to realize it easily fits on a few cheap HDDs raided together is still incredible to me.
Almost 9 GB, though it's significantly smaller if you download the version without images (which my Wikipedia reader can't render anyways.) IIRC pure text is only like 2 GB.
There is a torrent for them somewhere.
As large as that is, I am surprised that it wasn't huger than that.
My guess is not by much. Text doesn't require a lot of storage. After looking at a few random files my estimate is that the size of the full compressed dump would be about 30x (~10 GB)
It's most of the large articles, the entirety (stripped of citations and image metadata) and compressed is more like 4GB. For comparison, everything (including citations) as a wikimedia XML dump is about 7GB bzipped and thirty something uncompressed
I assume this must be only the text portion, and heavily compressed?
reply