Discussion:
a DVD with the Spanish Wikipedia (was [Argentina] WikiBrowse improvements)
Sean Moss-Pultz
2010-04-30 16:04:24 UTC
Permalink
Hi Alejandro
[snip]
Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.
I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.
We're doing this XML to HTML conversion as one of the steps in our process
of rendering Wikipedia for our WikiReader device. We can build Spanish
without issues.

All of our source code is here:

http://github.com/wikireader/wikireader

The specific portion you would need is the offline-renderer located here:


http://github.com/wikireader/wikireader/tree/master/host-tools/offline-renderer/

You'll probably need to modify the HTML output for your specific needs. Just
let me know if you get stuck.

Sean
Hernan Olivera
2010-04-30 17:53:11 UTC
Permalink
Post by Sean Moss-Pultz
Hi Alejandro
[snip]
Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.
I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.
We're doing this XML to HTML conversion as one of the steps in our process
of rendering Wikipedia for our WikiReader device. We can build Spanish
without issues.
http://github.com/wikireader/wikireader
http://github.com/wikireader/wikireader/tree/master/host-tools/offline-renderer/
You'll probably need to modify the HTML output for your specific needs.
Just let me know if you get stuck.
Sean
I'll take a look. Thanks!
--
Hernan Olivera
Alejandro J. Cura
2010-04-30 15:43:14 UTC
Permalink
Hi everyone, we need your help.

We are from Python Argentina, and we are working on adapting our
cdpedia project to make a DVD together with educ.ar and Wikimedia
Foundation, holding the entire Spanish Wikipedia that will be sent
soon to Argentinian schools.

Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.

I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.

Following is an email from Hernán describing the problems he's found.

thanks!
--
alecu - Python Argentina


2010/4/30 Hernan Olivera <lholivera-***@public.gmane.org>:
Hi everybody,

I've been working on making an up to date static html dump for the
spanish wikipedia, to use as a basis for the DVD.
I've followed the procedures detailed in the pages below, that were
used to generate the current (and out of date) static html dumps:

1) installing and setting up a mediawiki instance
2) importing the xml from [6] with mwdumper
3) exporting the static html with mediawiki's tool

The procedure finishes without throwing any errors, but the xml import
produces malformed html pages that have visible wikimarkup.

We would really need to have a successful import from the spanish xmls
to a mediawiki instance so we can produce the up to date static html
dump.

Links to the info I used:
[0] http://www.mediawiki.org/wiki/Manual:Installation_guide/es
[1] http://www.mediawiki.org/wiki/Manual:Running_MediaWiki_on_Ubuntu
[2] http://en.wikipedia.org/wiki/Wikipedia_database
[3] http://www.mediawiki.org/wiki/Manual:Importing_XML_dumps
[4] http://meta.wikimedia.org/wiki/Importing_a_Wikipedia_database_dump_into_MediaWiki
[5] http://meta.wikimedia.org/wiki/Data_dumps
[6] http://dumps.wikimedia.org/eswiki/20100331/
[7] http://www.mediawiki.org/wiki/Alternative_parsers
(among others)

Cheers,
--
Hernan Olivera

PS: unluckily I didn't write down every step in detail. I did a lot
more tests than what I wrote here. To make a detailed report I'd like
to go thru the procedure again writing down every option (and to check
if I missed something). I'm finishing installing a server just for
this, because this processes take forever and they blocked other tasks
while making this tests.
Jimbo - thanks for the spur to clean up the existing work.
All - Let's start by cleaning up the mailing lists and setting a few
short-term goals :-)  It's a good sign that we have both charity and love
converging to make something happen.
* For all-platform all-purpose wikireaders, let's use
Wikimania (Erik, were you going to set this up?  I think we agreed to
deprecate wiki-offline-reader-l and replace it with offline-l.)
* For wikireaders such as WikiBrowse and Infoslicer on the XO, please
I would like to see WikiBrowse become the 'sugarized' version of a reader
that combines the best of that and the openZim work.  A standalone DVD or
USB drive that comes with its own search tools would be another version of
the same.   As far as merging codebases goes, I don't think the WikiBrowse
developers are invested in the name.
I think we have a good first cut at selecting articles, weeding out stubs,
and including thumbnail images.  Maybe someone working on openZim can
suggest how to merge the search processes, and that file format seems
unambiguously better.
Kul - perhaps part of the work you've been helping along for standalone
usb-key snapshots would be useful here.
Please continue to update this page with your thoughts and progress!
http://meta.wikimedia.org/wiki/Offline_readers
SJ
My dream is quite simple: a DVD that can be shipped to millions of
people with an all-free-software solution for reading Wikipedia in Spanish.
 It should have a decent search solution, doesn't have to be perfect, but it
should be full-text.  It should be reasonably fast, but super-perfect is not
a consideration.
Hello! I am an educator, not a programmer. I can help selecting
articles or developing categories related to school issues.
Iris - you know the main page of WikiBrowse that you see when the reader
first loads?  You could help with a new version of that page.  Madeleine
(copied here) worked on the first one, but your thoughts on improving it
would be welcome.
Manuel Schneider
2010-04-30 20:24:38 UTC
Permalink
Hey,

have a look at Kiwix:

http://www.kiwix.org/index.php/Main_Page/es

As far as I know Emmanuel (maintainer of Kiwix) has made ZIM files for
es:wp.

Alternatively here is a description how to make them:
http://www.kiwix.org/index.php/Tools/en
(http://www.kiwix.org/index.php/Tools/es - not complete)

/Manuel
Post by Alejandro J. Cura
Hi everyone, we need your help.
We are from Python Argentina, and we are working on adapting our
cdpedia project to make a DVD together with educ.ar and Wikimedia
Foundation, holding the entire Spanish Wikipedia that will be sent
soon to Argentinian schools.
Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.
I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.
Following is an email from Hernán describing the problems he's found.
thanks!
--
Regards
Manuel Schneider

Wikimedia CH - Verein zur Förderung Freien Wissens
Wikimedia CH - Association for the advancement of free knowledge
www.wikimedia.ch
Pascal Martin
2010-05-01 12:11:17 UTC
Permalink
Hello,

you could test our solution also :
http://www.okawix.com/

If you want we could make .iso ready to use.

----- Original Message -----
From: "Alejandro J. Cura" <alecura-***@public.gmane.org>
To: "Samuel Klein" <meta.sj-***@public.gmane.org>
Cc: <ibarrags-***@public.gmane.org>; "Jimmy Wales" <jwales-zSU+zkDXJht8UrSeD/***@public.gmane.org>; "Madeleine
Ball" <mad-OsFVWbfNK3isTnJN9+***@public.gmane.org>; "Facundo Batista" <facundobatista-***@public.gmane.org>;
<Wiki-offline-reader-l-***@public.gmane.org>; "Offline Wikireaders"
<wikireader-oU9gvf+***@public.gmane.org>; "Cecilia Sagol" <csagol-***@public.gmane.org>; "Pomies
Patricia" <ppomies-***@public.gmane.org>; "Patricio Lorente"
<patricio.lorente-***@public.gmane.org>; "Enrique Chaparro" <cinabrium-***@public.gmane.org>;
"Sean Moss-Pultz" <sean-4Bgg8jF3iZdWk0Htik3J/***@public.gmane.org>; "Kul Takanao Wadhwa"
<kwadhwa-***@public.gmane.org>; "Emmanuel Engelhart"
<emmanuel.engelhart-***@public.gmane.org>; <godiard-***@public.gmane.org>; "Diego Mascialino"
<dmascialino-***@public.gmane.org>; "Hernan Olivera" <lholivera-***@public.gmane.org>;
<cjb-2X9k7bc8m7Mdnm+***@public.gmane.org>; "Iris Fernández" <irisfernandez-***@public.gmane.org>; "OpenZim
devel" <dev-l-***@public.gmane.org>
Sent: Friday, April 30, 2010 5:43 PM
Subject: [openZIM dev-l] a DVD with the Spanish Wikipedia (was
[Argentina]WikiBrowse improvements)


Hi everyone, we need your help.

We are from Python Argentina, and we are working on adapting our
cdpedia project to make a DVD together with educ.ar and Wikimedia
Foundation, holding the entire Spanish Wikipedia that will be sent
soon to Argentinian schools.

Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.

I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.

Following is an email from Hernán describing the problems he's found.

thanks!
--
alecu - Python Argentina


2010/4/30 Hernan Olivera <lholivera-***@public.gmane.org>:
Hi everybody,

I've been working on making an up to date static html dump for the
spanish wikipedia, to use as a basis for the DVD.
I've followed the procedures detailed in the pages below, that were
used to generate the current (and out of date) static html dumps:

1) installing and setting up a mediawiki instance
2) importing the xml from [6] with mwdumper
3) exporting the static html with mediawiki's tool

The procedure finishes without throwing any errors, but the xml import
produces malformed html pages that have visible wikimarkup.

We would really need to have a successful import from the spanish xmls
to a mediawiki instance so we can produce the up to date static html
dump.

Links to the info I used:
[0] http://www.mediawiki.org/wiki/Manual:Installation_guide/es
[1] http://www.mediawiki.org/wiki/Manual:Running_MediaWiki_on_Ubuntu
[2] http://en.wikipedia.org/wiki/Wikipedia_database
[3] http://www.mediawiki.org/wiki/Manual:Importing_XML_dumps
[4]
http://meta.wikimedia.org/wiki/Importing_a_Wikipedia_database_dump_into_MediaWiki
[5] http://meta.wikimedia.org/wiki/Data_dumps
[6] http://dumps.wikimedia.org/eswiki/20100331/
[7] http://www.mediawiki.org/wiki/Alternative_parsers
(among others)

Cheers,
--
Hernan Olivera

PS: unluckily I didn't write down every step in detail. I did a lot
more tests than what I wrote here. To make a detailed report I'd like
to go thru the procedure again writing down every option (and to check
if I missed something). I'm finishing installing a server just for
this, because this processes take forever and they blocked other tasks
while making this tests.
Jimbo - thanks for the spur to clean up the existing work.
All - Let's start by cleaning up the mailing lists and setting a few
short-term goals :-) It's a good sign that we have both charity and love
converging to make something happen.
* For all-platform all-purpose wikireaders, let's use
Wikimania (Erik, were you going to set this up? I think we agreed to
deprecate wiki-offline-reader-l and replace it with offline-l.)
* For wikireaders such as WikiBrowse and Infoslicer on the XO, please
I would like to see WikiBrowse become the 'sugarized' version of a reader
that combines the best of that and the openZim work. A standalone DVD or
USB drive that comes with its own search tools would be another version of
the same. As far as merging codebases goes, I don't think the WikiBrowse
developers are invested in the name.
I think we have a good first cut at selecting articles, weeding out stubs,
and including thumbnail images. Maybe someone working on openZim can
suggest how to merge the search processes, and that file format seems
unambiguously better.
Kul - perhaps part of the work you've been helping along for standalone
usb-key snapshots would be useful here.
Please continue to update this page with your thoughts and progress!
http://meta.wikimedia.org/wiki/Offline_readers
SJ
My dream is quite simple: a DVD that can be shipped to millions of
people with an all-free-software solution for reading Wikipedia in Spanish.
It should have a decent search solution, doesn't have to be perfect,
but it
should be full-text. It should be reasonably fast, but super-perfect is
not
a consideration.
Hello! I am an educator, not a programmer. I can help selecting
articles or developing categories related to school issues.
Iris - you know the main page of WikiBrowse that you see when the reader
first loads? You could help with a new version of that page. Madeleine
(copied here) worked on the first one, but your thoughts on improving it
would be welcome.
Jimmy Wales
2010-05-01 12:14:44 UTC
Permalink
You are all making me very happy with this important work. I am sad
that I'm not able to personally roll up my sleeves and help you. :) But
I am excited to see progress, thank you so much!
Hernan Olivera
2010-07-07 07:49:27 UTC
Permalink
I can think of a few different server farms that would be glad to run
your process :-) Which script is it that will take 3000 (!) hours to
run?
'Dump the Spanish Wikipedia in HTML'
php dumpHTML.php -d /home/hernan/html3 -s 1 -e 1824046 --checkpoint
/home/hernan/html3/check.txt --show-titles
SJ
Post by Hernan Olivera
Hi everyone. We need your help again.
We finally have a working mirror for generating the static html version
of
Post by Hernan Olivera
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of processing in
our little sempron server (4 months!).
How many time could it take in Wikimedia's servers?
Thanks
PD: excuse me if you receive this e-mail twice
Post by Jimmy Wales
You are all making me very happy with this important work. I am sad
that
Post by Hernan Olivera
Post by Jimmy Wales
I'm not able to personally roll up my sleeves and help you. :) But I am
excited to see progress, thank you so much!
--
Hernan Olivera
--
Samuel Klein identi.ca:sj w:user:sj
--
Hernan Olivera
Kul Takanao Wadhwa
2010-07-07 08:42:33 UTC
Permalink
I will also look into possible options. I just got to Gdansk for
Wikisym/Wikimania so give me a little time but if anyone else has ideas
feel free to chime in.

--Kul
El 7 de julio de 2010 04:38, Samuel Klein <meta.sj
I can think of a few different server farms that would be glad to run
your process :-) Which script is it that will take 3000 (!) hours to
run?
'Dump the Spanish Wikipedia in HTML'
php dumpHTML.php -d /home/hernan/html3 -s 1 -e 1824046 --checkpoint
/home/hernan/html3/check.txt --show-titles
SJ
Post by Hernan Olivera
Hi everyone. We need your help again.
We finally have a working mirror for generating the static html
version of
Post by Hernan Olivera
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of
processing in
Post by Hernan Olivera
our little sempron server (4 months!).
How many time could it take in Wikimedia's servers?
Thanks
PD: excuse me if you receive this e-mail twice
Post by Jimmy Wales
You are all making me very happy with this important work. I am
sad that
Post by Hernan Olivera
Post by Jimmy Wales
I'm not able to personally roll up my sleeves and help you. :)
But I am
Post by Hernan Olivera
Post by Jimmy Wales
excited to see progress, thank you so much!
--
Hernan Olivera
--
Samuel Klein identi.ca:sj w:user:sj
--
Hernan Olivera
Kul Takanao Wadhwa
2010-08-11 20:40:52 UTC
Permalink
Hernan,

We have some contacts that may be able to give you access to processing
power. Tomasz and I need to more about what your specific needs.

Just contact me and Tomasz directly and we'll go from there.

Kul
El 7 de julio de 2010 04:38, Samuel Klein <meta.sj
I can think of a few different server farms that would be glad to run
your process :-) Which script is it that will take 3000 (!) hours to
run?
'Dump the Spanish Wikipedia in HTML'
php dumpHTML.php -d /home/hernan/html3 -s 1 -e 1824046 --checkpoint
/home/hernan/html3/check.txt --show-titles
SJ
Post by Hernan Olivera
Hi everyone. We need your help again.
We finally have a working mirror for generating the static html
version of
Post by Hernan Olivera
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of
processing in
Post by Hernan Olivera
our little sempron server (4 months!).
How many time could it take in Wikimedia's servers?
Thanks
PD: excuse me if you receive this e-mail twice
Post by Jimmy Wales
You are all making me very happy with this important work. I am
sad that
Post by Hernan Olivera
Post by Jimmy Wales
I'm not able to personally roll up my sleeves and help you. :)
But I am
Post by Hernan Olivera
Post by Jimmy Wales
excited to see progress, thank you so much!
--
Hernan Olivera
--
Samuel Klein identi.ca:sj w:user:sj
--
Hernan Olivera
Emmanuel Engelhart
2010-09-14 05:34:28 UTC
Permalink
I have released an up-to-date ZIM file with all main namespace articles
with thumbnails:
http://tmp.kiwix.org/zim/0.9/wikipedia_es_all_09_2010_beta1.zim

You may extract the content in files using zimdump -D
http://openzim.org/Zimdump

You can have a look on it online (served by kiwix-serve, an HTTP server
which reads ZIM files) at :
http://library.kiwix.org:4214/

For windows users which simply want to have an all-in-one solution (ZIM
reader Kiwix + ZIM file + full search index + installer + autorun)
http://download.kiwix.org/portable/wikipedia_es_all.zip

Emmanuel
Post by Kul Takanao Wadhwa
Hernan,
We have some contacts that may be able to give you access to processing
power. Tomasz and I need to more about what your specific needs.
Just contact me and Tomasz directly and we'll go from there.
Kul
El 7 de julio de 2010 04:38, Samuel Klein <meta.sj
I can think of a few different server farms that would be glad to run
your process :-) Which script is it that will take 3000 (!) hours to
run?
'Dump the Spanish Wikipedia in HTML'
php dumpHTML.php -d /home/hernan/html3 -s 1 -e 1824046 --checkpoint
/home/hernan/html3/check.txt --show-titles
SJ
Post by Hernan Olivera
Hi everyone. We need your help again.
We finally have a working mirror for generating the static html
version of
Post by Hernan Olivera
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of
processing in
Post by Hernan Olivera
our little sempron server (4 months!).
How many time could it take in Wikimedia's servers?
Thanks
PD: excuse me if you receive this e-mail twice
Post by Jimmy Wales
You are all making me very happy with this important work. I am
sad that
Post by Hernan Olivera
Post by Jimmy Wales
I'm not able to personally roll up my sleeves and help you. :)
But I am
Post by Hernan Olivera
Post by Jimmy Wales
excited to see progress, thank you so much!
--
Hernan Olivera
--
Samuel Klein identi.ca:sj w:user:sj
--
Hernan Olivera
_______________________________________________
dev-l mailing list
https://intern.openzim.org/mailman/listinfo/dev-l
Samuel Klein
2010-07-07 07:38:05 UTC
Permalink
I can think of a few different server farms that would be glad to run
your process :-) Which script is it that will take 3000 (!) hours to
run?

SJ
Post by Hernan Olivera
Hi everyone. We need your help again.
We finally have a working mirror for generating the static html version of
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of processing in
our little sempron server (4 months!).
How many time could it take in Wikimedia's servers?
Thanks
PD: excuse me if you receive this e-mail twice
You are all making me very happy with this important work.  I am sad that
I'm not able to personally roll up my sleeves and help you. :)  But I am
excited to see progress, thank you so much!
--
Hernan Olivera
--
Samuel Klein identi.ca:sj w:user:sj
Hernan Olivera
2010-07-07 07:35:48 UTC
Permalink
Hi everyone. We need your help again.

We finally have a working mirror for generating the static html version of
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of processing in
our little sempron server (4 months!).

How many time could it take in Wikimedia's servers?


Thanks


PD: excuse me if you receive this e-mail twice
You are all making me very happy with this important work. I am sad that
I'm not able to personally roll up my sleeves and help you. :) But I am
excited to see progress, thank you so much!
--
Hernan Olivera
Ángel González
2010-06-01 21:53:28 UTC
Permalink
Post by Alejandro J. Cura
Hi everyone, we need your help.
We are from Python Argentina, and we are working on adapting our
cdpedia project to make a DVD together with educ.ar and Wikimedia
Foundation, holding the entire Spanish Wikipedia that will be sent
soon to Argentinian schools.
Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.
I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.
Following is an email from Hernán describing the problems he's found.
thanks!
-- alecu - Python Argentina 2010/4/30 Hernan Olivera
to date static html dump for the spanish wikipedia, to use as a basis
for the DVD. I've followed the procedures detailed in the pages below,
that were used to generate the current (and out of date) static html
dumps: 1) installing and setting up a mediawiki instance 2) importing
the xml from [6] with mwdumper 3) exporting the static html with
mediawiki's tool The procedure finishes without throwing any errors,
but the xml import produces malformed html pages that have visible
wikimarkup. We would really need to have a successful import from the
spanish xmls to a mediawiki instance so we can produce the up to date
static html dump. Links to the info I used: [0]
http://www.mediawiki.org/wiki/Manual:Installation_guide/es [1]
http://www.mediawiki.org/wiki/Manual:Running_MediaWiki_on_Ubuntu [2]
http://en.wikipedia.org/wiki/Wikipedia_database [3]
http://www.mediawiki.org/wiki/Manual:Importing_XML_dumps [4]
http://meta.wikimedia.org/wiki/Importing_a_Wikipedia_database_dump_into_MediaWiki
[5] http://meta.wikimedia.org/wiki/Data_dumps [6]
http://dumps.wikimedia.org/eswiki/20100331/ [7]
http://www.mediawiki.org/wiki/Alternative_parsers (among others)
Cheers, --
Hola Hernán,

You may have used one of the corrupted dumps. See
https://bugzilla.wikimedia.org/show_bug.cgi?id=18694
https://bugzilla.wikimedia.org/show_bug.cgi?id=23264

Otherwise, did you install parserfunctions and other extensions needed?
Hernan Olivera
2010-07-07 07:26:27 UTC
Permalink
Hi everyone. We need your help again.

We finally have a working mirror for generating the static html version of
ESWIKI we need for Cd-Pedia using DumpHTML extension.
But it seems that the process will take about 3000 hours of processing in
our little sempron server (4 months!).

How many time could it take in Wikimedia's servers?


Thanks

(This is intentional top-posting to update quickly the situation)
Post by Alejandro J. Cura
Post by Alejandro J. Cura
Hi everyone, we need your help.
We are from Python Argentina, and we are working on adapting our
cdpedia project to make a DVD together with educ.ar and Wikimedia
Foundation, holding the entire Spanish Wikipedia that will be sent
soon to Argentinian schools.
Hernán and Diego are the two interns tasked with updating the data
that cdpedia uses to make the cd (it currently uses a static html dump
dated June 2008), but they are encountering some problems while trying
to make an up to date static html es-wikipedia dump.
I'm ccing this list of people, because I'm sure you've faced similar
issues when making your offline wikipedias, or because maybe you know
someone who can help us.
Following is an email from Hernán describing the problems he's found.
thanks!
-- alecu - Python Argentina 2010/4/30 Hernan Olivera
to date static html dump for the spanish wikipedia, to use as a basis
for the DVD. I've followed the procedures detailed in the pages below,
that were used to generate the current (and out of date) static html
dumps: 1) installing and setting up a mediawiki instance 2) importing
the xml from [6] with mwdumper 3) exporting the static html with
mediawiki's tool The procedure finishes without throwing any errors,
but the xml import produces malformed html pages that have visible
wikimarkup. We would really need to have a successful import from the
spanish xmls to a mediawiki instance so we can produce the up to date
static html dump. Links to the info I used: [0]
http://www.mediawiki.org/wiki/Manual:Installation_guide/es [1]
http://www.mediawiki.org/wiki/Manual:Running_MediaWiki_on_Ubuntu [2]
http://en.wikipedia.org/wiki/Wikipedia_database [3]
http://www.mediawiki.org/wiki/Manual:Importing_XML_dumps [4]
http://meta.wikimedia.org/wiki/Importing_a_Wikipedia_database_dump_into_MediaWiki
Post by Alejandro J. Cura
[5] http://meta.wikimedia.org/wiki/Data_dumps [6]
http://dumps.wikimedia.org/eswiki/20100331/ [7]
http://www.mediawiki.org/wiki/Alternative_parsers (among others)
Cheers, --
Hola Hernán,
You may have used one of the corrupted dumps. See
https://bugzilla.wikimedia.org/show_bug.cgi?id=18694
https://bugzilla.wikimedia.org/show_bug.cgi?id=23264
Otherwise, did you install parserfunctions and other extensions needed?
--
Hernan Olivera
Loading...