Dbnary Extractor allows you to extract a lexical network from a wiktionary dump file. It currently supports extraction of French, English, German and Portuguese language editions of wiktionary. The extracted data is structured as a lexical network in RDF (most formats are supported). The organization of the network is based on the LEMON model.
Wiktionary support uses wiktionary dump files. Go to Wikimedia Dumps server and download the appropriate files. Currently Blexisma supports English, French, German and Portuguese wiktionaries
wget http://dumps.wikimedia.org/enwiktionary/latest/enwiktionary-latest-pages-articles.xml.bz2 wget http://dumps.wikimedia.org/frwiktionary/latest/frwiktionary-latest-pages-articles.xml.bz2 wget http://dumps.wikimedia.org/dewiktionary/latest/dewiktionary-latest-pages-articles.xml.bz2 wget http://dumps.wikimedia.org/ptwiktionary/latest/ptwiktionary-latest-pages-articles.xml.bz2
Then, uncompress the files using a bz2 uncompressor and convert them to UTF-16 encoding. Users with linux like system could use the iconv command line:
iconv -f UTF-8 -t UTF-16 < ...-articles.xml > xxwiktionary.utf16.xml (replace xx with fr, en, de or pt).
If your system does not provide an iconv utility. You may use the provided Iconv command line.
For the rest of this manual, we will assume that these files are available as /path/to/dumps/xxwiktionary.utf16.xml. and that you have write permissions to the /path/to/dumps/ directory.
All extracted versions are available at http://kaiko.getalp.org/dbnary/