spidernik84
|
2016-02-05
, 09:41
|
Posts: 27 |
Thanked: 35 times |
Joined on Jan 2016
@ Sweden
|
#271
|
|
2016-02-05
, 10:23
|
Posts: 105 |
Thanked: 205 times |
Joined on Dec 2015
@ Spain
|
#272
|
|
2016-02-06
, 23:10
|
|
Posts: 654 |
Thanked: 2,368 times |
Joined on Jul 2014
@ UK
|
#273
|
|
2016-02-07
, 18:22
|
Posts: 86 |
Thanked: 362 times |
Joined on Dec 2007
@ Paris / France
|
#274
|
The Following 14 Users Say Thank You to eber42 For This Useful Post: | ||
|
2016-02-19
, 10:13
|
Posts: 105 |
Thanked: 205 times |
Joined on Dec 2015
@ Spain
|
#275
|
Make sure you use the real "N" key and not "Ñ" (this is a known limitation that has been discussed earlier, and I have some ideas on how to fix it).
If it still does not work, send me your language files and logs (you have to enable them in the application and they can be found in ~/.local/share/okboard). And do not expect a quick answer (I did not even start to work on the weeks old transparency issue).
BTW, last week I worked on the engine that processes swipes. Now it is not that much more accurate, but it does no more assign bad scores to expected results. Next week I will start to re-implement the prediction engine (this is the part which handle language model and learning. At the moment it is completely broken) and after I will go back to regular bug fixing.
|
2016-03-26
, 12:15
|
Posts: 10 |
Thanked: 12 times |
Joined on Jun 2013
|
#276
|
The steps to do that are these:
-You need a linux environment (I'm using Archlinux, but Ubuntu or some other works too)
- You need to download the tarball first: http://git.tuxfamily.org/okboard/okb...master.tar.bz2 and uncompress it at your /home directory
- You need the dictionaries. I take it from https://github.com/titoBouzout/Dictionaries but it needs to be adjusted, so I attach the file already processed (see Spanish.dic.txt.zip on this post)
-You need the corpora files of your language (e.g. Spanish)
http://corpora2.informatik.uni-leipzig.de/download.html
http://www.cs.upc.edu/~nlp/wikicorpus/
http://opus.lingfil.uu.se/OpenSubtitles2016.php
http://www.lllf.uam.es/ESP/Corlec.html
https://tatoeba.org/spa/downloads
- You need the "aspell-es" package (in case of Spanish) instaled from the repos of your distro.
- You need "lbzip2" package installed in your system too.
-You need "rsync" installed in your system.
-You need "QT5" installed in your system.
- Now you need to create a folder somewhere and put the dictionary inside (e.g. /home/username/okboard/langs)
-If you have several corpora files, then:
- Open a terminal windowCode:cat file1 file2 file3 file4 file5 > corpus-es.txt
- And set the two environment variables:
Code:export CORPUS_DIR=/home/username/okboard/langs- You can see those variables withCode:export WORK_DIR=/home/username/okboard/langs
if you're curiousCode:echo $VARIABLE_NAME
- You need to compress the file (Spanish.dic.txt) you put before in /home/username/okboard/langs:
- Now should be named corpus-$LANG.txt.bz2 In our case: corpus-es.txt.bz2 because of SpanishCode:bzip2 Spanish.dic.txt
- There should be a single file inside.
- The next thing is to do is to move in okboard files inside the same Terminal window in our case "/home/username/okb-engine-master/". Here is the okboard's source code.
- In 'db' folder you must create a lang-es.cf file first. You can copy it from another .cf file in the same folder (e.g. copy lang-en.cf and rename it into lang-es.cf)Code:cd /home/username/okb-engine-master/
-And left only ASCII characteres on those files:
- ExecuteCode:lbzip2 -d < corpus.txt.bz2 | clean_corpus.py | lbzip2 > new_corpus.txt.bz2("es" in case of Spanish)Code:db/build.sh es
- After this, the script create the dictionaries for OKBoard with next list of files:
add-words-fr.txt
es-predict.dict
lang-fr.cf
clusters-es.log
es-test.txt.bz2
lang-nl.cf
clusters-es.txt
es.tre
predict-es.db
corpus-es.txt.bz2
grams-es-full.csv.bz2
predict-es.ng
db.version
grams-es-learn.csv.bz2
predict-es.rpt.bz2
es-full.dict
grams-es-test.csv.bz2
predict-es.txt.bz2
es-full.tre
lang-en.cf
words-es.txt
es-learn.txt.bz2
lang-es.cf
- So, now we have the Spanish dictionary created.
After this. I don't know what to do with these files. So any help is welcome
|
2016-03-26
, 17:10
|
Posts: 27 |
Thanked: 35 times |
Joined on Jan 2016
@ Sweden
|
#277
|
5) I think Finnish spellchecking doesn't use aspell, but Malaga based Voikko: http://voikko.puimula.org/ and if I'm not misunderstood, voikko is used by ispell for example. But how to get that finnish dictionary file is somehow unclear to me.
|
2016-03-26
, 17:54
|
Posts: 635 |
Thanked: 1,535 times |
Joined on Feb 2014
@ Germany
|
#278
|
|
2016-03-29
, 07:14
|
Posts: 10 |
Thanked: 12 times |
Joined on Jun 2013
|
#279
|
The Following 2 Users Say Thank You to uggeli For This Useful Post: | ||
|
2016-04-03
, 18:56
|
Posts: 25 |
Thanked: 7 times |
Joined on Nov 2012
|
#280
|
Tags |
bettertxtentry, huntnpeck sucks, okboard, sailfish, swype |
|