jnalanko / vomm Goto Github PK
View Code? Open in Web Editor NEWSpace-efficient variable-order Markov models
Space-efficient variable-order Markov models
reconstruct.cpp:134:75: error: no matching function for call to 'Global_Data::load_all_from_disk(std::__cxx11::string&, std::__cxx11::string&, bool&, bool)'
G.load_all_from_disk(C.modeldir, C.filename, C.run_length_coding, true);
^
In file included from Parent_Support.hh:10:0,
from reconstruct.cpp:18:
globals.hh:120:10: note: candidate: void Global_Data::load_all_from_disk(std::__cxx11::string, std::__cxx11::string, bool)
void load_all_from_disk(string directory, string filename_prefix, bool load_bibwt) {
could you please also offer a flag with which the user can print info of all contexts? e.g.:
if i don't specify --escapeprob with --lin-scoring i get:
score_string_optimized: score_string.cpp:85: void Scoring_Config::assert_all_ok(): Assertion `escapeprob != -1' failed.
but i don't think the competitors' paper uses escape probability?
i pulled the latest version from git and i started to experiment with the --store-depths flag. but i get the following error in many datasets:
Sat Mar 31 17:54:45 2018 Starting to build the model
Sat Mar 31 17:54:45 2018 Building the BiBWT
Sat Mar 31 17:55:10 2018 Building reverse suffix tree BPR and pruning marks
Sat Mar 31 17:57:03 2018 Pruning marks is all ones
Sat Mar 31 17:58:04 2018 Marking contexts and storing string depths of maxreps
Sat Mar 31 17:59:24 2018 Building the BPR of contexts only
Sat Mar 31 17:59:30 2018 Computing reverse BWT (todo: reuse the bibwt)
Sat Mar 31 17:59:42 2018 Writing model to directory: /tmp/cunial
terminate called after throwing an instance of 'std::runtime_error'
what(): Error writing to disk: /tmp/cunial/protein.txt.rev_bwt_bwt.dat
Don't do "return next();"
if there is no pruning, there is no need for pruning marks and the support structures at all.
The pruning marks will be a bit vector contain only ones. It's still built because the code is simpler
that way, but it should not be a big deal to change that.
this is important since its scoring function if very simple, and it will be probably faster than ours.
since its scoring is so simple, we might also end up noticing that we don't need the context data structures and probably just revST topology and rlbwt to implement it.
last week i was able to solve the problem with the cluster, and i started building the indexes on all bacteria. after five days and 13 hours, one of the indexes is still not built (the one with depth pruning at 8, the smallest i tried), and it seems stuck to "Building the BiBWT".
could you please provide a switch that allows one to reuse parts of an index at depth, say, 16, which i already have, without recomputing everything from scratch?
i think this is really crucial in order to index large files in practice.
Buffer input reading and optimize scoring itself.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.