|0.1.5||Mar 16, 2023|
|0.1.4||Jul 25, 2020|
|0.1.3||May 28, 2020|
120 downloads per month
Used in 4 crates (2 directly)
- Incorporate download counts into report
- Make things prettier and more visual - that way we can try again for a come-back :D
- See why RipGrep doesn't get any suggestions
- More reporting - right now the context gathering to see how much time is spent where is neglected.
- Suggest 'top-level' globs like
/README.mdif we know the matched file is on the top-level. Otherwise the pattern
README.mdwill actually match
- Count negation patterns in includes and excludes. The latter don't seem to be working, and if nobody is using them, Cargo can either make it work or reject them properly. Maybe. Maybe first create an issue for that and see what they think.
- On chunk download timeout, don't restart, but resume the download where it left off
- resilience: protect against ThreadPanics - they prevent the program from shutting down
- Futures has a wrapper to catch panics, even though we don't use it yet. A panic only brings down the future that panics, not the entire program.
- Graceful shutdown on Ctrl+C
- The current implementation relies on the database to handle aborted writes, and is no problem for that reason. However, it would be nice to have A well-behaving program.
- Git is slowing down report generation, as sending them to git and creating objects is slow. We could possibly multi-thread this by creating lose objects ourselves and sending these into an aggregator which puts them into the index. This is only interesting during the very first report generation though, so probably not worth it.
- Have each sub-section of criner use their own error type, which are aggregated in the crate level error. That way, individual errors will be smaller.
- Parse CSV files separately and index rows and fields - from there build everything on the fly without having to allocate and copy strings.
- probably warrants a different crate, and will really only be done if the 500MB budget isn't sufficient, that is things don't run on the Pie III
- futures::ThreadPools - panicking futures crash only one thread
- long-running futures need error and potentially panic recovery. Futures has a panick catcher that could be useful.
- sqlite needs a lot of massaging to work acceptably in concurrent applications. Takeaway: WAL_mode, and when writting, always use immediate transactions when writing. Retry yourself while waiting and set a busy handler which waits.
- Trying to optimize output HTML for git by prettifying failed - I just couldn't see it improve anything. For debugging HTML, it's easiest to use the browser.
When migrating to Sqlite
- is really not suited for many concurrent writers - you have to prepare for database locked errors, and the busy_handler doesn't help most of the time.
- writing many small objects is slow, and can only be alleviated with prepared statements which are not always feasible or nice to use with a persistence design inspired by sled. To alleviate, the whole application must embrace Sqlite and work with statements directly.
- Working with the lifetimes associated with transactions is a necessary evil, but it is painful too when trying to refactor anything! I just don't understand anymore what it tries to do, and have the feeling the compiler is confused itself (as in theory, there is no issue).
- sled databases are about 4 times bigger than an Sqlite database with the same content, and it would read about 1.2GB of a 14GB database at startup.
- sled is easy to handle in a threaded/concurrent environment, but iteration isn't possible across awaits as it's not sync
- Sqlite is not sync nor is it send, so it needs more treatment before it can be used with spawened futures
- Zero-copy is straigforward with Sled as it provides IVec structs, which are handles into an LRU which is the backing store.
- In retrospect, I would consider zero-copy a nice experiment, but also a premature optimization. It costs additinoal effort and when done from the beginning, you don't even know how much time is actually saved through that.