34 releases (breaking)

new 0.25.1 Jan 3, 2025
0.25.0 Dec 19, 2024
0.24.0 Nov 12, 2024
0.21.1 Jul 30, 2024
0.4.0 Jul 9, 2022

#392 in Machine learning

Download history 34/week @ 2024-09-18 132/week @ 2024-09-25 66/week @ 2024-10-02 17/week @ 2024-10-09 8/week @ 2024-10-16 14/week @ 2024-10-30 83/week @ 2024-11-06 61/week @ 2024-11-13 14/week @ 2024-11-20 5/week @ 2024-11-27 10/week @ 2024-12-04 6/week @ 2024-12-11 123/week @ 2024-12-18 147/week @ 2025-01-01

282 downloads per month
Used in 2 crates

Apache-2.0

345KB
7K SLoC

Mixture of experts

crates.io docs

egobox-moe provides a Rust implementation of mixture of experts algorithm. It is a Rust port of mixture of expert of the SMT Python library.

The big picture

egobox-moe is a library crate in the top-level package egobox.

Current state

egobox-moe currently implements mixture of gaussian processes provided by egobox-gp:

  • Clustering (linfa-clustering/gmm)
  • Hard recombination / Smooth recombination
  • Gaussian processe model choice: specify regression and correlation allowed models

Examples

There is some usage examples in the examples/ directory. To run, use:

$ cargo run --release --example clustering

License

Licensed under the Apache License, Version 2.0 http://www.apache.org/licenses/LICENSE-2.0

Dependencies

~15–34MB
~502K SLoC