10 releases (3 stable)
2.0.1 | Nov 6, 2024 |
---|---|
2.0.0 | Nov 2, 2024 |
1.0.0 | Nov 2, 2024 |
0.1.6 | Oct 30, 2024 |
#289 in Algorithms
905 downloads per month
110KB
2.5K
SLoC
Jaime's Artificial Inteligence and Machine learning Engine
J.a.i.m.e., pronounced as /hɑːɪmɛ/, is a all purpose ergonomic gradient descent engine. It can configure ANY* and ALL** models to find the best fit for your dataset. It will magicaly take care of the gradient computations with little effect on your coding style.
* not only neuronal
** derivability conditions apply
Concepts and explanation
- Input: For our purposes the input of our Model will be a vector of floating point numbers
- Output: For our purposes the output of our Model will be a vector of floating point numbers
- Dataset: a set of input-output pairs. Jaime will reconfigure the model to aproximate the behabiour described in the dataset.
- Model: a function that maps from input to ouput using a set of configuration parameters that define its behabiour. For our purposes small changes in the parameters should translate to small changes in the behabiour of the function. Examples of suitable models:
- Polinomial functions: Defined as
y = P_0 * x^0 + P_1 * x^1 + ... + P_n * x^n
. The vector[x]
will be our input, the vector[y]
will be our output, The vector[P_0, P_1, ... ,P_2]
will be our parameter vector. An example of this crate for this precise case can be found here - Neuronal networks: In their most basic form they are defined as consecutive matrix multiplications with delinearization steps in between. The classical meaning of parameters, input and output for a NN matches the concepts used in this crate. An example of this crate for this precise case can be found here
- Polinomial functions: Defined as
If you are able to define a model this crate will happily apply gradient descent to find some local minumum that aproximates the behabiour defined in the dataset.
Examples
To make sure this crate was as usable and performant as posible I've also implemented a few exercises that use its functions.
-
WIP Asteroid shape deduction
If you belive you understanding of what my crate does is clear I encourage you to study the voronoi image aproximator.
Geeky internal wizardry
Gradient calculation
If you are a little math savy and know how gradient descent works you may be wondering how am I able to do the partial derivatives for the parameters without knowing beforehand what operations will the model perform. The solution relies on Forward Mode Automatic Differentiation using dual numbers. Jaime will require you to define a generic function that manipulates a vector of float-oids and returns a vector of float-oids. That function will later be instanciated with a custom dual number type, that will allow me hijack the mathematic operations and keep track of the necesary extra data.
Rust, specificaly rust's generics and trait system, is perfect for this task. I can unambiguosly define what a float-oid is to rust as a set of traits that overload operators and other functionality.
After that the only thing remaining is to follow the calculated gradient towards victory, success and greatness.
Gradient following
The field of gradient descent has been thoroughly studied to make it kind of good. The naive aproach is prone to local minima and wasted time, in order to tacle this problems many gradient descent optimizers exist. j.a.i.m.e implements a few, more implementations are very very welcome! At this point the following optimizers are aviable:
-
In house naive asymptotic step gradient (AsymptoticGradientDescentTrainer). It will move the parameters towards the gradient direction some amount that decreases the cost. It will reduce the size as needed. This optimizer has the advantage of convergence.
-
ADAM optimizer, best explained over in this wikipedia page, it takes into account the previous steps to nugde the current step. It provides some protection towards local minima.
-
Genetic optimizer, explained here allows to train non smooth cost functions, at the cost of poor eficiency.
-
WIP: WOA: https://www.geeksforgeeks.org/whale-optimization-algorithm-woa/ https://www.sciencedirect.com/science/article/abs/pii/S0952197621003997
-
WIP: Analitical optimizer : resolve the gradient equation to find the values that make it cero
Usage Documentation
Comming soon, for now try having a look at the examples.
Contributing
Yes please. Make a PR to this repo and I will happily merge it.
A note on optimization
I heavily used Samply for profiling during this project.
Dependencies
~5–12MB
~141K SLoC