#agent #tether #lidar #system #clustering #consolidator #lidar2-d

yanked tether-lidar2d-consolidation-rs

Tether Lidar2D Consolidator Agent, Rust edition

0.3.1 Nov 7, 2023

#6 in #tether

Download history 3/week @ 2024-02-24 12/week @ 2024-03-30 37/week @ 2024-04-06 2/week @ 2024-04-13

51 downloads per month

MIT license

68KB
1.5K SLoC

Tether Lidar2D Consolidator, in Rust

This is more or a less a direct port of the original NodeJS Agent (here referred to as "the OG Agent" ✌️) into Rust 🦀.

All the essential features of the OG Agent have been implemented, but the structure of the application does not attempt to replicate the original.

The two main goals were

  • Learn to use Rust in a real-world application, following an existing codebase which was well known and where the fundamental engineering challenges had already been "solved"
  • Produce a usable version of the Lidar Consolidator Agent that was faster and lighter (e.g. on RAM) which could be deployed to lower-resource platforms (e.g. Raspberry Pi) without putting as much strain on the system

Dev dependencies

Paho Eclipse MQTT has some (non-Rust) dependencies of its own. On Mac, you might need to the following:

brew install openssh cmake

And on Linux:

sudo apt install libssl-dev build-essential cmake

Basic usage

If using cargo, you can simply use

cargo run

Or, for example running in release (recommended) and with some command-line arguments:

cargo run --release -- --loglevel debug

Alternatively, run the compiled binary (assuming you've already run cargo build --release), typically located in ./target/release/tether-lidar2d-consolidation-rs

Command-line configuration

You can see a full list of available command-line arguments by appending --help onto your executing command, e.g. cargo run -- --help

Options are deliberately made to be almost identical to those used in the OG Agent, with a few notable exceptions:

  • Here we use a flag --perspectiveTransform.includeOutside instead of setting a boolean ignoreOutside. This reflects the default usage better and makes more sense when using a flag (which defaults to false if not explicitly appended)
  • For Tether, you can optionally provide --tether.host, --tether.username, --tether.password if these differ from the defaults
  • There are no options relating to autoBroadcastConfig as there is no need for this behaviour; we save and (re)publish configuration essentially every time it changes, on startup and when requested

Notes on Libraries

MQTT Client

Initially we tried using mqtt-rs, as it seems relatively simple to use, but in the future mqttrs might be "better".

For now have settled on paho-mqtt since it seems well-supported and provides examples in realistic scenarios (especially async).

MessagePack encoding/decoding

rmp_serde is useful for both JSON and MsgPack serialisation/deserialisation. We might not be taking full advantage of zero-copy operations everywhere, but this will take a little more time to figure out.

In the beginning we tried msgpack-simple which warns that it is "not as performant as static solutions" but was much easier to use as a starting point.

Clustering

We tried the library kddbscan, but although this may well be more "accurate" it seems to run far too slowly. In any case, this is a very different algorithm from the DBSCAN used in the OG Agent.

We then settled for the more humble (but apparently much more performant) petal-clustering. This in turn requires something called ndarray which seems very similar (and likely based on) numpy for Python.

For now, we use the DBSCAN method as per the OG Agent, but in future it might be worth tested the other supported mode in this library, HDbscan which may be faster still (see the paper).

Another possibility might be the library linfa-clustering.

JSON serialisation / deserialisation

We are using a combination of the libraries serde and serde_json which makes it easy to handle JSON in various ways - including strongly typed corresponding to Rust types/structs, which is what we need here in the case of our Config loading/saving.

Perspective transformation

We are attempting to do a "quad to quad projection" from the ROI to a normalised "square" output quad, similar to perspective-transform as per the OG Agent.

So far

Finally, used a combination of ndarray (which was already installed, to support the clustering calculations) and nalgebra.

Logging

We are using log and env-logger. Log level has been set to INFO by default, but can be overridden, for example by prefixing with an environment variable, e.g.

RUST_LOG=debug cargo run

Command-line configuration

We are using clap which does command-line argument parsing only (no use of files, environment variables, etc.)

Something like more-config could be useful, since it includes functionality similar to the rc package for NodeJS.

Notes on implementation

  • As with OG Agent, the serial string for the LIDAR device is extracted from the topic, specifically the agentIdOrGroup part in lidar2d/{agentIdOrGroup}/scans
  • The samples are copied from the array of arrays; each "sample" is an array with the elements [angle, distance] and sometimes [angle, distance, quality] (although the latter is not handled yet). The samples are converted into points and the list(vector) of all converted points are "inserted" (i.e. replaced, if key already exists) into the hashmap which represents all LIDAR devices (with serial strings as keys).
    • There is possibly some allocation / copying of data going on; this needs to be reduced as far as is practical
    • There are some of instances of .unwrap() in this process; errors should be handled more carefully in some cases

TODO

  • Add (optional) tether-tracking-smooth functionality, built-in
  • Presence Zones should be defined within the real space (more sensible calibration) not the normalised tracking space
  • Presence Zones should be definable for shapes other than rectangles
  • Smoothing should also remove duplicate existing points that should be merged (do not leave "stale" points behind)
  • Ensure that this compiles on Raspberry Pi, then cross-compiles (e.g. from Mac)
  • Retain messages, for config publish - remove the need for "request config" topic
  • Debug loglevel should suppress MQTT log messages (too verbose)
  • Allow all critical settings (clustering, smoothing) to be updated live on command
  • Print warning/error (and optionally panic?) if no data received from known devices after timeout
  • If no Lidar Device configuration file is found, create one
  • Maintain "state" (config data) and publish this on startup this should allow the visualiser to start showing scan data
  • Use rmp-serde for MessagePack instead of msgpack-simple
  • Read/write config data from/to disk
  • Transform incoming scan points: rotation and position/offset
  • Apply Scan Mask Thresholds on incoming samples
  • Apply maxClusterSize filtering
  • Handle ROI, transformation (warping)
  • Allow incoming config (devices, ROI) to be saved via "saveLidarConfig"
  • Allow AutoMaskSampler to be created on request
  • Should test with multiple sensors
  • Tracking output should apply ignoreOutside, ignoreOutsideMargin logic
  • If receiving an unknown / unconfigured device, add it with some defaults to Config
  • Load/override some settings (e.g. Tether, clustering) from command-line args / defaults
  • Generate agent UUID automatically
  • Use "real" Tether Agent crate - although currently this is not set up for async
  • Move potentially blocking processes (e.g. Config write to disk) into a separate thread or non-blocking functions
  • Close the client properly on quit, so that the queue is also properly destroyed
  • Currently, if "scan samples" are tuples of the form (f64,f64) i.e. (angle,distance), then the system will panic if quality is included. This implies we either need an array without fixed length, or simply drop the quality "field" altogether
  • Possible memory leak when clearing and/or setting new AutoMask Samplers

Useful resources

Dependencies

~27–39MB
~720K SLoC