#prometheus-metrics #metrics #prometheus #open-telemetry

macro autometrics-macros

Easily add metrics to your code that actually help you spot and debug issues in production. Built on Prometheus and OpenTelemetry.

17 releases (2 stable)

1.0.1 Feb 12, 2024
1.0.0 Dec 1, 2023
0.6.0 Aug 8, 2023
0.5.0 Jun 2, 2023
0.3.2 Mar 22, 2023

#728 in Debugging

Download history 1016/week @ 2023-12-23 1688/week @ 2023-12-30 2578/week @ 2024-01-06 3115/week @ 2024-01-13 2553/week @ 2024-01-20 2506/week @ 2024-01-27 3305/week @ 2024-02-03 2726/week @ 2024-02-10 2544/week @ 2024-02-17 2868/week @ 2024-02-24 2864/week @ 2024-03-02 2920/week @ 2024-03-09 2928/week @ 2024-03-16 2961/week @ 2024-03-23 2292/week @ 2024-03-30 2340/week @ 2024-04-06

10,984 downloads per month
Used in 7 crates (via autometrics)

MIT/Apache

33KB
576 lines

GitHub_headerImage

Documentation Crates.io Discord Shield

Metrics are a powerful and cost-efficient tool for understanding the health and performance of your code in production. But it's hard to decide what metrics to track and even harder to write queries to understand the data.

Autometrics provides a macro that makes it trivial to instrument any function with the most useful metrics: request rate, error rate, and latency. It standardizes these metrics and then generates powerful Prometheus queries based on your function details to help you quickly identify and debug issues in production.

Benefits

  • #[autometrics] macro adds useful metrics to any function or impl block, without you thinking about what metrics to collect
  • 💡 Generates powerful Prometheus queries to help quickly identify and debug issues in production
  • 🔗 Injects links to live Prometheus charts directly into each function's doc comments
  • 📊 Grafana dashboards work without configuration to visualize the performance of functions & SLOs
  • 🔍 Correlates your code's version with metrics to help identify commits that introduced errors or latency
  • 📏 Standardizes metrics across services and teams to improve debugging
  • ⚖️ Function-level metrics provide useful granularity without exploding cardinality
  • ⚡ Minimal runtime overhead

Advanced Features

See autometrics.dev for more details on the ideas behind autometrics.

Example + Demo

use autometrics::autometrics;

#[autometrics]
pub async fn create_user() {
  // Now this function produces metrics! 📈
}

Here is a demo of jumping from function docs to live Prometheus charts:

https://github.com/autometrics-dev/autometrics-rs/assets/3262610/966ed140-1d6c-45f3-a607-64797d5f0233

Quickstart

  1. Add autometrics to your project:

    cargo add autometrics --features=prometheus-exporter
    
  2. Instrument your functions with the #[autometrics] macro

    use autometrics::autometrics;
    
    // Just add the autometrics annotation to your functions
    #[autometrics]
    pub async fn my_function() {
      // Now this function produces metrics!
    }
    
    struct MyStruct;
    
    // You can also instrument whole impl blocks
    #[autometrics]
    impl MyStruct {
      pub fn my_method() {
        // This method produces metrics too!
      }
    }
    
    Tip: Adding autometrics to all functions using the tracing::instrument macro

    You can use a search and replace to add autometrics to all functions instrumented with tracing::instrument.

    Replace:

    #[instrument]
    

    With:

    #[instrument]
    #[autometrics]
    

    And then let Rust Analyzer tell you which files you need to add use autometrics::autometrics at the top of.

    Tip: Adding autometrics to all pub functions (not necessarily recommended 😅)

    You can use a search and replace to add autometrics to all public functions. Yes, this is a bit nuts.

    Use a regular expression search to replace:

    (pub (?:async)? fn.*)
    

    With:

    #[autometrics]
    $1
    

    And then let Rust Analyzer tell you which files you need to add use autometrics::autometrics at the top of.

  3. Export the metrics for Prometheus

    For projects not currently using Prometheus metrics

    Autometrics includes optional functions to help collect and prepare metrics to be collected by Prometheus.

    In your main function, initialize the prometheus_exporter:

    pub fn main() {
      prometheus_exporter::init();
      // ...
    }
    

    And create a route on your API (probably mounted under /metrics) that returns the following:

    use autometrics::prometheus_exporter::{self, PrometheusResponse};
    
    /// Export metrics for Prometheus to scrape
    pub fn get_metrics() -> PrometheusResponse {
      prometheus_exporter::encode_http_response()
    }
    
    For projects already using custom Prometheus metrics

    Configure autometrics to use the same underlying metrics library you use with the feature flag corresponding to the crate and version you are using.

    [dependencies]
    autometrics = {
      version = "*",
      features = ["prometheus-0_13"],
      default-features = false
    }
    

    The autometrics metrics will be produced alongside yours.

    Note

    You must ensure that you are using the exact same version of the library as autometrics. If not, the autometrics metrics will not appear in your exported metrics. This is because Cargo will include both versions of the crate and the global statics used for the metrics registry will be different.

    You do not need to use the Prometheus exporter functions this library provides (you can leave out the prometheus-exporter feature flag) and you do not need a separate endpoint for autometrics' metrics.

  4. Run Prometheus locally with the Autometrics CLI or configure it manually to scrape your metrics endpoint

  5. (Optional) If you have Grafana, import the Autometrics dashboards for an overview and detailed view of the function metrics

API Docs

Examples

Open in Gitpod

To see autometrics in action:

  1. Install prometheus locally or download the Autometrics CLI which will install and configure Prometheus for you locally.

  2. Run the complete example:

    cargo run -p example-full-api
    
  3. Hover over the function names to see the generated query links (like in the image above) and view the Prometheus charts

Benchmarks

Using each of the following metrics libraries, tracking metrics with the autometrics macro adds approximately:

  • prometheus-0_13: 140-150 nanoseconds
  • prometheus-client-0_21: 150-250 nanoseconds
  • metrics-0_21: 550-650 nanoseconds
  • opentelemetry-0_20: 1700-2100 nanoseconds

These were calculated on a 2021 MacBook Pro with the M1 Max chip and 64 GB of RAM.

To run the benchmarks yourself, run the following command, replacing BACKEND with the metrics library of your choice:

cargo bench --features prometheus-exporter,BACKEND

Contributing

Issues, feature suggestions, and pull requests are very welcome!

If you are interested in getting involved:

Dependencies

~2.4–4MB
~71K SLoC