8 releases (4 breaking)

0.5.2 Jul 11, 2024
0.5.1 Jul 11, 2024
0.5.0 Jun 25, 2024
0.4.1 Jun 25, 2024
0.1.0 Jun 8, 2023

#230 in Debugging

Download history 135/week @ 2024-06-07 19/week @ 2024-06-14 282/week @ 2024-06-21 19/week @ 2024-06-28 159/week @ 2024-07-05 44/week @ 2024-07-12 36/week @ 2024-09-13 28/week @ 2024-09-20

64 downloads per month

GPL-3.0-only

44KB
610 lines

The service-skeleton is a Rust crate that provides the "bare bones" of a program which is intended to run as a service -- a long-lived, non-(directly)-human-interactive program that typically provides some sort of functionality to a larger system. It provides:

  • Initialisation and configuration of logging (partially implemented);
  • Configuration parsing and management, via environment variables;
  • Supervision of subunits of functionality, automatically restarting them if they crash (partially impemented);
  • A built-in Prometheus-compatible (OpenMetrics) metrics server and hooks for easily declaring and using metrics;

Features which are intended to be added in the future, but don't exist yet:

  • Signal handling, including built-in support for dynamic log-level adjustment, backtrace dumping, and graceful shutdown;
  • OpenTracing support;
  • HTTP-based introspection and control.

The general philosophy of service-skeleton is to be secure-by-default, provide features that have been found near-universally useful for service programs in modern deployment scenarios, and to prefer convention over configuration.

Installation

It's published on crates.io, so a cargo add service-skeleton should work.

Usage

In its simplest form, which enables most of the available features, you can make your main function look pretty much exactly like this:

# use std::time::Duration;
# // Yes, this is cheating
# fn sleep(_: Duration) { std::process::exit(0) }
use service_skeleton::service;

fn main() {
    service("SayHello").run(|_cfg: ()| say_hello());
}

fn say_hello() {
    println!("Hello world!");
    sleep(Duration::from_secs(5));
}

This example will cause the program to print "Hello world!" every five seconds to your terminal, until you stop it with Ctrl-C (or some other way, like a kill -9). Not the most exciting of services, but it does demonstrate one of service-skeleton's basic functions: service supervision.

The closure that you provide to service_skeleton::Service::run shouldn't ordinarily terminate -- the idea is that it'll live more-or-less forever, servicing whatever requests come its way. However, if the closure does terminate for any reason (whether via panic or otherwise) the closure will be run again, and the fact of the restart will be logged. Which is a nice segue into the next feature...

Logging / Tracing

One of the things that service-skeleton configures for you is logging and tracing, using the tracing crate's well-established facade. By default, all log messages with severity warn or higher will be printed to stderr with a bunch of related useful information. Again, you don't have to do anything special, just start instrumenting:

# use std::time::Duration;
# fn sleep(_: Duration) { std::process::exit(0) }
use service_skeleton::service;

fn main() {
    service("LogHello").run(|_cfg: ()| say_hello());
}

#[tracing::instrument]
fn say_hello() {
    tracing::info!("Hello, logs!");
    sleep(Duration::from_secs(5));
}

If you already have code that calls the log macros, such as log::info!(), log::debug!(), and so on, don't worry -- we automatically capture all log events and forward them into tracing.

This will print out the log message specified every five seconds. The default logging configuration is that everything at info level or above is logged. If you prefer a different default log level, to log to a file, modify the output format of the log, or set per-module levels, at the moment you'll have to rely on what RUST_LOG can do, but this will be expanded with various configuration "knobs" as demand requires.

Which is as good a time as any to talk about configuration.

Configuration

Most services need some sort of configuration. In line with the 12factor philosophy, service-skeleton encourages you to store your configuration in environment variables.

To declare your configuration, you'll need to declare your configuration struct, like this:

use service_skeleton::ServiceConfig;

#[derive(Clone, ServiceConfig, Debug)]
struct MyConfig {
    /// The name to say hello to
    #[config(default_value = "World")]
    name: String,
}

If you're familiar with clap, then hopefully the approach taken by service-skeleton's config support will feel comfortable, as it took significant inspiration from clap.

As you may have noticed from the previous examples, the configuration gets passed into the closure provided to run, so all you need to do is, in turn, pass that into your say_hello function, and you're off and running (as it were):

# use service_skeleton::ServiceConfig;
# #[derive(Clone, ServiceConfig, Debug)]
# struct MyConfig {
#     #[config(default_value = "World")]
#     name: String,
# }
# use std::time::Duration;
# fn sleep(_: Duration) { std::process::exit(0) }
use service_skeleton::service;

fn main() {
    service("Hello").run(|cfg| say_hello(cfg));
}

fn say_hello(cfg: MyConfig) {
    println!("Hello, {}!", cfg.name);
    sleep(Duration::from_secs(5));
}

By default, this will print "Hello, World!" every five seconds. However, you can now configure who to say hello to, by using an environment variable, like this:

HELLO_NAME=Bobbie cargo run

This will now instead print "Hello, Bobbie!" every five seconds.

The environment variable that service-skeleton will use to try and read the configuration value from is determined by the name of struct member, prefixed with the name of the service (what was passed to service), then turned into all-uppercase. If the environment variable is missing, the default value will be used (if specified), or the program will exit. If the value specified cannot be parsed into a value of the struct member's type, the program will log an error and exit.

Configuration Type Conversion

By default, service-skeleton uses str::parse() to convert the value in the environment variable (or the default_value, if provided and the environment variable is unset) into the type of the field in the configuration struct. Therefore, you can use any type that implements FromStr as the type of a configuration field. If the parsing fails, you'll get a pleasant runtime error.

However, if you have to parse into a type that, for whatever reason, you don't want to create via FromStr, you can instead define a value_parser, like this:

# use service_skeleton::ServiceConfig;
# fn parse_hex<const N: usize>(s: &str) -> Result<[u8; N], String> { Ok([0u8; N]) }
#[derive(Clone, ServiceConfig, Debug)]
struct MyConfig {
    #[config(value_parser = parse_hex::<4>)]
    some_id: [u8; 4],
}

Essentially, whatever is defined on the right-hand-side will be called as a function that takes &str and be expected to return Result<T, impl std::fmt::Display> (note that std::fmt::Display is a supertrait of std::error::Error, so you'll be OK with practically any error-producing parsing function out there, but you can make your own parsing functions return a String, which is so much easier for those ad-hoc parsing functions).

Secrets in Configuration

(I wanted to call this section "Environmental Protection Agency", but it seems someone else already got that name first)

One potential downside of using the process environment for configuration is that it is not necessarily entirely secret. On most systems, other processes with the same UID can read the environment out of /proc/<PID>/environ, procstat -e, and the like. An attacker with RCE can read the environment out of the current process, and the environment also gets passed, by default, to subprocesses. So many ways for environment variables to leak their contents.

The service-skeleton is aware of these problems, and wants to help.

Firstly, if you mark a field as #[config(sensitive)], it will be removed from the environment after it is read, meaning that it won't get passed to subprocesses, and it won't be trivially readable in-process. It looks like this:

# use secrecy::Secret;
# use service_skeleton::ServiceConfig;
#[derive(Clone, ServiceConfig, Debug)]
struct MyConfig {
    #[config(sensitive)]
    secret_name: String,
}

However, making a field as sensitive only really solves the subprocess problem, and to a lesser extent the read-it-from-the-current-process problem. The contents of these environment variables are still available in one way or another in most cases.

Also, some people like to store their application configuration in revision control, because they feel it's better to keep everything in one place. However, storing secrets (private keys, API tokens, and the like) in revision control is... unwise.

To prevent all these problems, we can mark one or more configuration items as #[config(encrypted)], and give the name of a field that specifies the file to read the decryption key from, like this:

# use service_skeleton::ServiceConfig;
#[derive(Clone, ServiceConfig, Debug)]
struct MyConfig {
    #[config(encrypted, key_file_field="secret_key")]
    api_token: String,
    #[config(encrypted, key_file_field="secret_key")]
    location_of_gold_bars: String,
}

The values marked encrypted will be decrypted at runtime, using the key read from a file, whose name is specified in an environment variable derived from the "pseudo-field" given in key_file_field. Ideally, you won't store that key file in revision control, but instead inject it into your application's filesystem at runtime using your provider's secrets management mechanism.

This is a lot of layers of indirection going on, I know... let's have an example.

If your application is called "SuperApp", and is using the MyConfig struct defined above, then the environment variable named SUPER_APP_SECRET_KEY will be consulted when the application starts up, looking for a filename. That filename will be read (relative to the working directory), and the contents parsed as a private key to decrypt the values specified in the SUPER_APP_API_TOKEN and SUPER_APP_LOCATION_OF_GOLD_BARS environment variables.

Encrypting Secrets

The final question is: how do we encrypt these secret values in the first place? For that matter, how do I get a private key?

Enter: a small CLI tool called sscrypt (aka "service-skeleton cryptography"). Using it is intended to be as straightforward as possible:

  1. Install it on your local machine using cargo install --locked sscrypt.

  2. Create a keypair by running sscrypt init <name>, where <name> is any identifier you like (such as prod, stage, or bruce, to keep things clear).

  • The private key will be printed to stdout, which you should copy into your secrets manager, and then forget you ever saw it. Probably best not to do it on a system running Windows Recall, either.
  • The public key will be written to <something>.key, and you can safely commit that to revision control.
  1. To encrypt a secret, run sscrypt encrypt <env var> <name>, where <env var> is the name of the environment variable whose value you wish to encrypt, and <name> is the identifier for your public key.
  • The public key to be used for encryption will be read out of <something>.key in the current working directory.
  • You will be prompted to enter in the value to be encrypted.
  • The value you enter will be encrypted by the public key, in such a way that it can only be used for the environment variable you specified.
  • The encrypted value, which you can safely store in revision control, will be printed to stdout.

By the way, all this magic also works with the FromStr type conversion functionality. So the encrypted secret will be decrypted, then parsed, and the final value of whatever type you specify will end up in the config struct instance ready for use.

Service Metrics

You can't manage what you don't measure. That's why service-skeleton comes with first-class support for Prometheus (aka "OpenMetrics") metrics collection and export.

Using metrics in service-skeleton has three separate parts, each of which we try to make as painless as possible.

  • First, you need to declare the metrics that you use, so that everyone is on the same page about what is being measured.

  • Next, you need to populate the metrics as your application runs, recording values of interest.

  • Finally, the values need to be exposed to the metrics collection server, for processing, display, and alerting.

Declaring metrics is done by configuring service-skeleton before your service starts running. We use the common Builder pattern to configure metrics (as well as everything else in service-skeleton). Thus, if we wanted to have a counter that exposed how many times the service has said hello, it would look like this:

use service_skeleton::service;

fn main() {
    service("InstrumentedHello")
        .counter::<()>("count", "Number of times we've said hello")
        .run(|_cfg: ()| say_hello());
}
# fn say_hello() { std::process::exit(0) }

There are also gauge and histogram methods that declare a metric of those types.

To access your newly-created counter, call the counter function, passing the metric name and label set, and a closure that manipulates the counter as needed:

# use service_skeleton::service;
# fn main() {
#    service("InstrumentedHello")
#        .counter::<()>("count", "Number of times we've said hello")
#        .run(|_cfg: ()| say_hello());
# }
# use std::time::Duration;
# fn sleep(_: Duration) { std::process::exit(0) }
use service_skeleton::metric::counter;

fn say_hello() {
    println!("Hello, Metrics!");
    counter("count", &(), |m| { m.inc(); });
    sleep(Duration::from_secs(5));
}

The reference to () in the counter call (and as the type in the counter declaration, ::<()>) refers to the label set; it is possible to provide arbitrary types as the labels for metrics calls. See the prometheus-client docs for more information on custom label set types.

Finally, you need to be able to scrape your metric to get it into your monitoring system. To do that, service-skeleton comes with a built-in metrics server, but for security it's turned off by default. It's simple to enable it, though: just set the INSTRUMENTED_HELLO_METRICS_SERVER_PORT environment variable, then you can hit the metrics server:

INSTRUMENTED_HELLO_METRICS_SERVER_PORT=9543 cargo run
# In another terminal, run
curl http://localhost:9543
# ... and you should see your metrics appear

Note that, like user-defined configuration, the environment variable name for the metrics port takes its prefix from the service name passed to start.

Further Reading

See the API docs for full(ish) details on everything that's available.

Licence

Unless otherwise stated, everything in this repo is covered by the following copyright notice:

    Copyright (C) 2023  Matt Palmer <matt@hezmatt.org>

    This program is free software: you can redistribute it and/or modify it
    under the terms of the GNU General Public License version 3, as
    published by the Free Software Foundation.

    This program is distributed in the hope that it will be useful,
    but WITHOUT ANY WARRANTY; without even the implied warranty of
    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
    GNU General Public License for more details.

    You should have received a copy of the GNU General Public License
    along with this program.  If not, see <http://www.gnu.org/licenses/>.

Dependencies

~10–20MB
~278K SLoC