1 unstable release
0.1.0 | Oct 14, 2023 |
---|
#61 in #probe
430KB
8K
SLoC
Dropshot is a general-purpose--but opinionated--crate for exposing REST APIs from a Rust program. It is meant to be simple and lightweight. It includes first-class support for OpenAPI in the form of precise specs generated directly from code. It supports primitives for consistent pagination including extensions an OpenAPI extension denoting it in the spec.
For information about configuration, design, and contributing, see the GitHub repo.
For information on the use the API see the documentation.
lib.rs
:
Carpio is a general-purpose crate for exposing REST APIs from a Rust program. Planned highlights include:
-
Suitability for production use on a largely untrusted network. Carpio-based systems should be high-performing, reliable, debuggable, and secure against basic denial of service attacks (intentional or otherwise).
-
First-class OpenAPI support, in the form of precise OpenAPI specs generated directly from code. This works because the functions that serve HTTP resources consume arguments and return values of specific types from which a schema can be statically generated.
-
Ease of integrating into a diverse team. An important use case for Carpio consumers is to have a team of engineers where individuals might add a few endpoints at a time to a complex server, and it should be relatively easy to do this. Part of this means an emphasis on the principle of least surprise: like Rust itself, we may choose abstractions that require more time to learn up front in order to make it harder to accidentally build systems that will not perform, will crash in corner cases, etc.
By "REST API", we primarily mean an API built atop existing HTTP primitives,
organized into hierarchical resources, and providing consistent, idempotent
mechanisms to create, update, list, and delete those resources. "REST" can
mean a range of things depending on who you talk to, and some people are
dogmatic about what is or isn't RESTy. We find such dogma not only
unhelpful, but poorly defined. (Consider such a simple case as trying to
update a resource in a REST API. Popular APIs sometimes use PUT
, PATCH
,
or POST
for the verb; and JSON Merge Patch or JSON Patch as the format.
(sometimes without even knowing it!). There's hardly a clear standard, yet
this is a really basic operation for any REST API.)
For a discussion of alternative crates considered, see Oxide RFD 10.
We hope Carpio will be fairly general-purpose, but it's primarily intended to address the needs of the Oxide control plane.
Usage
The bare minimum might look like this:
use carpio::ApiDescription;
use carpio::ConfigCarpio;
use carpio::HttpServerStarter;
use std::sync::Arc;
#[tokio::main]
async fn main() -> Result<(), String> {
// Describe the API.
let api = ApiDescription::new();
// Register API functions -- see detailed example or ApiDescription docs.
// Start the server.
let server =
HttpServerStarter::new(
&ConfigCarpio {
bind_address: "127.0.0.1:0".parse().unwrap(),
request_body_max_bytes: 1024,
},
api,
Arc::new(()),
)
.map_err(|error| format!("failed to start server: {}", error))?
.start();
server.await
}
This server returns a 404 for all resources because no API functions were
registered. See examples/basic.rs
for a simple, documented example that
provides a few resources using shared state.
For a given ApiDescription
, you can also print out an OpenAPI spec
describing the API. See ApiDescription::openapi
.
API Handler Functions
HTTP talks about resources. For a REST API, we often talk about endpoints or operations, which are identified by a combination of the HTTP method and the URI path.
Example endpoints for a resource called a "project" might include:
GET /projects
(list projects)POST /projects
(one way to create a project)GET /projects/my_project
(fetch one project)PUT /projects/my_project
(update (or possibly create) a project)DELETE /projects/my_project
(delete a project)
With Carpio, an incoming request for a given API endpoint is handled by a
particular Rust function. That function is called an entrypoint, an
endpoint handler, or a handler function. When you set up a Carpio
server, you configure the set of available API endpoints and which functions
will handle each one by setting up an ApiDescription
.
Typically, you define an endpoint with a handler function by using the
endpoint
macro. Here's an example of a single endpoint that lists
a hardcoded project:
use carpio::endpoint;
use carpio::ApiDescription;
use carpio::HttpError;
use carpio::HttpResponseOk;
use carpio::RequestContext;
use http::Method;
use schemars::JsonSchema;
use serde::Serialize;
use std::sync::Arc;
/** Represents a project in our API */
#[derive(Serialize, JsonSchema)]
struct Project {
/** name of the project */
name: String,
}
/** Fetch a project. */
#[endpoint {
method = GET,
path = "/projects/project1",
}]
async fn myapi_projects_get_project(
rqctx: RequestContext<()>,
) -> Result<HttpResponseOk<Project>, HttpError>
{
let project = Project { name: String::from("project1") };
Ok(HttpResponseOk(project))
}
fn main() {
let mut api = ApiDescription::new();
/*
* Register our endpoint and its handler function. The "endpoint" macro
* specifies the HTTP method and URI path that identify the endpoint,
* allowing this metadata to live right alongside the handler function.
*/
api.register(myapi_projects_get_project).unwrap();
/* ... (use `api` to set up an `HttpServer` ) */
}
There's quite a lot going on here:
- The
endpoint
macro specifies the HTTP method and URI path. When we invokeApiDescription::register()
, this information is used to register the endpoint that will be handled by our function. - The signature of our function indicates that on success, it returns a
HttpResponseOk<Project>
. This means that the function will return an HTTP 200 status code ("OK") with an object of typeProject
. - The function itself has a Rustdoc comment that will be used to document this endpoint in the OpenAPI schema.
From this information, Carpio can generate an OpenAPI specification for this API that describes the endpoint (which OpenAPI calls an "operation"), its documentation, the possible responses that it can return, and the schema for each type of response (which can also include documentation). This is largely known statically, though generated at runtime.
#[endpoint { ... }]
attribute parameters
The endpoint
attribute accepts parameters the affect the operation of
the endpoint as well as metadata that appears in the OpenAPI description
of it.
#[endpoint {
// Required fields
method = { DELETE | HEAD | GET | OPTIONS | PATCH | POST | PUT },
path = "/path/name/with/{named}/{variables}",
// Optional fields
tags = [ "all", "your", "OpenAPI", "tags" ],
}]
This is where you specify the HTTP method and path (including path variables) for the API endpoint. These are used as part of endpoint registration and appear in the OpenAPI spec output.
The tags field is used to categorize API endpoints and only impacts the OpenAPI spec output.
Function parameters
In general, a handler function looks like this:
async fn f(
rqctx: RequestContext<Context>,
[query_params: Query<Q>,]
[path_params: Path<P>,]
[body_param: TypedBody<J>,]
[body_param: UntypedBody,]
[body_param: StreamingBody,]
[raw_request: RawRequest,]
) -> Result<HttpResponse*, HttpError>
The RequestContext
must appear first. The Context
type is
caller-provided context which is provided when the server is created.
The types Query
, Path
, TypedBody
, UntypedBody
, and RawRequest
are
called Extractors because they cause information to be pulled out of the
request and made available to the handler function.
Query
<Q>
extracts parameters from a query string, deserializing them into an instance of typeQ
.Q
must implementserde::Deserialize
andschemars::JsonSchema
.Path
<P>
extracts parameters from HTTP path, deserializing them into an instance of typeP
.P
must implementserde::Deserialize
andschemars::JsonSchema
.TypedBody
<J>
extracts content from the request body by parsing the body as JSON (or form/url-encoded) and deserializing it into an instance of typeJ
.J
must implementserde::Deserialize
andschemars::JsonSchema
.UntypedBody
extracts the raw bytes of the request body.StreamingBody
provides the raw bytes of the request body as aStream
ofBytes
chunks.RawRequest
provides access to the underlyinghyper::Request
. The hope is that this would generally not be needed. It can be useful to implement functionality not provided by Carpio.
Query
and Path
impl SharedExtractor
. TypedBody
, UntypedBody
,
StreamingBody
, and RawRequest
impl ExclusiveExtractor
. Your function
may accept 0-3 extractors, but only one can be ExclusiveExtractor
, and it
must be the last one. Otherwise, the order of extractor arguments does not
matter.
If the handler accepts any extractors and the corresponding extraction cannot be completed, the request fails with status code 400 and an error message reflecting the error (usually a validation error).
As with any serde-deserializable type, you can make fields optional by having
the corresponding property of the type be an Option
. Here's an example of
an endpoint that takes two arguments via query parameters: "limit", a
required u32, and "marker", an optional string:
use http::StatusCode;
use carpio::HttpError;
use carpio::TypedBody;
use carpio::Query;
use carpio::RequestContext;
use hyper::Body;
use hyper::Response;
use schemars::JsonSchema;
use serde::Deserialize;
use std::sync::Arc;
#[derive(Deserialize, JsonSchema)]
struct MyQueryArgs {
limit: u32,
marker: Option<String>
}
struct MyContext {}
async fn myapi_projects_get(
rqctx: RequestContext<MyContext>,
query: Query<MyQueryArgs>)
-> Result<Response<Body>, HttpError>
{
let query_args = query.into_inner();
let context: &MyContext = rqctx.context();
let limit: u32 = query_args.limit;
let marker: Option<String> = query_args.marker;
Ok(Response::builder()
.status(StatusCode::OK)
.body(format!("limit = {}, marker = {:?}\n", limit, marker).into())?)
}
Endpoint function return types
Endpoint handler functions are async, so they always return a Future
. When
we say "return type" below, we use that as shorthand for the output of the
future.
An endpoint function must return a type that implements HttpResponse
.
Typically this should be a type that implements HttpTypedResponse
(either
one of the Carpio-provided ones or one of your own creation).
The more specific a type returned by the handler function, the more can be
validated at build-time, and the more specific an OpenAPI schema can be
generated from the source code. For example, a POST to an endpoint
"/projects" might return Result<HttpResponseCreated<Project>, HttpError>
.
As you might expect, on success, this turns into an HTTP 201 "Created"
response whose body is constructed by serializing the Project
. In this
example, OpenAPI tooling can identify at build time that this function
produces a 201 "Created" response on success with a body whose schema matches
Project
(which we already said implements Serialize
), and there would be
no way to violate this contract at runtime.
These are the implementations of HttpTypedResponse
with their associated
HTTP response code
on the HTTP method:
Return Type | HTTP status code |
---|---|
HttpResponseOk |
200 |
HttpResponseCreated |
201 |
HttpResponseAccepted |
202 |
HttpResponseDeleted |
204 |
HttpResponseUpdatedNoContent |
204 |
In situations where the response schema is not fixed, the endpoint should
return Response<Body>
, which also implements HttpResponse
. Note that
the OpenAPI spec will not include any status code or type information in
this case.
What about generic handlers that run on all requests?
There's no mechanism in Carpio for this. Instead, it's recommended that users commonize code using regular Rust functions and calling them. See the design notes in the README for more on this.
Support for paginated resources
"Pagination" here refers to the interface pattern where HTTP resources (or API endpoints) that provide a list of the items in a collection return a relatively small maximum number of items per request, often called a "page" of results. Each page includes some metadata that the client can use to make another request for the next page of results. The client can repeat this until they've gotten all the results. Limiting the number of results returned per request helps bound the resource utilization and time required for any request, which in turn facilities horizontal scalability, high availability, and protection against some denial of service attacks (intentional or otherwise). For more background, see the comments in carpio/src/pagination.rs.
Pagination support in Carpio implements this common pattern:
- This server exposes an API endpoint that returns the items contained within a collection.
- The client is not allowed to list the entire collection in one request. Instead, they list the collection using a sequence of requests to the one endpoint. We call this sequence of requests a scan of the collection, and we sometimes say that the client pages through the collection.
- The initial request in the scan may specify the scan parameters, which typically specify how the results are to be sorted (i.e., by which field(s) and whether the sort is ascending or descending), any filters to apply, etc.
- Each request returns a page of results at a time, along with a page token that's provided with the next request as a query parameter.
- The scan parameters cannot change between requests that are part of the same scan.
- With all requests: there's a default limit (e.g., 100 items returned at a
time). Clients can request a higher limit using a query parameter (e.g.,
limit=1000
). This limit is capped by a hard limit on the server. If the client asks for more than the hard limit, the server can use the hard limit or reject the request.
As an example, imagine that we have an API endpoint called "/animals"
. Each
item returned is an Animal
object that might look like this:
{
"name": "aardvark",
"class": "mammal",
"max_weight": "80", /* kilograms, typical */
}
There are at least 1.5 million known species of animal -- too many to return
in one API call! Our API supports paginating them by "name"
, which we'll
say is a unique field in our data set.
The first request to the API fetches "/animals"
(with no querystring
parameters) and returns:
{
"next_page": "abc123...",
"items": [
{
"name": "aardvark",
"class": "mammal",
"max_weight": "80",
},
...
{
"name": "badger",
"class": "mammal",
"max_weight": "12",
}
]
}
The subsequent request to the API fetches "/animals?page_token=abc123..."
.
The page token "abc123..."
is an opaque token to the client, but typically
encodes the scan parameters and the value of the last item seen
("name=badger"
). The client knows it has completed the scan when it
receives a response with no next_page
in it.
Our API endpoint can also support scanning in reverse order. In this case,
when the client makes the first request, it should fetch
"/animals?sort=name-descending"
. Now the first result might be "zebra"
.
Again, the page token must include the scan parameters so that in subsequent
requests, the API endpoint knows that we're scanning backwards, not forwards,
from the value we were given. It's not allowed to change directions or sort
order in the middle of a scan. (You can always start a new scan, but you
can't pick up from where you were in the previous scan.)
It's also possible to support sorting by multiple fields. For example, we
could support sort=class-name
, which we could define to mean that we'll
sort the results first by the animal's class, then by name. Thus we'd get
all the amphibians in sorted order, then all the mammals, then all the
reptiles. The main requirement is that the combination of fields used for
pagination must be unique. We cannot paginate by the animal's class alone.
(To see why: there are over 6,000 mammals. If the page size is, say, 1000,
then the page_token would say "mammal"
, but there's not enough information
there to see where we are within the list of mammals. It doesn't matter
whether there are 2 mammals or 6,000 because clients can limit the page size
to just one item if they want and that ought to work.)
Carpio interfaces for pagination
We can think of pagination in two parts: the input (handling the pagination query parameters) and the output (emitting a page of results, including the page token).
For input, a paginated API endpoint's handler function should accept a
Query
<
PaginationParams
<ScanParams, PageSelector>>
, where
ScanParams
is a consumer-defined type specifying the parameters of the scan
(typically including the sort fields, sort order, and filter options) and
PageSelector
is a consumer-defined type describing the page token. The
PageSelector will be serialized to JSON and base64-encoded to construct the
page token. This will be automatically parsed on the way back in.
For output, a paginated API endpoint's handler function can return
Result<
HttpResponseOk
<ResultsPage
<T>, HttpError>
where T: Serialize
is the item listed by the endpoint. You can also use your own
structure that contains a ResultsPage
(possibly using
#[serde(flatten)]
), if that's the behavior you want.
There are several complete, documented examples in the "examples" directory.
Advanced usage notes
It's possible to accept additional query parameters besides the pagination
parameters by having your API endpoint handler function take two different
arguments using Query
, like this:
use carpio::HttpError;
use carpio::HttpResponseOk;
use carpio::PaginationParams;
use carpio::Query;
use carpio::RequestContext;
use carpio::ResultsPage;
use carpio::endpoint;
use schemars::JsonSchema;
use serde::Deserialize;
use std::sync::Arc;
#[derive(Deserialize, JsonSchema)]
struct MyExtraQueryParams {
do_extra_stuff: bool,
}
#[endpoint {
method = GET,
path = "/list_stuff"
}]
async fn my_list_api(
rqctx: RequestContext<()>,
pag_params: Query<PaginationParams<MyScanParams, MyPageSelector>>,
extra_params: Query<MyExtraQueryParams>,
) -> Result<HttpResponseOk<ResultsPage<String>>, HttpError>
{
# unimplemented!();
/* ... */
}
You might expect that instead of doing this, you could define your own
structure that includes a PaginationParams
using #[serde(flatten)]
, and
this ought to work, but it currently doesn't due to serde_urlencoded#33,
which is really serde#1183.
Note that any parameters defined by MyScanParams
are effectively encoded
into the page token and need not be supplied with invocations when page_token
is specified. That is not the case for required parameters defined by
MyExtraQueryParams
--those must be supplied on each invocation.
OpenAPI extension
In generated OpenAPI documents, Carpio adds the x-carpio-pagination
extension to paginated operations. The value is currently a structure
with this format:
{
"required": [ .. ]
}
The string values in the required
array are the names of those query
parameters that are mandatory if page_token
is not specified (when
fetching the first page of data).
DTrace probes
Carpio optionally exposes two DTrace probes, request_start
and
request_finish
. These provide detailed information about each request,
such as their ID, the local and remote IPs, and the response information.
See the carpio::dtrace::RequestInfo
and carpio::dtrace::ResponseInfo
types for a complete listing of what's available.
These probes are implemented via the usdt
crate. They may require a
nightly toolchain if built on macOS prior to Rust version 1.66. Otherwise a
stable compiler >= v1.59 is required in order to present the necessary
features. Given these constraints, USDT functionality is behind the feature
flag "usdt-probes"
, which may become a default feature of this crate in
future releases.
Important: The probes are internally registered with the DTrace kernel module, making them visible via
dtrace(1M)
. This is done when anHttpServer
object is created, but it's possible that registration fails. The result of registration is stored in the server after creation, and can be accessed with the [HttpServer::probe_registration()
] method. This allows callers to decide how to handle failures, but ensures that probes are always enabled if possible.
Once in place, the probes can be seen via DTrace. For example, running:
$ cargo +nightly run --example basic --features usdt-probes
And making several requests to it with curl
, we can see the DTrace
probes with an invocation like:
## dtrace -Zq -n 'carpio*:::request-* { printf("%s\n", copyinstr(arg0)); }'
{"ok":{"id":"b793c62e-60e4-45c5-9274-198a04d9abb1","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:34286","method":"GET","path":"/counter","query":null}}
{"ok":{"id":"b793c62e-60e4-45c5-9274-198a04d9abb1","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:34286","status_code":200,"message":""}}
{"ok":{"id":"9050e30a-1ce3-4d6f-be1c-69a11c618800","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:41101","method":"PUT","path":"/counter","query":null}}
{"ok":{"id":"9050e30a-1ce3-4d6f-be1c-69a11c618800","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:41101","status_code":400,"message":"do not like the number 10"}}
{"ok":{"id":"a53696af-543d-452f-81b6-5a045dd9921d","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:57376","method":"PUT","path":"/counter","query":null}}
{"ok":{"id":"a53696af-543d-452f-81b6-5a045dd9921d","local_addr":"127.0.0.1:61028","remote_addr":"127.0.0.1:57376","status_code":204,"message":""}}
Dependencies
~19–30MB
~610K SLoC