12 unstable releases (3 breaking)
| new 0.4.0-alpha.0 | Apr 8, 2026 |
|---|---|
| 0.3.0 | Apr 7, 2026 |
| 0.3.0-alpha.0 | Mar 25, 2026 |
| 0.2.0 | Feb 26, 2026 |
| 0.1.6 | Nov 28, 2025 |
#755 in Debugging
Used in rialo-cdk
115KB
1.5K
SLoC
rialo-telemetry
A comprehensive telemetry library for distributed tracing and metrics in Rialo applications. This crate provides a unified interface for setting up OpenTelemetry distributed tracing, Prometheus metrics, and console logging with minimal configuration.
Features
- OpenTelemetry Integration: Full support for distributed tracing with OTLP HTTP exporters
- OpenTelemetry Metrics: Configuration support for OTLP metrics export (implementation pending)
- Distributed Tracing: Centralized utilities for HTTP trace context propagation
- Baggage Support: Complete baggage manipulation utilities for distributed metadata propagation
- Prometheus Metrics: Optional span latency metrics and custom registry support
- Console Logging: Configurable structured logging to console
- Environment Variable Configuration: Automatic configuration from standard OpenTelemetry environment variables
- Flexible Configuration: Builder pattern for programmatic configuration
- Feature Gated: Optional dependencies based on your needs
Optional Features
axum-headers- Enables HTTP server trace context extraction utilities for axumdistributed-tracing- Enables OpenTelemetry distributed tracing supportenv-context- Enables environment variable-based trace context propagation for subprocess communicationprometheus- Enables Prometheus metrics collectionreqwest-headers- Enables HTTP client trace context injection utilities for reqwest
Quick Start
Console-Only Logging
use rialo_telemetry::{TelemetryConfig, init_telemetry};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize with console logging only
let config = TelemetryConfig::new();
let handle = init_telemetry(config).await?;
// Your application code here
tracing::info!("Application started");
// Shutdown telemetry when done
handle.shutdown()?;
Ok(())
}
OpenTelemetry with Environment Variables
Enable the distributed-tracing feature and set environment variables:
export OTEL_SERVICE_NAME="my-service"
export OTEL_SERVICE_VERSION="1.0.0"
export OTEL_EXPORTER_OTLP_ENDPOINT="http://localhost:4318"
export OTEL_EXPORTER_OTLP_HEADERS="x-api-key=your-key"
use rialo_telemetry::{TelemetryConfig, init_telemetry};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize with OTLP support (reads from environment)
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Your traced application code here
tracing::info!("Application started with distributed tracing");
handle.shutdown()?;
Ok(())
}
Programmatic Configuration
use rialo_telemetry::{TelemetryConfig, OtlpConfig, init_telemetry};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let otlp_config = OtlpConfig::new()
.with_service_name("my-service")
.with_service_version("1.0.0")
.with_exporter_endpoint("https://api.honeycomb.io/v1/traces")
.with_console_enabled(true);
let config = TelemetryConfig::new()
.with_otlp_config(otlp_config)
.with_log_level("debug");
let handle = init_telemetry(config).await?;
// Your application code
handle.shutdown()?;
Ok(())
}
OpenTelemetry Metrics Configuration
The crate includes full configuration support for OpenTelemetry metrics export, though the actual metrics implementation is not yet active. All environment variables and configuration options are parsed and stored for future use:
use rialo_telemetry::{TelemetryConfig, OtlpConfig, init_telemetry};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let otlp_config = OtlpConfig::new()
.with_service_name("my-service")
// Traces endpoint
.with_traces_endpoint("http://jaeger:14268/api/traces")
// Metrics configuration (ready for future implementation)
.with_exporter_endpoint("http://otel-collector:4318"); // Base endpoint for metrics
let config = TelemetryConfig::new()
.with_otlp_config(otlp_config);
let handle = init_telemetry(config).await?;
// When metrics are implemented, they will automatically use the configured endpoints
handle.shutdown()?;
Ok(())
}
Note: While metrics configuration is fully supported, the actual metrics export implementation is planned for a future release. Currently, only tracing is actively exported via OTLP.
Prometheus Metrics
Enable the prometheus feature:
use rialo_telemetry::{TelemetryConfig, PrometheusConfig, init_telemetry};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let registry = prometheus::Registry::new();
let prometheus_config = PrometheusConfig::new(registry.clone())
.with_span_latency_buckets(20)
.with_span_latency_enabled(true);
let config = TelemetryConfig::new()
.with_prometheus_config(prometheus_config);
let handle = init_telemetry(config).await?;
// Your application code - span latencies will be recorded
handle.shutdown()?;
Ok(())
}
Distributed Tracing Context Propagation
The crate provides utilities for propagating trace context across HTTP requests, enabling distributed tracing across microservices.
HTTP Client (reqwest) - Trace Context Injection
Enable the reqwest-headers feature to inject trace context into outgoing HTTP requests:
use rialo_telemetry::{inject_trace_headers, apply_trace_headers_to_reqwest};
use reqwest::Client;
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize telemetry with OpenTelemetry
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
let client = Client::new();
// In a traced context
let span = tracing::info_span!("http_request", service = "api-call");
let _guard = span.enter();
// Inject trace headers into request
let trace_headers = inject_trace_headers();
let request = client.post("https://api.example.com/data");
let request = apply_trace_headers_to_reqwest(request, trace_headers);
let response = request.send().await?;
handle.shutdown()?;
Ok(())
}
HTTP Server (axum) - Trace Context Extraction
Enable the axum-headers feature to extract trace context from incoming HTTP requests:
use rialo_telemetry::extract_and_set_trace_context_axum;
use axum::{extract::State, http::HeaderMap, Json, response::Json as ResponseJson};
#[tracing::instrument]
async fn handler(
headers: HeaderMap,
Json(payload): Json<serde_json::Value>
) -> ResponseJson<serde_json::Value> {
// Extract and set trace context from incoming headers
extract_and_set_trace_context_axum(&headers);
// Your handler logic - this span will now be part of the distributed trace
tracing::info!("Processing request with distributed trace context");
ResponseJson(serde_json::json!({"status": "ok"}))
}
End-to-End Distributed Tracing Example
Combining both client and server utilities for full distributed tracing:
// Service A (HTTP client)
use rialo_telemetry::{TelemetryConfig, init_telemetry, inject_trace_headers, apply_trace_headers_to_reqwest};
async fn call_service_b() -> Result<(), Box<dyn std::error::Error>> {
let span = tracing::info_span!("call_service_b");
let _guard = span.enter();
let client = reqwest::Client::new();
let trace_headers = inject_trace_headers();
let request = client.post("http://service-b:8080/api/process");
let request = apply_trace_headers_to_reqwest(request, trace_headers);
let response = request.send().await?;
tracing::info!("Received response from service B");
Ok(())
}
// Service B (HTTP server)
use rialo_telemetry::extract_and_set_trace_context_axum;
use axum::{http::HeaderMap, Json};
#[tracing::instrument]
async fn process_request(
headers: HeaderMap,
Json(data): Json<serde_json::Value>
) -> Json<serde_json::Value> {
// Extract trace context - this creates a child span of service A's span
extract_and_set_trace_context_axum(&headers);
tracing::info!("Processing request in service B");
// This span is now part of the same distributed trace as service A
let result = process_business_logic(data).await;
Json(result)
}
Note: Both utilities require the distributed-tracing feature to be enabled along with their respective feature flags (reqwest-headers or axum-headers).
Environment Variable Context Propagation
Enable the env-context feature to propagate trace context across process boundaries using environment variables.
Trace Context Inheritance
When the env-context feature is enabled, you can manually extract trace context from environment variables after initializing telemetry. This allows subprocesses to connect to their parent's trace:
use rialo_telemetry::{TelemetryConfig, init_telemetry, extract_and_set_trace_context_env};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize telemetry first
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Then extract trace context from environment variables (if any)
extract_and_set_trace_context_env();
// This span is now part of parent's distributed trace!
tracing::info!("Child process started");
handle.shutdown()?;
Ok(())
}
Ergonomic Command Helper
Use inject_trace_env_to_cmd() for convenient one-liner subprocess trace propagation:
use rialo_telemetry::{TelemetryConfig, init_telemetry, inject_trace_env_to_cmd};
use std::process::Command;
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize telemetry first
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Then extract trace context from environment variables (if any)
extract_and_set_trace_context_env();
// Parent process: ergonomic one-liner to inject trace context
let span = tracing::info_span!("subprocess_execution", command = "worker");
let _guard = span.enter();
// One line replaces the manual env injection loop!
let cmd = inject_trace_env_to_cmd(Command::new("./worker"));
let output = cmd.arg("--task=process").output()?;
tracing::info!("Subprocess completed with status: {}", output.status);
handle.shutdown()?;
Ok(())
}
Manual Control (Advanced Usage)
For fine-grained control, you can still use the manual functions:
use rialo_telemetry::{inject_trace_env, extract_and_set_trace_context_env};
use std::process::Command;
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize telemetry first
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Manually extract trace context at a specific point
extract_and_set_trace_context_env();
// Manual injection (equivalent to inject_trace_env_to_cmd)
let trace_env = inject_trace_env();
let mut cmd = Command::new("./child-process");
for (key, value) in trace_env {
cmd.env(key, value);
}
let output = cmd.output()?;
handle.shutdown()?;
Ok(())
}
In the child process, initialize telemetry and extract trace context:
use rialo_telemetry::{TelemetryConfig, init_telemetry, extract_and_set_trace_context_env};
#[tracing::instrument]
fn main() -> Result<(), Box<dyn std::error::Error>> {
let rt = tokio::runtime::Runtime::new()?;
rt.block_on(async {
// Initialize telemetry first
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Then extract trace context from environment variables
extract_and_set_trace_context_env();
// This span is now part of the parent process's distributed trace
tracing::info!("Child process started with inherited trace context");
do_work().await;
tracing::info!("Child process completed");
handle.shutdown()
})
}
async fn do_work() {
let span = tracing::info_span!("child_work");
let _guard = span.enter();
tracing::info!("Performing work in child process");
// This work is automatically traced as part of the parent's trace
}
You can also extract from a custom environment map instead of the current process environment:
use rialo_telemetry::extract_and_set_trace_context_from_env_map;
use std::collections::HashMap;
fn handle_custom_environment(custom_env: &HashMap<String, String>) {
// Extract trace context from a specific environment map
extract_and_set_trace_context_from_env_map(custom_env);
// Current span now has the extracted trace context as parent
tracing::info!("Working with custom trace context");
}
Cross-Process Distributed Tracing Example
Complete example showing trace propagation from parent to child process:
// parent.rs - The main process
use rialo_telemetry::{TelemetryConfig, init_telemetry, inject_trace_env_to_cmd};
use std::process::Command;
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Start a distributed trace
let span = tracing::info_span!("batch_job", job_id = "12345");
let _guard = span.enter();
tracing::info!("Starting batch job with multiple workers");
// Launch multiple child processes with ergonomic one-liners
for worker_id in 1..=3 {
let worker_span = tracing::info_span!("launch_worker", worker_id = worker_id);
let _worker_guard = worker_span.enter();
// Ergonomic one-liner - inject trace context and spawn
let cmd = inject_trace_env_to_cmd(Command::new("./worker"))
.arg(worker_id.to_string());
tracing::info!("Launching worker {}", worker_id);
cmd.spawn()?;
}
tracing::info!("All workers launched");
handle.shutdown()?;
Ok(())
}
// worker.rs - Child process
use rialo_telemetry::{TelemetryConfig, init_telemetry, extract_and_set_trace_context_env};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let worker_id = std::env::args().nth(1).unwrap_or("unknown".to_string());
// Initialize telemetry first
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Then extract trace context from environment variables
extract_and_set_trace_context_env();
// Create main span for this worker - now inherits from parent
let worker_span = tracing::info_span!("worker_process", worker_id = worker_id);
let _guard = worker_span.enter();
tracing::info!("Worker {} started with inherited trace", worker_id);
// Do work - all automatically part of the original batch job trace
process_batch_items().await;
tracing::info!("Worker {} completed", worker_id);
handle.shutdown()?;
Ok(())
}
async fn process_batch_items() {
for item in 1..=10 {
let item_span = tracing::info_span!("process_item", item_id = item);
let _guard = item_span.enter();
tracing::info!("Processing item {}", item);
// Simulate work
tokio::time::sleep(std::time::Duration::from_millis(100)).await;
}
}
Note: Environment variable context propagation requires the distributed-tracing feature to be enabled along with the env-context feature flag.
Baggage Support
Baggage provides a way to propagate key-value metadata across distributed systems alongside trace context. It's useful for passing cross-cutting concerns like user IDs, feature flags, request priorities, or any other data that should be available throughout a distributed trace.
The crate provides comprehensive baggage manipulation utilities when the distributed-tracing feature is enabled:
Basic Baggage Operations
use rialo_telemetry::{get_baggage, get_all_baggage};
use rialo_telemetry::{TelemetryConfig, init_telemetry};
use opentelemetry::{baggage::{Baggage, BaggageExt, BaggageMetadata}, Context, Key, Value};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize telemetry with OpenTelemetry (baggage requires this)
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// To set baggage, create a context with baggage and attach it within a scope
let span = tracing::info_span!("my_operation");
let _guard = span.enter();
// Create baggage and attach to context
{
let mut baggage = Baggage::new();
baggage.insert_with_metadata(
Key::new("user_id".to_string()),
Value::from("12345".to_string()),
BaggageMetadata::default(),
);
let context = Context::current().with_baggage(baggage);
let _baggage_guard = context.attach();
// Baggage is now available within this scope
let user_id = get_baggage("user_id"); // Some("12345")
let request_id = get_baggage("request_id"); // None
// Get all baggage as HashMap
let all_baggage = get_all_baggage();
println!("Current baggage: {:?}", all_baggage);
}
handle.shutdown()?;
Ok(())
}
Distributed Baggage Propagation
Baggage automatically propagates across distributed systems through the same mechanisms as trace context:
use rialo_telemetry::{get_baggage, inject_trace_headers, apply_trace_headers_to_reqwest};
use reqwest::Client;
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
let span = tracing::info_span!("call_service_b");
let _guard = span.enter();
// HTTP call - baggage automatically included in trace context headers if present
let client = Client::new();
let trace_headers = inject_trace_headers(); // Includes any baggage from current context
let request = client.post("http://service-b:8080/process");
let request = apply_trace_headers_to_reqwest(request, trace_headers);
let response = request.send().await?;
handle.shutdown()?;
Ok(())
}
// Service B: Automatically receives baggage
use rialo_telemetry::{extract_and_set_trace_context_axum, get_baggage};
use axum::{http::HeaderMap, Json};
#[tracing::instrument]
async fn process_request(
headers: HeaderMap,
Json(data): Json<serde_json::Value>
) -> Json<serde_json::Value> {
// Extract trace context (includes baggage)
extract_and_set_trace_context_axum(&headers);
// Baggage is now available in service B
let user_id = get_baggage("user_id"); // Some("user-12345")
let tenant_id = get_baggage("tenant_id"); // Some("tenant-abc")
let priority = get_baggage("request_priority"); // Some("high")
tracing::info!("Processing request for user {:?} in tenant {:?} with priority {:?}",
user_id, tenant_id, priority);
// Business logic can use baggage data
if priority == Some("high".to_string()) {
process_with_high_priority(data).await
} else {
process_normally(data).await
}
}
Cross-Process Baggage Propagation
Baggage also propagates across process boundaries when using environment variable context propagation:
use rialo_telemetry::inject_trace_env_to_cmd;
use std::process::Command;
use opentelemetry::baggage::BaggageExt;
use opentelemetry::Context;
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Parent process: Set baggage using OpenTelemetry context
let baggage = opentelemetry::baggage::Baggage::builder()
.with_entry("batch_id", "batch-2024-001")
.with_entry("processing_mode", "parallel")
.build();
let cx = Context::current().with_baggage(baggage);
let _guard = cx.attach();
let span = tracing::info_span!("launch_worker");
let _span_guard = span.enter();
// Launch subprocess with baggage propagation
let cmd = inject_trace_env_to_cmd(Command::new("./worker"));
let output = cmd.arg("--task=process").output()?;
handle.shutdown()?;
Ok(())
}
// Worker process: Inherits baggage
use rialo_telemetry::{TelemetryConfig, init_telemetry, extract_and_set_trace_context_env, get_baggage};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
// Initialize telemetry first
let config = TelemetryConfig::new().with_otlp();
let handle = init_telemetry(config).await?;
// Extract trace context and baggage from environment variables
extract_and_set_trace_context_env();
// Baggage from parent is now available
let batch_id = get_baggage("batch_id"); // Some("batch-2024-001")
let mode = get_baggage("processing_mode"); // Some("parallel")
tracing::info!("Worker started for batch {:?} in mode {:?}", batch_id, mode);
// Use baggage to influence processing
if mode == Some("parallel".to_string()) {
process_in_parallel().await;
} else {
process_sequentially().await;
}
handle.shutdown()?;
Ok(())
}
Baggage Best Practices
Use Cases:
- User identification across services
- Feature flags and A/B testing
- Request prioritization and routing
- Tenant or organization context
- Debug flags and trace sampling decisions
Performance Considerations:
- Keep baggage small (recommended: < 1KB total)
- Use short, meaningful keys
- Create new contexts with updated baggage when values need to change
- Drop context guards when exiting scopes to restore previous context
Security Notes:
- Don't put sensitive data in baggage (it's propagated in headers)
- Baggage is visible to all services in the trace
- Consider baggage as public metadata within your distributed system
Baggage Configuration
Baggage propagation is enabled by default when OpenTelemetry is configured. The propagators include both trace context and baggage:
// Baggage is included in default propagators
// OTEL_PROPAGATORS="tracecontext,baggage" (default)
// To customize propagators (not recommended unless you have specific needs):
std::env::set_var("OTEL_PROPAGATORS", "tracecontext,baggage,b3");
Note: Baggage utilities require the distributed-tracing feature to be enabled.
Environment Variables
The crate supports all standard OpenTelemetry environment variables:
Service Configuration
OTEL_SERVICE_NAME- Service name (default: "rialo")OTEL_SERVICE_VERSION- Service version (default: "unknown")OTEL_RESOURCE_ATTRIBUTES- Additional resource attributes
Endpoint Configuration
OTEL_EXPORTER_OTLP_ENDPOINT- General OTLP endpoint (default: "http://localhost:4318")OTEL_EXPORTER_OTLP_TRACES_ENDPOINT- Traces-specific endpoint (overrides general)OTEL_EXPORTER_OTLP_METRICS_ENDPOINT- Metrics-specific endpoint (overrides general)OTEL_EXPORTER_OTLP_INSECURE- Use insecure connection for general endpoint (default: false)OTEL_EXPORTER_OTLP_TRACES_INSECURE- Use insecure connection for traces (default: false)OTEL_EXPORTER_OTLP_METRICS_INSECURE- Use insecure connection for metrics (default: false)
Headers
OTEL_EXPORTER_OTLP_HEADERS- General headers for authenticationOTEL_EXPORTER_OTLP_TRACES_HEADERS- Traces-specific headers (merged with general)OTEL_EXPORTER_OTLP_METRICS_HEADERS- Metrics-specific headers (merged with general)
Protocol Configuration
OTEL_EXPORTER_OTLP_PROTOCOL- General export protocol: "grpc", "http/protobuf", "http/json" (default: "http/protobuf")OTEL_EXPORTER_OTLP_TRACES_PROTOCOL- Traces-specific protocol (overrides general)OTEL_EXPORTER_OTLP_METRICS_PROTOCOL- Metrics-specific protocol (overrides general)
Feature Toggles
OTEL_TRACES_ENABLED- Enable/disable traces (default: true)OTEL_METRICS_ENABLED- Enable/disable metrics (default: true)OTEL_LOG_LEVEL- Log level (default: "info")
Metrics Configuration
OTEL_EXPORTER_OTLP_METRICS_PERIOD- Metrics reporting interval (default: "30s")
Propagation
OTEL_PROPAGATORS- Trace context propagators, comma-separated (default: "tracecontext,baggage")
Local Jaeger Setup
For local development and testing, you can easily connect to a local Jaeger instance to visualize your traces.
Running Jaeger
Jaeger can be run either as an all-in-one Docker container or built locally. A Nix recipe is included for convenience in the rialo-nix-toolchain.
Docker (Recommended for Quick Testing)
Run the all-in-one Jaeger container:
docker run -d --name jaeger \
-e COLLECTOR_ZIPKIN_HTTP_PORT=9411 \
-p 5775:5775/udp \
-p 6831:6831/udp \
-p 6832:6832/udp \
-p 5778:5778 \
-p 16686:16686 \
-p 14268:14268 \
-p 9411:9411 \
jaegertracing/all-in-one:1.6.0
This exposes the following ports:
- 16686: Jaeger UI (http://localhost:16686)
- 14268: Jaeger collector (HTTP)
- 4318: OTLP gRPC endpoint (if using OTLP collector)
Nix (For Development Environment)
If you're using the rialo-nix-toolchain, you can run Jaeger with:
nix run .#jaeger
Connecting Your Application
Once Jaeger is running, configure your application to send traces to it using environment variables:
export OTEL_EXPORTER_OTLP_ENDPOINT=http://localhost:4318
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=http://localhost:4318/v1/traces
export OTEL_EXPORTER_OTLP_PROTOCOL=http/protobuf
Or create a .env file:
# .env
OTEL_EXPORTER_OTLP_ENDPOINT=http://localhost:4318
OTEL_EXPORTER_OTLP_TRACES_ENDPOINT=http://localhost:4318/v1/traces
OTEL_EXPORTER_OTLP_PROTOCOL=http/protobuf
Viewing Traces
- Start your application with the environment variables set
- Generate some traces by using your application
- Open http://localhost:16686 in your browser
- Select your service from the dropdown
- Click "Find Traces" to see your traces
Configuration Precedence
Configuration values are resolved in this order (highest to lowest precedence):
- Programmatic configuration via builder methods
- Environment variables
- Default values
For endpoints and headers, signal-specific settings override general settings:
OTEL_EXPORTER_OTLP_TRACES_ENDPOINToverridesOTEL_EXPORTER_OTLP_ENDPOINTfor tracesOTEL_EXPORTER_OTLP_METRICS_ENDPOINToverridesOTEL_EXPORTER_OTLP_ENDPOINTfor metrics- Signal-specific headers are merged with general headers (signal-specific takes precedence for conflicts)
Examples
Honeycomb.io Integration
export OTEL_SERVICE_NAME="my-service"
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT="https://api.honeycomb.io/v1/traces"
export OTEL_EXPORTER_OTLP_HEADERS="x-honeycomb-team=your-api-key,x-honeycomb-dataset=my-dataset"
# If using metrics (when implemented):
# export OTEL_EXPORTER_OTLP_METRICS_ENDPOINT="https://api.honeycomb.io/v1/metrics"
Jaeger Integration
export OTEL_SERVICE_NAME="my-service"
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT="http://localhost:14268/api/traces"
export OTEL_EXPORTER_OTLP_PROTOCOL="http/protobuf"
# Jaeger doesn't support metrics via OTLP, so disable them:
# export OTEL_METRICS_ENABLED="false"
Separate Traces and Metrics Endpoints
export OTEL_SERVICE_NAME="my-service"
# Send traces to Jaeger
export OTEL_EXPORTER_OTLP_TRACES_ENDPOINT="http://jaeger:14268/api/traces"
# Send metrics to Prometheus-compatible endpoint (when implemented)
export OTEL_EXPORTER_OTLP_METRICS_ENDPOINT="http://prometheus:9090/api/v1/otlp/v1/metrics"
# Different authentication for each
export OTEL_EXPORTER_OTLP_TRACES_HEADERS="authorization=Bearer traces-token"
export OTEL_EXPORTER_OTLP_METRICS_HEADERS="authorization=Bearer metrics-token"
Development Setup
# Console logging only for development
export RUST_LOG="debug"
# No OTEL_EXPORTER_OTLP_ENDPOINT set - will use console only
Error Handling
The library handles common error scenarios gracefully:
- Invalid endpoints: Empty or invalid endpoints disable OpenTelemetry export
- Network failures: Export failures don't crash the application
- Configuration errors: Invalid environment variables fall back to defaults
- Global subscriber conflicts: Handles multiple initialization attempts gracefully
Performance Considerations
- Batched Export: Uses OpenTelemetry's batched span processor for efficient export
- Conditional Compilation: Feature gates ensure zero overhead when features are disabled
- Efficient Headers: Headers are parsed once and reused
- Resource Detection: Uses OpenTelemetry's resource detection for optimal metadata
Testing
# Test with no features
cargo nextest run -p rialo-telemetry
# Test with Prometheus
cargo nextest run -p rialo-telemetry --features prometheus
License
Licensed under the Apache License, Version 2.0.
Contributing
This crate is part of the Rialo project. See the main repository for contribution guidelines.
Dependencies
~8–29MB
~320K SLoC