Compare commits
29 Commits
Author | SHA1 | Date | |
---|---|---|---|
66b51df2af | |||
28d05d1d0e | |||
a774f92bf7 | |||
0b9769dc39 | |||
9dadd9edc3 | |||
c6230be91e | |||
796b2a8412 | |||
6a3bc7bab3 | |||
da13a371b4 | |||
a1a5a6b8be | |||
96afb77844 | |||
c372d79d1d | |||
b5141d6a70 | |||
694f8ba1d3 | |||
09b52f92d7 | |||
785abdec8d | |||
ce203a4618 | |||
3c994a7343 | |||
f69c3e9c3c | |||
fac5b1f275 | |||
a37559b93e | |||
faac09ac0c | |||
62a7605869 | |||
0996167865 | |||
87344dd601 | |||
39e6611602 | |||
355d246010 | |||
e5836a8b97 | |||
f6d160a7f8 |
1
.github/copilot-instructions.md
vendored
Symbolic link
1
.github/copilot-instructions.md
vendored
Symbolic link
@ -0,0 +1 @@
|
||||
../CLAUDE.md
|
34
CHANGELOG.md
Normal file
34
CHANGELOG.md
Normal file
@ -0,0 +1,34 @@
|
||||
# Release Notes - v0.5.1
|
||||
|
||||
## Observability Enhancements
|
||||
|
||||
### OTLP Metrics Support
|
||||
- **New `metrics/` package** - OpenTelemetry-native metrics with OTLP export support for structured metrics collection
|
||||
- **Centralized OTLP configuration** - Refactored configuration to `internal/tracerconfig/` to eliminate code duplication across tracing, logging, and metrics
|
||||
- **HTTP retry support** - Added consistent retry configuration for all HTTP OTLP exporters to improve reliability
|
||||
|
||||
### Enhanced Logging
|
||||
- **Buffering exporter** - Added OTLP log buffering to queue logs until tracing configuration is available
|
||||
- **TLS support for logs** - Client certificate authentication support for secure OTLP log export
|
||||
- **Improved logfmt formatting** - Better structured output for log messages
|
||||
|
||||
### Tracing Improvements
|
||||
- **HTTP retry support** - OTLP trace requests now automatically retry on failure when using HTTP transport
|
||||
|
||||
## Build System
|
||||
|
||||
### Version Package Enhancements
|
||||
- **Unix epoch build time support** - Build time can now be injected as Unix timestamps (`$(date +%s)`) in addition to RFC3339 format
|
||||
- **Simplified build commands** - Reduces complexity of ldflags injection while maintaining backward compatibility
|
||||
- **Consistent output format** - All build times normalize to RFC3339 format regardless of input
|
||||
|
||||
## API Changes
|
||||
|
||||
### New Public Interfaces
|
||||
- `metrics.NewMeterProvider()` - Create OTLP metrics provider with centralized configuration
|
||||
- `metrics.Shutdown()` - Graceful shutdown for metrics exporters
|
||||
- `internal/tracerconfig` - Shared OTLP configuration utilities (internal package)
|
||||
|
||||
### Dependencies
|
||||
- Added explicit OpenTelemetry metrics dependencies to `go.mod`
|
||||
- Updated tracing dependencies for retry support
|
163
CLAUDE.md
Normal file
163
CLAUDE.md
Normal file
@ -0,0 +1,163 @@
|
||||
# CLAUDE.md
|
||||
|
||||
This file provides guidance to Claude Code (claude.ai/code) when working with code in this repository.
|
||||
|
||||
## Commands
|
||||
|
||||
### Testing
|
||||
- Run all tests: `go test ./...`
|
||||
- Run tests with verbose output: `go test -v ./...`
|
||||
- Run tests for specific package: `go test ./config`
|
||||
- Run specific test: `go test -run TestConfigBool ./config`
|
||||
|
||||
### Building
|
||||
- Build all packages: `go build ./...`
|
||||
- Check module dependencies: `go mod tidy`
|
||||
- Verify dependencies: `go mod verify`
|
||||
|
||||
### Code Quality
|
||||
- Format code: `go fmt ./...`
|
||||
- Vet code: `go vet ./...`
|
||||
- Run static analysis: `staticcheck ./...` (if available)
|
||||
|
||||
## Architecture
|
||||
|
||||
This is a common library (`go.ntppool.org/common`) providing shared infrastructure for the NTP Pool project. The codebase emphasizes observability, security, and modern Go practices.
|
||||
|
||||
### Core Components
|
||||
|
||||
**Web Service Foundation:**
|
||||
- `ekko/` - Enhanced Echo web framework with pre-configured middleware (OpenTelemetry, Prometheus, logging, security headers)
|
||||
- `health/` - Standalone health check HTTP server with `/__health` endpoint
|
||||
- `metricsserver/` - Prometheus metrics exposure via `/metrics` endpoint
|
||||
|
||||
**Observability Stack:**
|
||||
- `logger/` - Structured logging with OpenTelemetry trace integration and multiple output formats
|
||||
- `tracing/` - OpenTelemetry distributed tracing with OTLP export support
|
||||
- `metricsserver/` - Prometheus metrics with custom registry
|
||||
|
||||
**Configuration & Environment:**
|
||||
- `config/` - Environment-based configuration with code-generated accessors (`config_accessor.go`)
|
||||
- `version/` - Build metadata and version information with Cobra CLI integration
|
||||
|
||||
**Security & Communication:**
|
||||
- `apitls/` - TLS certificate management with automatic renewal via certman
|
||||
- `kafka/` - Kafka client wrapper with TLS support for log streaming
|
||||
- `xff/fastlyxff/` - Fastly CDN IP range management for trusted proxy handling
|
||||
|
||||
**Utilities:**
|
||||
- `ulid/` - Thread-safe ULID generation with monotonic ordering
|
||||
- `timeutil/` - JSON-serializable duration types
|
||||
- `types/` - Shared data structures (LogScoreAttributes for NTP server scoring)
|
||||
|
||||
### Key Patterns
|
||||
|
||||
**Functional Options:** Used extensively in `ekko/` for flexible service configuration
|
||||
**Interface-Based Design:** `CertificateProvider` in `apitls/` for pluggable certificate management
|
||||
**Context Propagation:** Throughout the codebase for cancellation and tracing
|
||||
**Graceful Shutdown:** Implemented in web servers and background services
|
||||
|
||||
### Dependencies
|
||||
|
||||
The codebase heavily uses:
|
||||
- Echo web framework with custom middleware stack
|
||||
- OpenTelemetry for observability (traces, metrics, logs)
|
||||
- Prometheus for metrics collection
|
||||
- Kafka for message streaming
|
||||
- Cobra for CLI applications
|
||||
|
||||
### Code Generation
|
||||
|
||||
`config/config_accessor.go` is generated - modify `config.go` and regenerate accessors when adding new configuration options.
|
||||
|
||||
## Package Overview
|
||||
|
||||
### `apitls/`
|
||||
TLS certificate management with automatic renewal support via certman. Provides a CA pool for trusted certificates and interfaces for pluggable certificate providers. Used for secure inter-service communication.
|
||||
|
||||
### `config/`
|
||||
Environment-based configuration system with code-generated accessor methods. Handles deployment mode, hostname configuration, and TLS settings. Provides URL building utilities for web and management interfaces.
|
||||
|
||||
### `ekko/`
|
||||
Enhanced Echo web framework wrapper with pre-configured middleware stack including OpenTelemetry tracing, Prometheus metrics, structured logging, gzip compression, and security headers. Supports HTTP/2 with graceful shutdown.
|
||||
|
||||
### `health/`
|
||||
Standalone HTTP health check server that runs independently from the main application. Exposes `/__health` endpoint with configurable health handlers, timeouts, and graceful shutdown capabilities.
|
||||
|
||||
### `kafka/`
|
||||
Kafka client wrapper with TLS support for secure log streaming. Provides connection management, broker discovery, and reader/writer factories with compression and batching optimizations.
|
||||
|
||||
### `logger/`
|
||||
Structured logging system with OpenTelemetry trace integration. Supports multiple output formats (text, OTLP) with configurable log levels, systemd compatibility, and context-aware logging.
|
||||
|
||||
### `metricsserver/`
|
||||
Dedicated Prometheus metrics HTTP server with custom registry isolation. Exposes `/metrics` endpoint with OpenMetrics support and graceful shutdown handling.
|
||||
|
||||
### `timeutil/`
|
||||
JSON-serializable duration types that support both string parsing ("30s", "5m") and numeric nanosecond values. Compatible with configuration files and REST APIs.
|
||||
|
||||
### `tracing/`
|
||||
OpenTelemetry distributed tracing setup with support for OTLP export via gRPC or HTTP. Handles resource detection, propagation, and automatic instrumentation with configurable TLS.
|
||||
|
||||
### `types/`
|
||||
Shared data structures for the NTP Pool project. Currently contains `LogScoreAttributes` for NTP server scoring with JSON and SQL database compatibility.
|
||||
|
||||
### `ulid/`
|
||||
Thread-safe ULID (Universally Unique Lexicographically Sortable Identifier) generation using cryptographically secure randomness. Optimized for simplicity and performance in high-concurrency environments.
|
||||
|
||||
### `version/`
|
||||
Build metadata and version information system with Git integration. Provides CLI commands for Cobra and Kong frameworks, Prometheus build info metrics, and semantic version validation.
|
||||
|
||||
### `xff/fastlyxff/`
|
||||
Fastly CDN IP range management for trusted proxy handling. Parses Fastly's IP ranges JSON file and generates Echo framework trust options for proper client IP extraction.
|
||||
|
||||
## Go Development Best Practices
|
||||
|
||||
### Code Style
|
||||
- Follow standard Go formatting (`go fmt ./...`)
|
||||
- Use `go vet ./...` for static analysis
|
||||
- Run `staticcheck ./...` when available
|
||||
- Prefer short, descriptive variable names
|
||||
- Use interfaces for testability and flexibility
|
||||
|
||||
### Error Handling
|
||||
- Always handle errors explicitly
|
||||
- Use `errors.Join()` for combining multiple errors
|
||||
- Wrap errors with context using `fmt.Errorf("context: %w", err)`
|
||||
- Return early on errors to reduce nesting
|
||||
|
||||
### Testing
|
||||
- Write table-driven tests when testing multiple scenarios
|
||||
- Use `t.Helper()` in test helper functions
|
||||
- Test error conditions, not just happy paths
|
||||
- Use `testing.Short()` for integration tests that can be skipped
|
||||
|
||||
### Concurrency
|
||||
- Use contexts for cancellation and timeouts
|
||||
- Prefer channels for communication over shared memory
|
||||
- Use `sync.Once` for one-time initialization
|
||||
- Always call `defer cancel()` after `context.WithCancel()`
|
||||
|
||||
### Performance
|
||||
- Use `sync.Pool` for frequently allocated objects
|
||||
- Prefer slices over arrays for better performance
|
||||
- Use `strings.Builder` for string concatenation in loops
|
||||
- Profile before optimizing with `go tool pprof`
|
||||
|
||||
### Observability
|
||||
- Use structured logging with key-value pairs
|
||||
- Add OpenTelemetry spans for external calls
|
||||
- Include trace IDs in error messages
|
||||
- Use metrics for monitoring application health
|
||||
|
||||
### Dependencies
|
||||
- Keep dependencies minimal and well-maintained
|
||||
- Use `go mod tidy` to clean up unused dependencies
|
||||
- Pin major versions to avoid breaking changes
|
||||
- Prefer standard library when possible
|
||||
|
||||
### Security
|
||||
- Never log sensitive information (passwords, tokens)
|
||||
- Use `crypto/rand` for cryptographic randomness
|
||||
- Validate all inputs at API boundaries
|
||||
- Use TLS for all network communication
|
20
README.md
Normal file
20
README.md
Normal file
@ -0,0 +1,20 @@
|
||||
|
||||
Common library for the NTP Pool project with shared infrastructure components.
|
||||
|
||||
## Packages
|
||||
|
||||
- **apitls** - TLS setup for NTP Pool internal services with embedded CA
|
||||
- **config** - NTP Pool project configuration with environment variables
|
||||
- **ekko** - Enhanced Echo web framework with observability middleware
|
||||
- **health** - Standalone health check HTTP server
|
||||
- **kafka** - Kafka client wrapper with TLS support
|
||||
- **logger** - Structured logging with OpenTelemetry integration
|
||||
- **metricsserver** - Prometheus metrics HTTP server
|
||||
- **timeutil** - JSON-serializable duration types
|
||||
- **tracing** - OpenTelemetry distributed tracing setup
|
||||
- **types** - Shared data structures for NTP Pool
|
||||
- **ulid** - Thread-safe ULID generation
|
||||
- **version** - Build metadata and version information
|
||||
- **xff/fastlyxff** - Fastly CDN IP range management
|
||||
|
||||
[](https://pkg.go.dev/go.ntppool.org/common)
|
@ -1,3 +1,14 @@
|
||||
// Package apitls provides TLS certificate management with automatic renewal support.
|
||||
//
|
||||
// This package handles TLS certificate provisioning and management for secure
|
||||
// inter-service communication within the NTP Pool project infrastructure.
|
||||
// It provides both server and client certificate management through the
|
||||
// CertificateProvider interface and includes a trusted CA certificate pool
|
||||
// for validating certificates.
|
||||
//
|
||||
// The package integrates with certman for automatic certificate renewal
|
||||
// and includes embedded CA certificates for establishing trust relationships
|
||||
// between services.
|
||||
package apitls
|
||||
|
||||
import (
|
||||
@ -13,11 +24,32 @@ import (
|
||||
//go:embed ca.pem
|
||||
var caBytes []byte
|
||||
|
||||
// CertificateProvider defines the interface for providing TLS certificates
|
||||
// for both server and client connections. Implementations should handle
|
||||
// certificate retrieval, caching, and renewal as needed.
|
||||
//
|
||||
// This interface supports both server-side certificate provisioning
|
||||
// (via GetCertificate) and client-side certificate authentication
|
||||
// (via GetClientCertificate).
|
||||
type CertificateProvider interface {
|
||||
// GetCertificate retrieves a server certificate based on the client hello information.
|
||||
// This method is typically used in tls.Config.GetCertificate for server-side TLS.
|
||||
GetCertificate(hello *tls.ClientHelloInfo) (*tls.Certificate, error)
|
||||
|
||||
// GetClientCertificate retrieves a client certificate for mutual TLS authentication.
|
||||
// This method is used in tls.Config.GetClientCertificate for client-side TLS.
|
||||
GetClientCertificate(certRequestInfo *tls.CertificateRequestInfo) (*tls.Certificate, error)
|
||||
}
|
||||
|
||||
// CAPool returns a certificate pool containing trusted CA certificates
|
||||
// for validating TLS connections within the NTP Pool infrastructure.
|
||||
//
|
||||
// The CA certificates are embedded in the binary and include the trusted
|
||||
// certificate authorities used for inter-service communication.
|
||||
// This pool should be used in tls.Config.RootCAs for client connections
|
||||
// or tls.Config.ClientCAs for server connections requiring client certificates.
|
||||
//
|
||||
// Returns an error if the embedded CA certificates cannot be parsed or loaded.
|
||||
func CAPool() (*x509.CertPool, error) {
|
||||
capool := x509.NewCertPool()
|
||||
if !capool.AppendCertsFromPEM(caBytes) {
|
||||
@ -30,7 +62,6 @@ func CAPool() (*x509.CertPool, error) {
|
||||
// GetCertman sets up certman for the specified cert / key pair. It is
|
||||
// used in the monitor-api and (for now) in the client
|
||||
func GetCertman(certFile, keyFile string) (*certman.CertMan, error) {
|
||||
|
||||
cm, err := certman.New(certFile, keyFile)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
|
@ -1,5 +1,18 @@
|
||||
// Package config provides NTP Pool specific
|
||||
// configuration tools.
|
||||
// Package config provides environment-based configuration management for NTP Pool services.
|
||||
//
|
||||
// This package handles configuration loading from environment variables and provides
|
||||
// utilities for constructing URLs for web and management interfaces. It supports
|
||||
// deployment-specific settings including hostname configuration, TLS settings,
|
||||
// and deployment modes.
|
||||
//
|
||||
// Configuration is loaded automatically from environment variables:
|
||||
// - deployment_mode: The deployment environment (devel, production, etc.)
|
||||
// - manage_hostname: Hostname for management interface
|
||||
// - web_hostname: Comma-separated list of web hostnames (first is primary)
|
||||
// - manage_tls: Enable TLS for management interface (yes, no, true, false)
|
||||
// - web_tls: Enable TLS for web interface (yes, no, true, false)
|
||||
//
|
||||
// The package includes code generation for accessor methods using the accessory tool.
|
||||
package config
|
||||
|
||||
import (
|
||||
@ -11,8 +24,11 @@ import (
|
||||
"go.ntppool.org/common/logger"
|
||||
)
|
||||
|
||||
//go:generate accessory -type Config
|
||||
//go:generate go tool github.com/masaushi/accessory -type Config
|
||||
|
||||
// Config holds environment-based configuration for NTP Pool services.
|
||||
// It manages hostnames, TLS settings, and deployment modes loaded from
|
||||
// environment variables. The struct includes code-generated accessor methods.
|
||||
type Config struct {
|
||||
deploymentMode string `accessor:"getter"`
|
||||
|
||||
@ -26,6 +42,16 @@ type Config struct {
|
||||
valid bool `accessor:"getter"`
|
||||
}
|
||||
|
||||
// New creates a new Config instance by loading configuration from environment variables.
|
||||
// It automatically parses hostnames, TLS settings, and deployment mode from the environment.
|
||||
// The configuration is considered valid if at least one web hostname is provided.
|
||||
//
|
||||
// Environment variables used:
|
||||
// - deployment_mode: Deployment environment identifier
|
||||
// - manage_hostname: Management interface hostname
|
||||
// - web_hostname: Comma-separated web hostnames (first becomes primary)
|
||||
// - manage_tls: Management interface TLS setting
|
||||
// - web_tls: Web interface TLS setting
|
||||
func New() *Config {
|
||||
c := Config{}
|
||||
c.deploymentMode = os.Getenv("deployment_mode")
|
||||
@ -46,10 +72,30 @@ func New() *Config {
|
||||
return &c
|
||||
}
|
||||
|
||||
// WebURL constructs a complete URL for the web interface using the primary web hostname.
|
||||
// It automatically selects HTTP or HTTPS based on the web_tls configuration setting.
|
||||
//
|
||||
// Parameters:
|
||||
// - path: URL path component (should start with "/")
|
||||
// - query: Optional URL query parameters (can be nil)
|
||||
//
|
||||
// Returns a complete URL string suitable for web interface requests.
|
||||
func (c *Config) WebURL(path string, query *url.Values) string {
|
||||
return baseURL(c.webHostname, c.webTLS, path, query)
|
||||
}
|
||||
|
||||
// ManageURL constructs a complete URL for the management interface using the management hostname.
|
||||
// It automatically selects HTTP or HTTPS based on the manage_tls configuration setting.
|
||||
//
|
||||
// Parameters:
|
||||
// - path: URL path component (should start with "/")
|
||||
// - query: Optional URL query parameters (can be nil)
|
||||
//
|
||||
// Returns a complete URL string suitable for management interface requests.
|
||||
func (c *Config) ManageURL(path string, query *url.Values) string {
|
||||
return baseURL(c.manageHostname, c.webTLS, path, query)
|
||||
}
|
||||
|
||||
func baseURL(host string, tls bool, path string, query *url.Values) string {
|
||||
uri := url.URL{}
|
||||
uri.Host = host
|
||||
|
@ -7,7 +7,6 @@ import (
|
||||
)
|
||||
|
||||
func TestBaseURL(t *testing.T) {
|
||||
|
||||
os.Setenv("web_hostname", "www.ntp.dev, web.ntppool.dev")
|
||||
os.Setenv("web_tls", "yes")
|
||||
|
||||
@ -22,5 +21,4 @@ func TestBaseURL(t *testing.T) {
|
||||
if u != "https://www.ntp.dev/foo?foo=bar" {
|
||||
t.Fatalf("unexpected WebURL: %s", u)
|
||||
}
|
||||
|
||||
}
|
||||
|
18
config/depenv/context.go
Normal file
18
config/depenv/context.go
Normal file
@ -0,0 +1,18 @@
|
||||
package depenv
|
||||
|
||||
import "context"
|
||||
|
||||
type contextKey struct{}
|
||||
|
||||
// NewContext adds the deployment environment to the context
|
||||
func NewContext(ctx context.Context, d DeploymentEnvironment) context.Context {
|
||||
return context.WithValue(ctx, contextKey{}, d)
|
||||
}
|
||||
|
||||
// FromContext retrieves the deployment environment from the context
|
||||
func FromContext(ctx context.Context) DeploymentEnvironment {
|
||||
if d, ok := ctx.Value(contextKey{}).(DeploymentEnvironment); ok {
|
||||
return d
|
||||
}
|
||||
return DeployUndefined
|
||||
}
|
133
config/depenv/depenv.go
Normal file
133
config/depenv/depenv.go
Normal file
@ -0,0 +1,133 @@
|
||||
// Package depenv provides deployment environment management for NTP Pool services.
|
||||
//
|
||||
// This package handles different deployment environments (development, test, production)
|
||||
// and provides environment-specific configuration including API endpoints, management URLs,
|
||||
// and monitoring domains. It supports string-based environment identification and
|
||||
// automatic URL construction for various service endpoints.
|
||||
//
|
||||
// The package defines three main deployment environments:
|
||||
// - DeployDevel: Development environment with dev-specific endpoints
|
||||
// - DeployTest: Test/beta environment for staging
|
||||
// - DeployProd: Production environment with live endpoints
|
||||
//
|
||||
// Environment detection supports both short and long forms:
|
||||
// - "dev" or "devel" → DeployDevel
|
||||
// - "test" or "beta" → DeployTest
|
||||
// - "prod" → DeployProd
|
||||
package depenv
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"os"
|
||||
)
|
||||
|
||||
var manageServers = map[DeploymentEnvironment]string{
|
||||
DeployDevel: "https://manage.askdev.grundclock.com",
|
||||
DeployTest: "https://manage.beta.grundclock.com",
|
||||
DeployProd: "https://manage.ntppool.org",
|
||||
}
|
||||
|
||||
var apiServers = map[DeploymentEnvironment]string{
|
||||
DeployDevel: "https://dev-api.ntppool.dev",
|
||||
DeployTest: "https://beta-api.ntppool.dev",
|
||||
DeployProd: "https://api.ntppool.dev",
|
||||
}
|
||||
|
||||
// var validationServers = map[DeploymentEnvironment]string{
|
||||
// DeployDevel: "https://v.ntp.dev/d/",
|
||||
// DeployTest: "https://v.ntp.dev/b/",
|
||||
// DeployProd: "https://v.ntp.dev/p/",
|
||||
// }
|
||||
|
||||
const (
|
||||
// DeployUndefined represents an unrecognized or unset deployment environment.
|
||||
DeployUndefined DeploymentEnvironment = iota
|
||||
// DeployDevel represents the development environment.
|
||||
DeployDevel
|
||||
// DeployTest represents the test/beta environment.
|
||||
DeployTest
|
||||
// DeployProd represents the production environment.
|
||||
DeployProd
|
||||
)
|
||||
|
||||
// DeploymentEnvironment represents a deployment environment type.
|
||||
// It provides methods for environment-specific URL construction and
|
||||
// supports text marshaling/unmarshaling for configuration files.
|
||||
type DeploymentEnvironment uint8
|
||||
|
||||
// DeploymentEnvironmentFromString parses a string into a DeploymentEnvironment.
|
||||
// It supports both short and long forms of environment names:
|
||||
// - "dev" or "devel" → DeployDevel
|
||||
// - "test" or "beta" → DeployTest
|
||||
// - "prod" → DeployProd
|
||||
// - any other value → DeployUndefined
|
||||
func DeploymentEnvironmentFromString(s string) DeploymentEnvironment {
|
||||
switch s {
|
||||
case "devel", "dev":
|
||||
return DeployDevel
|
||||
case "test", "beta":
|
||||
return DeployTest
|
||||
case "prod":
|
||||
return DeployProd
|
||||
default:
|
||||
return DeployUndefined
|
||||
}
|
||||
}
|
||||
|
||||
// String returns the canonical string representation of the deployment environment.
|
||||
// Returns "prod", "test", "devel", or panics for invalid environments.
|
||||
func (d DeploymentEnvironment) String() string {
|
||||
switch d {
|
||||
case DeployProd:
|
||||
return "prod"
|
||||
case DeployTest:
|
||||
return "test"
|
||||
case DeployDevel:
|
||||
return "devel"
|
||||
default:
|
||||
panic("invalid DeploymentEnvironment")
|
||||
}
|
||||
}
|
||||
|
||||
// APIHost returns the API server URL for this deployment environment.
|
||||
// It first checks the API_HOST environment variable for overrides,
|
||||
// then falls back to the environment-specific default API endpoint.
|
||||
func (d DeploymentEnvironment) APIHost() string {
|
||||
if apiHost := os.Getenv("API_HOST"); apiHost != "" {
|
||||
return apiHost
|
||||
}
|
||||
return apiServers[d]
|
||||
}
|
||||
|
||||
// ManageURL constructs a management interface URL for this deployment environment.
|
||||
// It combines the environment-specific management server base URL with the provided path.
|
||||
//
|
||||
// The path parameter should start with "/" for proper URL construction.
|
||||
func (d DeploymentEnvironment) ManageURL(path string) string {
|
||||
return manageServers[d] + path
|
||||
}
|
||||
|
||||
// MonitorDomain returns the monitoring domain for this deployment environment.
|
||||
// The domain follows the pattern: {environment}.mon.ntppool.dev
|
||||
// For example: "devel.mon.ntppool.dev" for the development environment.
|
||||
func (d DeploymentEnvironment) MonitorDomain() string {
|
||||
return d.String() + ".mon.ntppool.dev"
|
||||
}
|
||||
|
||||
// UnmarshalText implements the encoding.TextUnmarshaler interface.
|
||||
// It allows DeploymentEnvironment to be unmarshaled from configuration files
|
||||
// and other text-based formats. Empty strings are treated as valid (no-op).
|
||||
//
|
||||
// Returns an error if the text represents an invalid deployment environment.
|
||||
func (d *DeploymentEnvironment) UnmarshalText(text []byte) error {
|
||||
s := string(text)
|
||||
if s == "" {
|
||||
return nil
|
||||
}
|
||||
env := DeploymentEnvironmentFromString(s)
|
||||
if env == DeployUndefined {
|
||||
return fmt.Errorf("invalid deployment environment: %s", s)
|
||||
}
|
||||
*d = env
|
||||
return nil
|
||||
}
|
40
config/depenv/monitor_names.go
Normal file
40
config/depenv/monitor_names.go
Normal file
@ -0,0 +1,40 @@
|
||||
package depenv
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"strings"
|
||||
)
|
||||
|
||||
var monitorApiServers = map[DeploymentEnvironment]string{
|
||||
DeployDevel: "https://api.devel.mon.ntppool.dev",
|
||||
DeployTest: "https://api.test.mon.ntppool.dev",
|
||||
DeployProd: "https://api.mon.ntppool.dev",
|
||||
}
|
||||
|
||||
func (d DeploymentEnvironment) MonitorAPIHost() string {
|
||||
return monitorApiServers[d]
|
||||
}
|
||||
|
||||
func GetDeploymentEnvironmentFromName(clientName string) (DeploymentEnvironment, error) {
|
||||
clientName = strings.ToLower(clientName)
|
||||
|
||||
if !strings.HasSuffix(clientName, ".mon.ntppool.dev") {
|
||||
return DeployUndefined, fmt.Errorf("invalid client name %s", clientName)
|
||||
}
|
||||
|
||||
if clientName == "api.mon.ntppool.dev" {
|
||||
return DeployProd, nil
|
||||
}
|
||||
|
||||
prefix := clientName[:strings.Index(clientName, ".mon.ntppool.dev")]
|
||||
parts := strings.Split(prefix, ".")
|
||||
if len(parts) != 2 {
|
||||
return DeployUndefined, fmt.Errorf("invalid client name %s", clientName)
|
||||
}
|
||||
|
||||
if d := DeploymentEnvironmentFromString(parts[1]); d != DeployUndefined {
|
||||
return d, nil
|
||||
}
|
||||
|
||||
return DeployUndefined, fmt.Errorf("invalid client name %s (unknown environment %s)", clientName, parts[1])
|
||||
}
|
72
database/config.go
Normal file
72
database/config.go
Normal file
@ -0,0 +1,72 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"os"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
// Config represents the database configuration structure
|
||||
type Config struct {
|
||||
MySQL DBConfig `yaml:"mysql"`
|
||||
}
|
||||
|
||||
// DBConfig represents the MySQL database configuration
|
||||
type DBConfig struct {
|
||||
DSN string `default:"" flag:"dsn" usage:"Database DSN"`
|
||||
User string `default:"" flag:"user"`
|
||||
Pass string `default:"" flag:"pass"`
|
||||
DBName string // Optional database name override
|
||||
}
|
||||
|
||||
// ConfigOptions allows customization of database opening behavior
|
||||
type ConfigOptions struct {
|
||||
// ConfigFiles is a list of config file paths to search for database configuration
|
||||
ConfigFiles []string
|
||||
|
||||
// EnablePoolMonitoring enables connection pool metrics collection
|
||||
EnablePoolMonitoring bool
|
||||
|
||||
// PrometheusRegisterer for metrics collection. If nil, no metrics are collected.
|
||||
PrometheusRegisterer prometheus.Registerer
|
||||
|
||||
// Connection pool settings
|
||||
MaxOpenConns int
|
||||
MaxIdleConns int
|
||||
ConnMaxLifetime time.Duration
|
||||
}
|
||||
|
||||
// getConfigFiles returns the list of config files to search for database configuration.
|
||||
// If DATABASE_CONFIG_FILE environment variable is set, it returns that single file.
|
||||
// Otherwise, it returns the default paths.
|
||||
func getConfigFiles() []string {
|
||||
if configFile := os.Getenv("DATABASE_CONFIG_FILE"); configFile != "" {
|
||||
return []string{configFile}
|
||||
}
|
||||
return []string{"database.yaml", "/vault/secrets/database.yaml"}
|
||||
}
|
||||
|
||||
// DefaultConfigOptions returns the standard configuration options used by API package
|
||||
func DefaultConfigOptions() ConfigOptions {
|
||||
return ConfigOptions{
|
||||
ConfigFiles: getConfigFiles(),
|
||||
EnablePoolMonitoring: true,
|
||||
PrometheusRegisterer: prometheus.DefaultRegisterer,
|
||||
MaxOpenConns: 25,
|
||||
MaxIdleConns: 10,
|
||||
ConnMaxLifetime: 3 * time.Minute,
|
||||
}
|
||||
}
|
||||
|
||||
// MonitorConfigOptions returns configuration options optimized for Monitor package
|
||||
func MonitorConfigOptions() ConfigOptions {
|
||||
return ConfigOptions{
|
||||
ConfigFiles: getConfigFiles(),
|
||||
EnablePoolMonitoring: false, // Monitor doesn't need metrics
|
||||
PrometheusRegisterer: nil, // No Prometheus dependency
|
||||
MaxOpenConns: 10,
|
||||
MaxIdleConns: 5,
|
||||
ConnMaxLifetime: 3 * time.Minute,
|
||||
}
|
||||
}
|
81
database/config_test.go
Normal file
81
database/config_test.go
Normal file
@ -0,0 +1,81 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
func TestDefaultConfigOptions(t *testing.T) {
|
||||
opts := DefaultConfigOptions()
|
||||
|
||||
// Verify expected defaults for API package
|
||||
if opts.MaxOpenConns != 25 {
|
||||
t.Errorf("Expected MaxOpenConns=25, got %d", opts.MaxOpenConns)
|
||||
}
|
||||
if opts.MaxIdleConns != 10 {
|
||||
t.Errorf("Expected MaxIdleConns=10, got %d", opts.MaxIdleConns)
|
||||
}
|
||||
if opts.ConnMaxLifetime != 3*time.Minute {
|
||||
t.Errorf("Expected ConnMaxLifetime=3m, got %v", opts.ConnMaxLifetime)
|
||||
}
|
||||
if !opts.EnablePoolMonitoring {
|
||||
t.Error("Expected EnablePoolMonitoring=true")
|
||||
}
|
||||
if opts.PrometheusRegisterer != prometheus.DefaultRegisterer {
|
||||
t.Error("Expected PrometheusRegisterer to be DefaultRegisterer")
|
||||
}
|
||||
if len(opts.ConfigFiles) == 0 {
|
||||
t.Error("Expected ConfigFiles to be non-empty")
|
||||
}
|
||||
}
|
||||
|
||||
func TestMonitorConfigOptions(t *testing.T) {
|
||||
opts := MonitorConfigOptions()
|
||||
|
||||
// Verify expected defaults for Monitor package
|
||||
if opts.MaxOpenConns != 10 {
|
||||
t.Errorf("Expected MaxOpenConns=10, got %d", opts.MaxOpenConns)
|
||||
}
|
||||
if opts.MaxIdleConns != 5 {
|
||||
t.Errorf("Expected MaxIdleConns=5, got %d", opts.MaxIdleConns)
|
||||
}
|
||||
if opts.ConnMaxLifetime != 3*time.Minute {
|
||||
t.Errorf("Expected ConnMaxLifetime=3m, got %v", opts.ConnMaxLifetime)
|
||||
}
|
||||
if opts.EnablePoolMonitoring {
|
||||
t.Error("Expected EnablePoolMonitoring=false")
|
||||
}
|
||||
if opts.PrometheusRegisterer != nil {
|
||||
t.Error("Expected PrometheusRegisterer to be nil")
|
||||
}
|
||||
if len(opts.ConfigFiles) == 0 {
|
||||
t.Error("Expected ConfigFiles to be non-empty")
|
||||
}
|
||||
}
|
||||
|
||||
func TestConfigStructures(t *testing.T) {
|
||||
// Test that configuration structures can be created and populated
|
||||
config := Config{
|
||||
MySQL: DBConfig{
|
||||
DSN: "user:pass@tcp(localhost:3306)/dbname",
|
||||
User: "testuser",
|
||||
Pass: "testpass",
|
||||
DBName: "testdb",
|
||||
},
|
||||
}
|
||||
|
||||
if config.MySQL.DSN == "" {
|
||||
t.Error("Expected DSN to be set")
|
||||
}
|
||||
if config.MySQL.User != "testuser" {
|
||||
t.Errorf("Expected User='testuser', got '%s'", config.MySQL.User)
|
||||
}
|
||||
if config.MySQL.Pass != "testpass" {
|
||||
t.Errorf("Expected Pass='testpass', got '%s'", config.MySQL.Pass)
|
||||
}
|
||||
if config.MySQL.DBName != "testdb" {
|
||||
t.Errorf("Expected DBName='testdb', got '%s'", config.MySQL.DBName)
|
||||
}
|
||||
}
|
88
database/connector.go
Normal file
88
database/connector.go
Normal file
@ -0,0 +1,88 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql/driver"
|
||||
"errors"
|
||||
"fmt"
|
||||
"os"
|
||||
|
||||
"github.com/go-sql-driver/mysql"
|
||||
"gopkg.in/yaml.v3"
|
||||
)
|
||||
|
||||
// from https://github.com/Boostport/dynamic-database-config
|
||||
|
||||
// CreateConnectorFunc is a function that creates a database connector
|
||||
type CreateConnectorFunc func() (driver.Connector, error)
|
||||
|
||||
// Driver implements the sql/driver interface with dynamic configuration
|
||||
type Driver struct {
|
||||
CreateConnectorFunc CreateConnectorFunc
|
||||
}
|
||||
|
||||
// Driver returns the driver instance
|
||||
func (d Driver) Driver() driver.Driver {
|
||||
return d
|
||||
}
|
||||
|
||||
// Connect creates a new database connection using the dynamic connector
|
||||
func (d Driver) Connect(ctx context.Context) (driver.Conn, error) {
|
||||
connector, err := d.CreateConnectorFunc()
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("error creating connector from function: %w", err)
|
||||
}
|
||||
|
||||
return connector.Connect(ctx)
|
||||
}
|
||||
|
||||
// Open is not supported for dynamic configuration
|
||||
func (d Driver) Open(name string) (driver.Conn, error) {
|
||||
return nil, errors.New("open is not supported")
|
||||
}
|
||||
|
||||
// createConnector creates a connector function that reads configuration from a file
|
||||
func createConnector(configFile string) CreateConnectorFunc {
|
||||
return func() (driver.Connector, error) {
|
||||
dbFile, err := os.Open(configFile)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
defer dbFile.Close()
|
||||
|
||||
dec := yaml.NewDecoder(dbFile)
|
||||
cfg := Config{}
|
||||
|
||||
err = dec.Decode(&cfg)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
dsn := cfg.MySQL.DSN
|
||||
if len(dsn) == 0 {
|
||||
dsn = os.Getenv("DATABASE_DSN")
|
||||
if len(dsn) == 0 {
|
||||
return nil, fmt.Errorf("dsn config in database.yaml or DATABASE_DSN environment variable required")
|
||||
}
|
||||
}
|
||||
|
||||
dbcfg, err := mysql.ParseDSN(dsn)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if user := cfg.MySQL.User; len(user) > 0 {
|
||||
dbcfg.User = user
|
||||
}
|
||||
|
||||
if pass := cfg.MySQL.Pass; len(pass) > 0 {
|
||||
dbcfg.Passwd = pass
|
||||
}
|
||||
|
||||
if name := cfg.MySQL.DBName; len(name) > 0 {
|
||||
dbcfg.DBName = name
|
||||
}
|
||||
|
||||
return mysql.NewConnector(dbcfg)
|
||||
}
|
||||
}
|
117
database/integration_test.go
Normal file
117
database/integration_test.go
Normal file
@ -0,0 +1,117 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"testing"
|
||||
)
|
||||
|
||||
// Mock types for testing SQLC integration patterns
|
||||
type mockQueries struct {
|
||||
db DBTX
|
||||
}
|
||||
|
||||
type mockQueriesTx struct {
|
||||
*mockQueries
|
||||
tx *sql.Tx
|
||||
}
|
||||
|
||||
// Mock the Begin method pattern that SQLC generates
|
||||
func (q *mockQueries) Begin(ctx context.Context) (*mockQueriesTx, error) {
|
||||
// This would normally be: tx, err := q.db.(*sql.DB).BeginTx(ctx, nil)
|
||||
// For our test, we return a mock
|
||||
return &mockQueriesTx{mockQueries: q, tx: nil}, nil
|
||||
}
|
||||
|
||||
func (qtx *mockQueriesTx) Commit(ctx context.Context) error {
|
||||
return nil // Mock implementation
|
||||
}
|
||||
|
||||
func (qtx *mockQueriesTx) Rollback(ctx context.Context) error {
|
||||
return nil // Mock implementation
|
||||
}
|
||||
|
||||
// This test verifies that our common database interfaces are compatible with SQLC-generated code
|
||||
func TestSQLCIntegration(t *testing.T) {
|
||||
// Test that SQLC's DBTX interface matches our DBTX interface
|
||||
t.Run("DBTX Interface Compatibility", func(t *testing.T) {
|
||||
// Test interface compatibility by assignment without execution
|
||||
var ourDBTX DBTX
|
||||
|
||||
// Test with sql.DB (should implement DBTX)
|
||||
var db *sql.DB
|
||||
ourDBTX = db // This will compile only if interfaces are compatible
|
||||
_ = ourDBTX // Use the variable to avoid "unused" warning
|
||||
|
||||
// Test with sql.Tx (should implement DBTX)
|
||||
var tx *sql.Tx
|
||||
ourDBTX = tx // This will compile only if interfaces are compatible
|
||||
_ = ourDBTX // Use the variable to avoid "unused" warning
|
||||
|
||||
// If we reach here, interfaces are compatible
|
||||
t.Log("DBTX interface is compatible with sql.DB and sql.Tx")
|
||||
})
|
||||
|
||||
t.Run("Transaction Interface Compatibility", func(t *testing.T) {
|
||||
// This test verifies our transaction interfaces work with SQLC patterns
|
||||
// We can't define methods inside a function, so we test interface compatibility
|
||||
|
||||
// Verify our DB interface is compatible with what SQLC expects
|
||||
var dbInterface DB[*mockQueriesTx]
|
||||
var mockDB *mockQueries = &mockQueries{}
|
||||
dbInterface = mockDB
|
||||
|
||||
// Test that our transaction helper can work with this pattern
|
||||
err := WithTransaction(context.Background(), dbInterface, func(ctx context.Context, qtx *mockQueriesTx) error {
|
||||
// This would be where you'd call SQLC-generated query methods
|
||||
return nil
|
||||
})
|
||||
if err != nil {
|
||||
t.Errorf("Transaction helper failed: %v", err)
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
// Test that demonstrates how the common package would be used with real SQLC patterns
|
||||
func TestRealWorldUsagePattern(t *testing.T) {
|
||||
// This test shows how a package would typically use our common database code
|
||||
|
||||
t.Run("Database Opening Pattern", func(t *testing.T) {
|
||||
// Test that our configuration options work as expected
|
||||
opts := DefaultConfigOptions()
|
||||
|
||||
// Modify for test environment (no actual database connection)
|
||||
opts.ConfigFiles = []string{} // No config files for unit test
|
||||
opts.PrometheusRegisterer = nil // No metrics for unit test
|
||||
|
||||
// This would normally open a database: db, err := OpenDB(ctx, opts)
|
||||
// For our unit test, we just verify the options are reasonable
|
||||
if opts.MaxOpenConns <= 0 {
|
||||
t.Error("MaxOpenConns should be positive")
|
||||
}
|
||||
if opts.MaxIdleConns <= 0 {
|
||||
t.Error("MaxIdleConns should be positive")
|
||||
}
|
||||
if opts.ConnMaxLifetime <= 0 {
|
||||
t.Error("ConnMaxLifetime should be positive")
|
||||
}
|
||||
})
|
||||
|
||||
t.Run("Monitor Package Configuration", func(t *testing.T) {
|
||||
opts := MonitorConfigOptions()
|
||||
|
||||
// Verify monitor-specific settings
|
||||
if opts.EnablePoolMonitoring {
|
||||
t.Error("Monitor package should not enable pool monitoring")
|
||||
}
|
||||
if opts.PrometheusRegisterer != nil {
|
||||
t.Error("Monitor package should not have Prometheus registerer")
|
||||
}
|
||||
if opts.MaxOpenConns != 10 {
|
||||
t.Errorf("Expected MaxOpenConns=10 for monitor, got %d", opts.MaxOpenConns)
|
||||
}
|
||||
if opts.MaxIdleConns != 5 {
|
||||
t.Errorf("Expected MaxIdleConns=5 for monitor, got %d", opts.MaxIdleConns)
|
||||
}
|
||||
})
|
||||
}
|
34
database/interfaces.go
Normal file
34
database/interfaces.go
Normal file
@ -0,0 +1,34 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
)
|
||||
|
||||
// DBTX matches the interface expected by SQLC-generated code
|
||||
// This interface is implemented by both *sql.DB and *sql.Tx
|
||||
type DBTX interface {
|
||||
ExecContext(context.Context, string, ...interface{}) (sql.Result, error)
|
||||
PrepareContext(context.Context, string) (*sql.Stmt, error)
|
||||
QueryContext(context.Context, string, ...interface{}) (*sql.Rows, error)
|
||||
QueryRowContext(context.Context, string, ...interface{}) *sql.Row
|
||||
}
|
||||
|
||||
// BaseQuerier provides basic query functionality
|
||||
// This interface should be implemented by package-specific Queries types
|
||||
type BaseQuerier interface {
|
||||
WithTx(tx *sql.Tx) BaseQuerier
|
||||
}
|
||||
|
||||
// BaseQuerierTx provides transaction functionality
|
||||
// This interface should be implemented by package-specific Queries types
|
||||
type BaseQuerierTx interface {
|
||||
BaseQuerier
|
||||
Begin(ctx context.Context) (BaseQuerierTx, error)
|
||||
Commit(ctx context.Context) error
|
||||
Rollback(ctx context.Context) error
|
||||
}
|
||||
|
||||
// TransactionFunc represents a function that operates within a database transaction
|
||||
// This is used by the shared transaction helpers in transaction.go
|
||||
type TransactionFunc[Q any] func(ctx context.Context, q Q) error
|
93
database/metrics.go
Normal file
93
database/metrics.go
Normal file
@ -0,0 +1,93 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"fmt"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
// DatabaseMetrics holds the Prometheus metrics for database connection pool monitoring
|
||||
type DatabaseMetrics struct {
|
||||
ConnectionsOpen prometheus.Gauge
|
||||
ConnectionsIdle prometheus.Gauge
|
||||
ConnectionsInUse prometheus.Gauge
|
||||
ConnectionsWaitCount prometheus.Counter
|
||||
ConnectionsWaitDuration prometheus.Histogram
|
||||
}
|
||||
|
||||
// NewDatabaseMetrics creates a new set of database metrics and registers them
|
||||
func NewDatabaseMetrics(registerer prometheus.Registerer) *DatabaseMetrics {
|
||||
metrics := &DatabaseMetrics{
|
||||
ConnectionsOpen: prometheus.NewGauge(prometheus.GaugeOpts{
|
||||
Name: "database_connections_open",
|
||||
Help: "Number of open database connections",
|
||||
}),
|
||||
ConnectionsIdle: prometheus.NewGauge(prometheus.GaugeOpts{
|
||||
Name: "database_connections_idle",
|
||||
Help: "Number of idle database connections",
|
||||
}),
|
||||
ConnectionsInUse: prometheus.NewGauge(prometheus.GaugeOpts{
|
||||
Name: "database_connections_in_use",
|
||||
Help: "Number of database connections in use",
|
||||
}),
|
||||
ConnectionsWaitCount: prometheus.NewCounter(prometheus.CounterOpts{
|
||||
Name: "database_connections_wait_count_total",
|
||||
Help: "Total number of times a connection had to wait",
|
||||
}),
|
||||
ConnectionsWaitDuration: prometheus.NewHistogram(prometheus.HistogramOpts{
|
||||
Name: "database_connections_wait_duration_seconds",
|
||||
Help: "Time spent waiting for a database connection",
|
||||
Buckets: prometheus.DefBuckets,
|
||||
}),
|
||||
}
|
||||
|
||||
if registerer != nil {
|
||||
registerer.MustRegister(
|
||||
metrics.ConnectionsOpen,
|
||||
metrics.ConnectionsIdle,
|
||||
metrics.ConnectionsInUse,
|
||||
metrics.ConnectionsWaitCount,
|
||||
metrics.ConnectionsWaitDuration,
|
||||
)
|
||||
}
|
||||
|
||||
return metrics
|
||||
}
|
||||
|
||||
// monitorConnectionPool runs a background goroutine to collect connection pool metrics
|
||||
func monitorConnectionPool(ctx context.Context, db *sql.DB, registerer prometheus.Registerer) {
|
||||
if registerer == nil {
|
||||
return // No metrics collection if no registerer provided
|
||||
}
|
||||
|
||||
metrics := NewDatabaseMetrics(registerer)
|
||||
ticker := time.NewTicker(30 * time.Second)
|
||||
defer ticker.Stop()
|
||||
|
||||
for {
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
return
|
||||
case <-ticker.C:
|
||||
stats := db.Stats()
|
||||
|
||||
metrics.ConnectionsOpen.Set(float64(stats.OpenConnections))
|
||||
metrics.ConnectionsIdle.Set(float64(stats.Idle))
|
||||
metrics.ConnectionsInUse.Set(float64(stats.InUse))
|
||||
metrics.ConnectionsWaitCount.Add(float64(stats.WaitCount))
|
||||
|
||||
if stats.WaitDuration > 0 {
|
||||
metrics.ConnectionsWaitDuration.Observe(stats.WaitDuration.Seconds())
|
||||
}
|
||||
|
||||
// Log connection pool stats for high usage or waiting
|
||||
if stats.OpenConnections > 20 || stats.WaitCount > 0 {
|
||||
fmt.Printf("Connection pool stats: open=%d idle=%d in_use=%d wait_count=%d wait_duration=%s\n",
|
||||
stats.OpenConnections, stats.Idle, stats.InUse, stats.WaitCount, stats.WaitDuration)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
78
database/pool.go
Normal file
78
database/pool.go
Normal file
@ -0,0 +1,78 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"fmt"
|
||||
"os"
|
||||
|
||||
"go.ntppool.org/common/logger"
|
||||
)
|
||||
|
||||
// OpenDB opens a database connection with the specified configuration options
|
||||
func OpenDB(ctx context.Context, options ConfigOptions) (*sql.DB, error) {
|
||||
log := logger.Setup()
|
||||
|
||||
configFile, err := findConfigFile(options.ConfigFiles)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
dbconn := sql.OpenDB(Driver{
|
||||
CreateConnectorFunc: createConnector(configFile),
|
||||
})
|
||||
|
||||
// Set connection pool parameters
|
||||
dbconn.SetConnMaxLifetime(options.ConnMaxLifetime)
|
||||
dbconn.SetMaxOpenConns(options.MaxOpenConns)
|
||||
dbconn.SetMaxIdleConns(options.MaxIdleConns)
|
||||
|
||||
err = dbconn.Ping()
|
||||
if err != nil {
|
||||
log.Error("could not connect to database", "err", err)
|
||||
return nil, err
|
||||
}
|
||||
|
||||
// Start optional connection pool monitoring
|
||||
if options.EnablePoolMonitoring && options.PrometheusRegisterer != nil {
|
||||
go monitorConnectionPool(ctx, dbconn, options.PrometheusRegisterer)
|
||||
}
|
||||
|
||||
return dbconn, nil
|
||||
}
|
||||
|
||||
// OpenDBWithConfigFile opens a database connection using an explicit config file path
|
||||
// This is a convenience function for API package compatibility
|
||||
func OpenDBWithConfigFile(ctx context.Context, configFile string) (*sql.DB, error) {
|
||||
options := DefaultConfigOptions()
|
||||
options.ConfigFiles = []string{configFile}
|
||||
return OpenDB(ctx, options)
|
||||
}
|
||||
|
||||
// OpenDBMonitor opens a database connection with monitor-specific defaults
|
||||
// This is a convenience function for Monitor package compatibility
|
||||
func OpenDBMonitor() (*sql.DB, error) {
|
||||
options := MonitorConfigOptions()
|
||||
return OpenDB(context.Background(), options)
|
||||
}
|
||||
|
||||
// findConfigFile searches for the first existing config file from the list
|
||||
func findConfigFile(configFiles []string) (string, error) {
|
||||
var firstErr error
|
||||
|
||||
for _, configFile := range configFiles {
|
||||
if configFile == "" {
|
||||
continue
|
||||
}
|
||||
if _, err := os.Stat(configFile); err == nil {
|
||||
return configFile, nil
|
||||
} else if firstErr == nil {
|
||||
firstErr = err
|
||||
}
|
||||
}
|
||||
|
||||
if firstErr != nil {
|
||||
return "", fmt.Errorf("no config file found: %w", firstErr)
|
||||
}
|
||||
return "", fmt.Errorf("no valid config files provided")
|
||||
}
|
69
database/transaction.go
Normal file
69
database/transaction.go
Normal file
@ -0,0 +1,69 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
|
||||
"go.ntppool.org/common/logger"
|
||||
)
|
||||
|
||||
// DB interface for database operations that can begin transactions
|
||||
type DB[Q any] interface {
|
||||
Begin(ctx context.Context) (Q, error)
|
||||
}
|
||||
|
||||
// TX interface for transaction operations
|
||||
type TX interface {
|
||||
Commit(ctx context.Context) error
|
||||
Rollback(ctx context.Context) error
|
||||
}
|
||||
|
||||
// WithTransaction executes a function within a database transaction
|
||||
// Handles proper rollback on error and commit on success
|
||||
func WithTransaction[Q TX](ctx context.Context, db DB[Q], fn func(ctx context.Context, q Q) error) error {
|
||||
tx, err := db.Begin(ctx)
|
||||
if err != nil {
|
||||
return fmt.Errorf("failed to begin transaction: %w", err)
|
||||
}
|
||||
|
||||
var committed bool
|
||||
defer func() {
|
||||
if !committed {
|
||||
if rbErr := tx.Rollback(ctx); rbErr != nil {
|
||||
// Log rollback error but don't override original error
|
||||
log := logger.FromContext(ctx)
|
||||
log.ErrorContext(ctx, "failed to rollback transaction", "error", rbErr)
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
if err := fn(ctx, tx); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
err = tx.Commit(ctx)
|
||||
committed = true // Mark as committed regardless of commit success/failure
|
||||
if err != nil {
|
||||
return fmt.Errorf("failed to commit transaction: %w", err)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// WithReadOnlyTransaction executes a read-only function within a transaction
|
||||
// Always rolls back at the end (for consistent read isolation)
|
||||
func WithReadOnlyTransaction[Q TX](ctx context.Context, db DB[Q], fn func(ctx context.Context, q Q) error) error {
|
||||
tx, err := db.Begin(ctx)
|
||||
if err != nil {
|
||||
return fmt.Errorf("failed to begin read-only transaction: %w", err)
|
||||
}
|
||||
|
||||
defer func() {
|
||||
if rbErr := tx.Rollback(ctx); rbErr != nil {
|
||||
log := logger.FromContext(ctx)
|
||||
log.ErrorContext(ctx, "failed to rollback read-only transaction", "error", rbErr)
|
||||
}
|
||||
}()
|
||||
|
||||
return fn(ctx, tx)
|
||||
}
|
69
database/transaction_base.go
Normal file
69
database/transaction_base.go
Normal file
@ -0,0 +1,69 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"fmt"
|
||||
|
||||
"go.ntppool.org/common/logger"
|
||||
)
|
||||
|
||||
// Shared interface definitions that both packages use identically
|
||||
type BaseBeginner interface {
|
||||
Begin(context.Context) (sql.Tx, error)
|
||||
}
|
||||
|
||||
type BaseTx interface {
|
||||
BaseBeginner
|
||||
Commit(ctx context.Context) error
|
||||
Rollback(ctx context.Context) error
|
||||
}
|
||||
|
||||
// BeginTransactionForQuerier contains the shared Begin() logic from both packages
|
||||
func BeginTransactionForQuerier(ctx context.Context, db DBTX) (DBTX, error) {
|
||||
if sqlDB, ok := db.(*sql.DB); ok {
|
||||
tx, err := sqlDB.BeginTx(ctx, &sql.TxOptions{})
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return tx, nil
|
||||
} else {
|
||||
// Handle transaction case
|
||||
if beginner, ok := db.(BaseBeginner); ok {
|
||||
tx, err := beginner.Begin(ctx)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
return &tx, nil
|
||||
}
|
||||
return nil, fmt.Errorf("database connection does not support transactions")
|
||||
}
|
||||
}
|
||||
|
||||
// CommitTransactionForQuerier contains the shared Commit() logic from both packages
|
||||
func CommitTransactionForQuerier(ctx context.Context, db DBTX) error {
|
||||
if sqlTx, ok := db.(*sql.Tx); ok {
|
||||
return sqlTx.Commit()
|
||||
}
|
||||
|
||||
tx, ok := db.(BaseTx)
|
||||
if !ok {
|
||||
log := logger.FromContext(ctx)
|
||||
log.ErrorContext(ctx, "could not get a Tx", "type", fmt.Sprintf("%T", db))
|
||||
return sql.ErrTxDone
|
||||
}
|
||||
return tx.Commit(ctx)
|
||||
}
|
||||
|
||||
// RollbackTransactionForQuerier contains the shared Rollback() logic from both packages
|
||||
func RollbackTransactionForQuerier(ctx context.Context, db DBTX) error {
|
||||
if sqlTx, ok := db.(*sql.Tx); ok {
|
||||
return sqlTx.Rollback()
|
||||
}
|
||||
|
||||
tx, ok := db.(BaseTx)
|
||||
if !ok {
|
||||
return sql.ErrTxDone
|
||||
}
|
||||
return tx.Rollback(ctx)
|
||||
}
|
157
database/transaction_test.go
Normal file
157
database/transaction_test.go
Normal file
@ -0,0 +1,157 @@
|
||||
package database
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"testing"
|
||||
)
|
||||
|
||||
// Mock implementations for testing
|
||||
type mockDB struct {
|
||||
beginError error
|
||||
txMock *mockTX
|
||||
}
|
||||
|
||||
func (m *mockDB) Begin(ctx context.Context) (*mockTX, error) {
|
||||
if m.beginError != nil {
|
||||
return nil, m.beginError
|
||||
}
|
||||
return m.txMock, nil
|
||||
}
|
||||
|
||||
type mockTX struct {
|
||||
commitError error
|
||||
rollbackError error
|
||||
commitCalled bool
|
||||
rollbackCalled bool
|
||||
}
|
||||
|
||||
func (m *mockTX) Commit(ctx context.Context) error {
|
||||
m.commitCalled = true
|
||||
return m.commitError
|
||||
}
|
||||
|
||||
func (m *mockTX) Rollback(ctx context.Context) error {
|
||||
m.rollbackCalled = true
|
||||
return m.rollbackError
|
||||
}
|
||||
|
||||
func TestWithTransaction_Success(t *testing.T) {
|
||||
tx := &mockTX{}
|
||||
db := &mockDB{txMock: tx}
|
||||
|
||||
var functionCalled bool
|
||||
err := WithTransaction(context.Background(), db, func(ctx context.Context, q *mockTX) error {
|
||||
functionCalled = true
|
||||
if q != tx {
|
||||
t.Error("Expected transaction to be passed to function")
|
||||
}
|
||||
return nil
|
||||
})
|
||||
if err != nil {
|
||||
t.Errorf("Expected no error, got %v", err)
|
||||
}
|
||||
if !functionCalled {
|
||||
t.Error("Expected function to be called")
|
||||
}
|
||||
if !tx.commitCalled {
|
||||
t.Error("Expected commit to be called")
|
||||
}
|
||||
if tx.rollbackCalled {
|
||||
t.Error("Expected rollback NOT to be called on success")
|
||||
}
|
||||
}
|
||||
|
||||
func TestWithTransaction_FunctionError(t *testing.T) {
|
||||
tx := &mockTX{}
|
||||
db := &mockDB{txMock: tx}
|
||||
|
||||
expectedError := errors.New("function error")
|
||||
err := WithTransaction(context.Background(), db, func(ctx context.Context, q *mockTX) error {
|
||||
return expectedError
|
||||
})
|
||||
|
||||
if err != expectedError {
|
||||
t.Errorf("Expected error %v, got %v", expectedError, err)
|
||||
}
|
||||
if tx.commitCalled {
|
||||
t.Error("Expected commit NOT to be called on function error")
|
||||
}
|
||||
if !tx.rollbackCalled {
|
||||
t.Error("Expected rollback to be called on function error")
|
||||
}
|
||||
}
|
||||
|
||||
func TestWithTransaction_BeginError(t *testing.T) {
|
||||
expectedError := errors.New("begin error")
|
||||
db := &mockDB{beginError: expectedError}
|
||||
|
||||
err := WithTransaction(context.Background(), db, func(ctx context.Context, q *mockTX) error {
|
||||
t.Error("Function should not be called when Begin fails")
|
||||
return nil
|
||||
})
|
||||
|
||||
if err == nil || !errors.Is(err, expectedError) {
|
||||
t.Errorf("Expected wrapped begin error, got %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestWithTransaction_CommitError(t *testing.T) {
|
||||
commitError := errors.New("commit error")
|
||||
tx := &mockTX{commitError: commitError}
|
||||
db := &mockDB{txMock: tx}
|
||||
|
||||
err := WithTransaction(context.Background(), db, func(ctx context.Context, q *mockTX) error {
|
||||
return nil
|
||||
})
|
||||
|
||||
if err == nil || !errors.Is(err, commitError) {
|
||||
t.Errorf("Expected wrapped commit error, got %v", err)
|
||||
}
|
||||
if !tx.commitCalled {
|
||||
t.Error("Expected commit to be called")
|
||||
}
|
||||
if tx.rollbackCalled {
|
||||
t.Error("Expected rollback NOT to be called when commit fails")
|
||||
}
|
||||
}
|
||||
|
||||
func TestWithReadOnlyTransaction_Success(t *testing.T) {
|
||||
tx := &mockTX{}
|
||||
db := &mockDB{txMock: tx}
|
||||
|
||||
var functionCalled bool
|
||||
err := WithReadOnlyTransaction(context.Background(), db, func(ctx context.Context, q *mockTX) error {
|
||||
functionCalled = true
|
||||
return nil
|
||||
})
|
||||
if err != nil {
|
||||
t.Errorf("Expected no error, got %v", err)
|
||||
}
|
||||
if !functionCalled {
|
||||
t.Error("Expected function to be called")
|
||||
}
|
||||
if tx.commitCalled {
|
||||
t.Error("Expected commit NOT to be called in read-only transaction")
|
||||
}
|
||||
if !tx.rollbackCalled {
|
||||
t.Error("Expected rollback to be called in read-only transaction")
|
||||
}
|
||||
}
|
||||
|
||||
func TestWithReadOnlyTransaction_FunctionError(t *testing.T) {
|
||||
tx := &mockTX{}
|
||||
db := &mockDB{txMock: tx}
|
||||
|
||||
expectedError := errors.New("function error")
|
||||
err := WithReadOnlyTransaction(context.Background(), db, func(ctx context.Context, q *mockTX) error {
|
||||
return expectedError
|
||||
})
|
||||
|
||||
if err != expectedError {
|
||||
t.Errorf("Expected error %v, got %v", expectedError, err)
|
||||
}
|
||||
if !tx.rollbackCalled {
|
||||
t.Error("Expected rollback to be called")
|
||||
}
|
||||
}
|
78
ekko/ekko.go
78
ekko/ekko.go
@ -1,3 +1,32 @@
|
||||
// Package ekko provides an enhanced Echo web framework wrapper with pre-configured middleware.
|
||||
//
|
||||
// This package wraps the Echo web framework with a comprehensive middleware stack including:
|
||||
// - OpenTelemetry distributed tracing with request context propagation
|
||||
// - Prometheus metrics collection with per-service subsystems
|
||||
// - Structured logging with trace ID correlation
|
||||
// - Security headers (HSTS, content security policy)
|
||||
// - Gzip compression for response optimization
|
||||
// - Recovery middleware with detailed error logging
|
||||
// - HTTP/2 support with H2C (HTTP/2 Cleartext) capability
|
||||
//
|
||||
// The package uses functional options pattern for flexible configuration
|
||||
// and supports graceful shutdown with configurable timeouts. It's designed
|
||||
// as the standard web service foundation for NTP Pool project services.
|
||||
//
|
||||
// Example usage:
|
||||
//
|
||||
// ekko, err := ekko.New("myservice",
|
||||
// ekko.WithPort(8080),
|
||||
// ekko.WithPrometheus(prometheus.DefaultRegisterer),
|
||||
// ekko.WithEchoSetup(func(e *echo.Echo) error {
|
||||
// e.GET("/health", healthHandler)
|
||||
// return nil
|
||||
// }),
|
||||
// )
|
||||
// if err != nil {
|
||||
// log.Fatal(err)
|
||||
// }
|
||||
// err = ekko.Start(ctx)
|
||||
package ekko
|
||||
|
||||
import (
|
||||
@ -16,9 +45,29 @@ import (
|
||||
"go.opentelemetry.io/contrib/instrumentation/github.com/labstack/echo/otelecho"
|
||||
"go.opentelemetry.io/otel/attribute"
|
||||
"go.opentelemetry.io/otel/trace"
|
||||
"golang.org/x/net/http2"
|
||||
"golang.org/x/sync/errgroup"
|
||||
)
|
||||
|
||||
// New creates a new Ekko instance with the specified service name and functional options.
|
||||
// The name parameter is used for OpenTelemetry service identification, Prometheus metrics
|
||||
// subsystem naming, and server identification headers.
|
||||
//
|
||||
// Default configuration includes:
|
||||
// - 60 second write timeout
|
||||
// - 30 second read header timeout
|
||||
// - HTTP/2 support with H2C
|
||||
// - Standard middleware stack (tracing, metrics, logging, security)
|
||||
//
|
||||
// Use functional options to customize behavior:
|
||||
// - WithPort(): Set server port (required for Start())
|
||||
// - WithPrometheus(): Enable Prometheus metrics
|
||||
// - WithEchoSetup(): Configure routes and handlers
|
||||
// - WithLogFilters(): Filter access logs
|
||||
// - WithOtelMiddleware(): Custom OpenTelemetry middleware
|
||||
// - WithWriteTimeout(): Custom write timeout
|
||||
// - WithReadHeaderTimeout(): Custom read header timeout
|
||||
// - WithGzipConfig(): Custom gzip compression settings
|
||||
func New(name string, options ...func(*Ekko)) (*Ekko, error) {
|
||||
ek := &Ekko{
|
||||
writeTimeout: 60 * time.Second,
|
||||
@ -31,13 +80,25 @@ func New(name string, options ...func(*Ekko)) (*Ekko, error) {
|
||||
return ek, nil
|
||||
}
|
||||
|
||||
// Setup Echo; only intended for testing
|
||||
// SetupEcho creates and configures an Echo instance without starting the server.
|
||||
// This method is primarily intended for testing scenarios where you need access
|
||||
// to the configured Echo instance without starting the HTTP server.
|
||||
//
|
||||
// The returned Echo instance includes all configured middleware and routes
|
||||
// but requires manual server lifecycle management.
|
||||
func (ek *Ekko) SetupEcho(ctx context.Context) (*echo.Echo, error) {
|
||||
return ek.setup(ctx)
|
||||
}
|
||||
|
||||
// Setup Echo and start the server. Will return if the http server
|
||||
// returns or the context is done.
|
||||
// Start creates the Echo instance and starts the HTTP server with graceful shutdown support.
|
||||
// The server runs until either an error occurs or the provided context is cancelled.
|
||||
//
|
||||
// The server supports HTTP/2 with H2C (HTTP/2 Cleartext) and includes a 5-second
|
||||
// graceful shutdown timeout when the context is cancelled. Server configuration
|
||||
// (port, timeouts, middleware) must be set via functional options during New().
|
||||
//
|
||||
// Returns an error if server startup fails or if shutdown doesn't complete within
|
||||
// the timeout period. Returns nil for clean shutdown via context cancellation.
|
||||
func (ek *Ekko) Start(ctx context.Context) error {
|
||||
log := logger.Setup()
|
||||
|
||||
@ -50,7 +111,8 @@ func (ek *Ekko) Start(ctx context.Context) error {
|
||||
g.Go(func() error {
|
||||
e.Server.Addr = fmt.Sprintf(":%d", ek.port)
|
||||
log.Info("server starting", "port", ek.port)
|
||||
err := e.Server.ListenAndServe()
|
||||
// err := e.Server.ListenAndServe()
|
||||
err := e.StartH2CServer(e.Server.Addr, &http2.Server{})
|
||||
if err == http.ErrServerClosed {
|
||||
return nil
|
||||
}
|
||||
@ -120,7 +182,13 @@ func (ek *Ekko) setup(ctx context.Context) (*echo.Echo, error) {
|
||||
e.Use(middleware.Gzip())
|
||||
}
|
||||
|
||||
e.Use(middleware.Secure())
|
||||
secureConfig := middleware.DefaultSecureConfig
|
||||
// secureConfig.ContentSecurityPolicy = "default-src *"
|
||||
secureConfig.ContentSecurityPolicy = ""
|
||||
secureConfig.HSTSMaxAge = int(time.Hour * 168 * 30 / time.Second)
|
||||
secureConfig.HSTSPreloadEnabled = true
|
||||
|
||||
e.Use(middleware.SecureWithConfig(secureConfig))
|
||||
|
||||
e.Use(
|
||||
func(next echo.HandlerFunc) echo.HandlerFunc {
|
||||
|
@ -9,6 +9,9 @@ import (
|
||||
slogecho "github.com/samber/slog-echo"
|
||||
)
|
||||
|
||||
// Ekko represents an enhanced Echo web server with pre-configured middleware stack.
|
||||
// It encapsulates server configuration, middleware options, and lifecycle management
|
||||
// for NTP Pool web services. Use New() with functional options to configure.
|
||||
type Ekko struct {
|
||||
name string
|
||||
prom prometheus.Registerer
|
||||
@ -22,50 +25,76 @@ type Ekko struct {
|
||||
readHeaderTimeout time.Duration
|
||||
}
|
||||
|
||||
// RouteFn defines a function type for configuring Echo routes and handlers.
|
||||
// It receives a configured Echo instance and should register all application
|
||||
// routes, middleware, and handlers. Return an error to abort server startup.
|
||||
type RouteFn func(e *echo.Echo) error
|
||||
|
||||
// WithPort sets the HTTP server port. This option is required when using Start().
|
||||
// The port should be available and the process should have permission to bind to it.
|
||||
func WithPort(port int) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.port = port
|
||||
}
|
||||
}
|
||||
|
||||
// WithPrometheus enables Prometheus metrics collection using the provided registerer.
|
||||
// Metrics include HTTP request duration, request count, and response size histograms.
|
||||
// The service name is used as the metrics subsystem for namespacing.
|
||||
func WithPrometheus(reg prometheus.Registerer) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.prom = reg
|
||||
}
|
||||
}
|
||||
|
||||
// WithEchoSetup configures application routes and handlers via a setup function.
|
||||
// The provided function receives the configured Echo instance after all middleware
|
||||
// is applied and should register routes, custom middleware, and handlers.
|
||||
func WithEchoSetup(rfn RouteFn) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.routeFn = rfn
|
||||
}
|
||||
}
|
||||
|
||||
// WithLogFilters configures access log filtering to reduce log noise.
|
||||
// Filters can exclude specific paths, methods, or status codes from access logs.
|
||||
// Useful for excluding health checks, metrics endpoints, and other high-frequency requests.
|
||||
func WithLogFilters(f []slogecho.Filter) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.logFilters = f
|
||||
}
|
||||
}
|
||||
|
||||
// WithOtelMiddleware replaces the default OpenTelemetry middleware with a custom implementation.
|
||||
// The default middleware provides distributed tracing for all requests. Use this option
|
||||
// when you need custom trace configuration or want to disable tracing entirely.
|
||||
func WithOtelMiddleware(mw echo.MiddlewareFunc) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.otelmiddleware = mw
|
||||
}
|
||||
}
|
||||
|
||||
// WithWriteTimeout configures the HTTP server write timeout.
|
||||
// This is the maximum duration before timing out writes of the response.
|
||||
// Default is 60 seconds. Should be longer than expected response generation time.
|
||||
func WithWriteTimeout(t time.Duration) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.writeTimeout = t
|
||||
}
|
||||
}
|
||||
|
||||
// WithReadHeaderTimeout configures the HTTP server read header timeout.
|
||||
// This is the amount of time allowed to read request headers.
|
||||
// Default is 30 seconds. Should be sufficient for slow clients and large headers.
|
||||
func WithReadHeaderTimeout(t time.Duration) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.readHeaderTimeout = t
|
||||
}
|
||||
}
|
||||
|
||||
// WithGzipConfig provides custom gzip compression configuration.
|
||||
// By default, gzip compression is enabled with standard settings.
|
||||
// Use this option to customize compression level, skip patterns, or disable compression.
|
||||
func WithGzipConfig(gzipConfig *middleware.GzipConfig) func(*Ekko) {
|
||||
return func(ek *Ekko) {
|
||||
ek.gzipConfig = gzipConfig
|
||||
|
21
go.mod
21
go.mod
@ -1,15 +1,15 @@
|
||||
module go.ntppool.org/common
|
||||
|
||||
go 1.23
|
||||
|
||||
toolchain go1.23.4
|
||||
go 1.23.5
|
||||
|
||||
require (
|
||||
github.com/abh/certman v0.4.0
|
||||
github.com/go-sql-driver/mysql v1.9.3
|
||||
github.com/labstack/echo-contrib v0.17.2
|
||||
github.com/labstack/echo/v4 v4.13.3
|
||||
github.com/oklog/ulid/v2 v2.1.0
|
||||
github.com/prometheus/client_golang v1.20.5
|
||||
github.com/prometheus/client_model v0.6.1
|
||||
github.com/remychantenay/slog-otel v1.3.2
|
||||
github.com/samber/slog-echo v1.14.8
|
||||
github.com/samber/slog-multi v1.2.4
|
||||
@ -19,20 +19,28 @@ require (
|
||||
go.opentelemetry.io/contrib/exporters/autoexport v0.58.0
|
||||
go.opentelemetry.io/contrib/instrumentation/github.com/labstack/echo/otelecho v0.58.0
|
||||
go.opentelemetry.io/otel v1.33.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlplog/otlploggrpc v0.9.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlplog/otlploghttp v0.9.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlpmetric/otlpmetricgrpc v1.33.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlpmetric/otlpmetrichttp v1.33.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlptrace v1.33.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracegrpc v1.33.0
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracehttp v1.33.0
|
||||
go.opentelemetry.io/otel/log v0.9.0
|
||||
go.opentelemetry.io/otel/metric v1.33.0
|
||||
go.opentelemetry.io/otel/sdk v1.33.0
|
||||
go.opentelemetry.io/otel/sdk/log v0.9.0
|
||||
go.opentelemetry.io/otel/sdk/metric v1.33.0
|
||||
go.opentelemetry.io/otel/trace v1.33.0
|
||||
golang.org/x/mod v0.22.0
|
||||
golang.org/x/net v0.33.0
|
||||
golang.org/x/sync v0.10.0
|
||||
google.golang.org/grpc v1.69.2
|
||||
gopkg.in/yaml.v3 v3.0.1
|
||||
)
|
||||
|
||||
require (
|
||||
filippo.io/edwards25519 v1.1.0 // indirect
|
||||
github.com/beorn7/perks v1.0.1 // indirect
|
||||
github.com/cenkalti/backoff/v4 v4.3.0 // indirect
|
||||
github.com/cespare/xxhash/v2 v2.3.0 // indirect
|
||||
@ -49,7 +57,6 @@ require (
|
||||
github.com/munnerz/goautoneg v0.0.0-20191010083416-a7dc8b61c822 // indirect
|
||||
github.com/pierrec/lz4/v4 v4.1.22 // indirect
|
||||
github.com/pkg/errors v0.9.1 // indirect
|
||||
github.com/prometheus/client_model v0.6.1 // indirect
|
||||
github.com/prometheus/common v0.61.0 // indirect
|
||||
github.com/prometheus/procfs v0.15.1 // indirect
|
||||
github.com/samber/lo v1.47.0 // indirect
|
||||
@ -58,16 +65,10 @@ require (
|
||||
github.com/valyala/fasttemplate v1.2.2 // indirect
|
||||
go.opentelemetry.io/auto/sdk v1.1.0 // indirect
|
||||
go.opentelemetry.io/contrib/bridges/prometheus v0.58.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlplog/otlploggrpc v0.9.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlplog/otlploghttp v0.9.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlpmetric/otlpmetricgrpc v1.33.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/otlp/otlpmetric/otlpmetrichttp v1.33.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/prometheus v0.55.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/stdout/stdoutlog v0.9.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/stdout/stdoutmetric v1.33.0 // indirect
|
||||
go.opentelemetry.io/otel/exporters/stdout/stdouttrace v1.33.0 // indirect
|
||||
go.opentelemetry.io/otel/metric v1.33.0 // indirect
|
||||
go.opentelemetry.io/otel/sdk/metric v1.33.0 // indirect
|
||||
go.opentelemetry.io/proto/otlp v1.4.0 // indirect
|
||||
golang.org/x/crypto v0.31.0 // indirect
|
||||
golang.org/x/sys v0.28.0 // indirect
|
||||
|
12
go.sum
12
go.sum
@ -1,3 +1,5 @@
|
||||
filippo.io/edwards25519 v1.1.0 h1:FNf4tywRC1HmFuKW5xopWpigGjJKiJSV0Cqo0cJWDaA=
|
||||
filippo.io/edwards25519 v1.1.0/go.mod h1:BxyFTGdWcka3PhytdK4V28tE5sGfRvvvRV7EaN4VDT4=
|
||||
github.com/abh/certman v0.4.0 h1:XHoDtb0YyRQPclaHMrBDlKTVZpNjTK6vhB0S3Bd/Sbs=
|
||||
github.com/abh/certman v0.4.0/go.mod h1:x8QhpKVZifmV1Hdiwdg9gLo2GMPAxezz1s3zrVnPs+I=
|
||||
github.com/beorn7/perks v1.0.1 h1:VlbKKnNfV8bJzeqoa4cOKqO6bYr3WgKZxO8Z16+hsOM=
|
||||
@ -17,6 +19,8 @@ github.com/go-logr/logr v1.4.2 h1:6pFjapn8bFcIbiKo3XT4j/BhANplGihG6tvd+8rYgrY=
|
||||
github.com/go-logr/logr v1.4.2/go.mod h1:9T104GzyrTigFIr8wt5mBrctHMim0Nb2HLGrmQ40KvY=
|
||||
github.com/go-logr/stdr v1.2.2 h1:hSWxHoqTgW2S2qGc0LTAI563KZ5YKYRhT3MFKZMbjag=
|
||||
github.com/go-logr/stdr v1.2.2/go.mod h1:mMo/vtBO5dYbehREoey6XUKy/eSumjCCveDpRre4VKE=
|
||||
github.com/go-sql-driver/mysql v1.9.3 h1:U/N249h2WzJ3Ukj8SowVFjdtZKfu9vlLZxjPXV1aweo=
|
||||
github.com/go-sql-driver/mysql v1.9.3/go.mod h1:qn46aNg1333BRMNU69Lq93t8du/dwxI64Gl8i5p1WMU=
|
||||
github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek=
|
||||
github.com/golang/protobuf v1.5.4/go.mod h1:lnTiLA8Wa4RWRcIUkrtSVa5nRhsEGBg48fD6rSs7xps=
|
||||
github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI=
|
||||
@ -30,6 +34,10 @@ github.com/inconshreveable/mousetrap v1.1.0/go.mod h1:vpF70FUmC8bwa3OWnCshd2FqLf
|
||||
github.com/klauspost/compress v1.15.9/go.mod h1:PhcZ0MbTNciWF3rruxRgKxI5NkcHHrHUDtV4Yw2GlzU=
|
||||
github.com/klauspost/compress v1.17.11 h1:In6xLpyWOi1+C7tXUUWv2ot1QvBjxevKAaI6IXrJmUc=
|
||||
github.com/klauspost/compress v1.17.11/go.mod h1:pMDklpSncoRMuLFrf1W9Ss9KT+0rH90U12bZKk7uwG0=
|
||||
github.com/kr/pretty v0.3.1 h1:flRD4NNwYAUpkphVc1HcthR4KEIFJ65n8Mw5qdRn3LE=
|
||||
github.com/kr/pretty v0.3.1/go.mod h1:hoEshYVHaxMs3cyo3Yncou5ZscifuDolrwPKZanG3xk=
|
||||
github.com/kr/text v0.2.0 h1:5Nx0Ya0ZqY2ygV366QzturHI13Jq95ApcVaJBhpS+AY=
|
||||
github.com/kr/text v0.2.0/go.mod h1:eLer722TekiGuMkidMxC/pM04lWEeraHUUmBw8l2grE=
|
||||
github.com/kylelemons/godebug v1.1.0 h1:RPNrshWIDI6G2gRW9EHilWtl7Z6Sb1BR0xunSBf0SNc=
|
||||
github.com/kylelemons/godebug v1.1.0/go.mod h1:9/0rRGxNHcop5bhtWyNeEfOS8JIWk580+fNqagV/RAw=
|
||||
github.com/labstack/echo-contrib v0.17.2 h1:K1zivqmtcC70X9VdBFdLomjPDEVHlrcAObqmuFj1c6w=
|
||||
@ -65,6 +73,8 @@ github.com/prometheus/procfs v0.15.1 h1:YagwOFzUgYfKKHX6Dr+sHT7km/hxC76UB0leargg
|
||||
github.com/prometheus/procfs v0.15.1/go.mod h1:fB45yRUv8NstnjriLhBQLuOUt+WW4BsoGhij/e3PBqk=
|
||||
github.com/remychantenay/slog-otel v1.3.2 h1:ZBx8qnwfLJ6e18Vba4e9Xp9B7khTmpIwFsU1sAmActw=
|
||||
github.com/remychantenay/slog-otel v1.3.2/go.mod h1:gKW4tQ8cGOKoA+bi7wtYba/tcJ6Tc9XyQ/EW8gHA/2E=
|
||||
github.com/rogpeppe/go-internal v1.13.1 h1:KvO1DLK/DRN07sQ1LQKScxyZJuNnedQ5/wKSR38lUII=
|
||||
github.com/rogpeppe/go-internal v1.13.1/go.mod h1:uMEvuHeurkdAXX61udpOXGD/AzZDWNMNyH2VO9fmH0o=
|
||||
github.com/russross/blackfriday/v2 v2.1.0/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM=
|
||||
github.com/samber/lo v1.47.0 h1:z7RynLwP5nbyRscyvcD043DWYoOcYRv3mV8lBeqOCLc=
|
||||
github.com/samber/lo v1.47.0/go.mod h1:RmDH9Ct32Qy3gduHQuKJ3gW1fMHAnE/fAzQuf6He5cU=
|
||||
@ -211,6 +221,8 @@ google.golang.org/grpc v1.69.2/go.mod h1:vyjdE6jLBI76dgpDojsFGNaHlxdjXN9ghpnd2o7
|
||||
google.golang.org/protobuf v1.36.1 h1:yBPeRvTftaleIgM3PZ/WBIZ7XM/eEYAaEyCwvyjq/gk=
|
||||
google.golang.org/protobuf v1.36.1/go.mod h1:9fA7Ob0pmnwhb644+1+CVWFRbNajQ6iRojtC/QF5bRE=
|
||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk=
|
||||
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c/go.mod h1:JHkPIbrfpd72SG/EVd6muEfDQjcINNoR0C8j2r3qZ4Q=
|
||||
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
|
||||
gopkg.in/yaml.v3 v3.0.1/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||
|
@ -1,3 +1,9 @@
|
||||
// Package health provides a standalone HTTP server for health checks.
|
||||
//
|
||||
// This package implements a simple health check server that can be used
|
||||
// to expose health status endpoints for monitoring and load balancing.
|
||||
// It supports custom health check handlers and provides structured logging
|
||||
// with graceful shutdown capabilities.
|
||||
package health
|
||||
|
||||
import (
|
||||
@ -11,11 +17,19 @@ import (
|
||||
"golang.org/x/sync/errgroup"
|
||||
)
|
||||
|
||||
// Server is a standalone HTTP server dedicated to health checks.
|
||||
// It runs separately from the main application server to ensure health
|
||||
// checks remain available even if the main server is experiencing issues.
|
||||
//
|
||||
// The server includes built-in timeouts, graceful shutdown, and structured
|
||||
// logging for monitoring and debugging health check behavior.
|
||||
type Server struct {
|
||||
log *slog.Logger
|
||||
healthFn http.HandlerFunc
|
||||
}
|
||||
|
||||
// NewServer creates a new health check server with the specified health handler.
|
||||
// If healthFn is nil, a default handler that returns HTTP 200 "ok" is used.
|
||||
func NewServer(healthFn http.HandlerFunc) *Server {
|
||||
if healthFn == nil {
|
||||
healthFn = basicHealth
|
||||
@ -27,10 +41,13 @@ func NewServer(healthFn http.HandlerFunc) *Server {
|
||||
return srv
|
||||
}
|
||||
|
||||
// SetLogger replaces the default logger with a custom one.
|
||||
func (srv *Server) SetLogger(log *slog.Logger) {
|
||||
srv.log = log
|
||||
}
|
||||
|
||||
// Listen starts the health server on the specified port and blocks until ctx is cancelled.
|
||||
// The server exposes the health handler at "/__health" with graceful shutdown support.
|
||||
func (srv *Server) Listen(ctx context.Context, port int) error {
|
||||
srv.log.Info("starting health listener", "port", port)
|
||||
|
||||
@ -59,11 +76,10 @@ func (srv *Server) Listen(ctx context.Context, port int) error {
|
||||
|
||||
<-ctx.Done()
|
||||
|
||||
ctx, cancel := context.WithTimeout(ctx, 2*time.Second)
|
||||
defer cancel()
|
||||
|
||||
g.Go(func() error {
|
||||
if err := hsrv.Shutdown(ctx); err != nil {
|
||||
shCtx, cancel := context.WithTimeout(context.Background(), 2*time.Second)
|
||||
defer cancel()
|
||||
if err := hsrv.Shutdown(shCtx); err != nil {
|
||||
srv.log.Error("health check server shutdown failed", "err", err)
|
||||
return err
|
||||
}
|
||||
@ -73,8 +89,7 @@ func (srv *Server) Listen(ctx context.Context, port int) error {
|
||||
return g.Wait()
|
||||
}
|
||||
|
||||
// HealthCheckListener runs simple http server on the specified port for
|
||||
// health check probes
|
||||
// HealthCheckListener runs a simple HTTP server on the specified port for health check probes.
|
||||
func HealthCheckListener(ctx context.Context, port int, log *slog.Logger) error {
|
||||
srv := NewServer(nil)
|
||||
srv.SetLogger(log)
|
||||
|
@ -8,7 +8,6 @@ import (
|
||||
)
|
||||
|
||||
func TestHealthHandler(t *testing.T) {
|
||||
|
||||
req := httptest.NewRequest(http.MethodGet, "/__health", nil)
|
||||
w := httptest.NewRecorder()
|
||||
|
||||
|
378
internal/tracerconfig/config.go
Normal file
378
internal/tracerconfig/config.go
Normal file
@ -0,0 +1,378 @@
|
||||
// Package tracerconfig provides a bridge to eliminate circular dependencies between
|
||||
// the logger and tracing packages. It stores tracer configuration and provides
|
||||
// factory functions that can be used by the logger package without importing tracing.
|
||||
package tracerconfig
|
||||
|
||||
import (
|
||||
"context"
|
||||
"crypto/tls"
|
||||
"crypto/x509"
|
||||
"errors"
|
||||
"fmt"
|
||||
"net/url"
|
||||
"os"
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlplog/otlploggrpc"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlplog/otlploghttp"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlpmetric/otlpmetricgrpc"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlpmetric/otlpmetrichttp"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlptrace"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracegrpc"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracehttp"
|
||||
sdklog "go.opentelemetry.io/otel/sdk/log"
|
||||
sdkmetric "go.opentelemetry.io/otel/sdk/metric"
|
||||
sdktrace "go.opentelemetry.io/otel/sdk/trace"
|
||||
"google.golang.org/grpc/credentials"
|
||||
)
|
||||
|
||||
const (
|
||||
otelExporterOTLPProtoEnvKey = "OTEL_EXPORTER_OTLP_PROTOCOL"
|
||||
otelExporterOTLPTracesProtoEnvKey = "OTEL_EXPORTER_OTLP_TRACES_PROTOCOL"
|
||||
otelExporterOTLPLogsProtoEnvKey = "OTEL_EXPORTER_OTLP_LOGS_PROTOCOL"
|
||||
otelExporterOTLPMetricsProtoEnvKey = "OTEL_EXPORTER_OTLP_METRICS_PROTOCOL"
|
||||
)
|
||||
|
||||
var errInvalidOTLPProtocol = errors.New("invalid OTLP protocol - should be one of ['grpc', 'http/protobuf']")
|
||||
|
||||
// newInvalidProtocolError creates a specific error message for invalid protocols
|
||||
func newInvalidProtocolError(protocol, signalType string) error {
|
||||
return fmt.Errorf("invalid OTLP protocol '%s' for %s - should be one of ['grpc', 'http/protobuf', 'http/json']", protocol, signalType)
|
||||
}
|
||||
|
||||
// Validate checks the configuration for common errors and inconsistencies
|
||||
func (c *Config) Validate() error {
|
||||
var errs []error
|
||||
|
||||
// Check that both Endpoint and EndpointURL are not specified
|
||||
if c.Endpoint != "" && c.EndpointURL != "" {
|
||||
errs = append(errs, errors.New("cannot specify both Endpoint and EndpointURL - use one or the other"))
|
||||
}
|
||||
|
||||
// Validate EndpointURL format if specified
|
||||
if c.EndpointURL != "" {
|
||||
if _, err := url.Parse(c.EndpointURL); err != nil {
|
||||
errs = append(errs, fmt.Errorf("invalid EndpointURL format: %w", err))
|
||||
}
|
||||
}
|
||||
|
||||
// Validate Endpoint format if specified
|
||||
if c.Endpoint != "" {
|
||||
// Basic validation - should not contain protocol scheme
|
||||
if strings.Contains(c.Endpoint, "://") {
|
||||
errs = append(errs, errors.New("Endpoint should not include protocol scheme (use EndpointURL for full URLs)"))
|
||||
}
|
||||
// Should not be empty after trimming whitespace
|
||||
if strings.TrimSpace(c.Endpoint) == "" {
|
||||
errs = append(errs, errors.New("Endpoint cannot be empty or whitespace"))
|
||||
}
|
||||
}
|
||||
|
||||
// Validate TLS configuration consistency
|
||||
if c.CertificateProvider != nil && c.RootCAs == nil {
|
||||
// This is just a warning - client cert without custom CAs is valid
|
||||
// but might indicate a configuration issue
|
||||
}
|
||||
|
||||
// Validate service name if specified
|
||||
if c.ServiceName != "" && strings.TrimSpace(c.ServiceName) == "" {
|
||||
errs = append(errs, errors.New("ServiceName cannot be empty or whitespace"))
|
||||
}
|
||||
|
||||
// Combine all errors
|
||||
if len(errs) > 0 {
|
||||
var errMsgs []string
|
||||
for _, err := range errs {
|
||||
errMsgs = append(errMsgs, err.Error())
|
||||
}
|
||||
return fmt.Errorf("configuration validation failed: %s", strings.Join(errMsgs, "; "))
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// ValidateAndStore validates the configuration before storing it
|
||||
func ValidateAndStore(ctx context.Context, cfg *Config, logFactory LogExporterFactory, metricFactory MetricExporterFactory, traceFactory TraceExporterFactory) error {
|
||||
if cfg != nil {
|
||||
if err := cfg.Validate(); err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
Store(ctx, cfg, logFactory, metricFactory, traceFactory)
|
||||
return nil
|
||||
}
|
||||
|
||||
// GetClientCertificate defines a function type for providing client certificates for mutual TLS.
|
||||
// This is used when exporting telemetry data to secured OTLP endpoints that require
|
||||
// client certificate authentication.
|
||||
type GetClientCertificate func(*tls.CertificateRequestInfo) (*tls.Certificate, error)
|
||||
|
||||
// Config provides configuration options for OpenTelemetry tracing setup.
|
||||
// It supplements standard OpenTelemetry environment variables with additional
|
||||
// NTP Pool-specific configuration including TLS settings for secure OTLP export.
|
||||
type Config struct {
|
||||
ServiceName string // Service name for resource identification (overrides OTEL_SERVICE_NAME)
|
||||
Environment string // Deployment environment (development, staging, production)
|
||||
Endpoint string // OTLP endpoint hostname/port (e.g., "otlp.example.com:4317")
|
||||
EndpointURL string // Complete OTLP endpoint URL (e.g., "https://otlp.example.com:4317/v1/traces")
|
||||
CertificateProvider GetClientCertificate // Client certificate provider for mutual TLS
|
||||
RootCAs *x509.CertPool // CA certificate pool for server verification
|
||||
}
|
||||
|
||||
// LogExporterFactory creates an OTLP log exporter using the provided configuration.
|
||||
// This allows the logger package to create exporters without importing the tracing package.
|
||||
type LogExporterFactory func(context.Context, *Config) (sdklog.Exporter, error)
|
||||
|
||||
// MetricExporterFactory creates an OTLP metric exporter using the provided configuration.
|
||||
// This allows the metrics package to create exporters without importing the tracing package.
|
||||
type MetricExporterFactory func(context.Context, *Config) (sdkmetric.Exporter, error)
|
||||
|
||||
// TraceExporterFactory creates an OTLP trace exporter using the provided configuration.
|
||||
// This allows for consistent trace exporter creation across packages.
|
||||
type TraceExporterFactory func(context.Context, *Config) (sdktrace.SpanExporter, error)
|
||||
|
||||
// Global state for sharing configuration between packages
|
||||
var (
|
||||
globalConfig *Config
|
||||
globalContext context.Context
|
||||
logExporterFactory LogExporterFactory
|
||||
metricExporterFactory MetricExporterFactory
|
||||
traceExporterFactory TraceExporterFactory
|
||||
configMu sync.RWMutex
|
||||
)
|
||||
|
||||
// Store saves the tracer configuration and exporter factories for use by other packages.
|
||||
// This should be called by the tracing package during initialization.
|
||||
func Store(ctx context.Context, cfg *Config, logFactory LogExporterFactory, metricFactory MetricExporterFactory, traceFactory TraceExporterFactory) {
|
||||
configMu.Lock()
|
||||
defer configMu.Unlock()
|
||||
globalConfig = cfg
|
||||
globalContext = ctx
|
||||
logExporterFactory = logFactory
|
||||
metricExporterFactory = metricFactory
|
||||
traceExporterFactory = traceFactory
|
||||
}
|
||||
|
||||
// GetLogExporter returns the stored configuration and log exporter factory.
|
||||
// Returns nil values if no configuration has been stored yet.
|
||||
func GetLogExporter() (*Config, context.Context, LogExporterFactory) {
|
||||
configMu.RLock()
|
||||
defer configMu.RUnlock()
|
||||
return globalConfig, globalContext, logExporterFactory
|
||||
}
|
||||
|
||||
// GetMetricExporter returns the stored configuration and metric exporter factory.
|
||||
// Returns nil values if no configuration has been stored yet.
|
||||
func GetMetricExporter() (*Config, context.Context, MetricExporterFactory) {
|
||||
configMu.RLock()
|
||||
defer configMu.RUnlock()
|
||||
return globalConfig, globalContext, metricExporterFactory
|
||||
}
|
||||
|
||||
// GetTraceExporter returns the stored configuration and trace exporter factory.
|
||||
// Returns nil values if no configuration has been stored yet.
|
||||
func GetTraceExporter() (*Config, context.Context, TraceExporterFactory) {
|
||||
configMu.RLock()
|
||||
defer configMu.RUnlock()
|
||||
return globalConfig, globalContext, traceExporterFactory
|
||||
}
|
||||
|
||||
// Get returns the stored tracer configuration, context, and log exporter factory.
|
||||
// This maintains backward compatibility for the logger package.
|
||||
// Returns nil values if no configuration has been stored yet.
|
||||
func Get() (*Config, context.Context, LogExporterFactory) {
|
||||
return GetLogExporter()
|
||||
}
|
||||
|
||||
// IsConfigured returns true if tracer configuration has been stored.
|
||||
func IsConfigured() bool {
|
||||
configMu.RLock()
|
||||
defer configMu.RUnlock()
|
||||
return globalConfig != nil && globalContext != nil
|
||||
}
|
||||
|
||||
// Clear removes the stored configuration. This is primarily useful for testing.
|
||||
func Clear() {
|
||||
configMu.Lock()
|
||||
defer configMu.Unlock()
|
||||
globalConfig = nil
|
||||
globalContext = nil
|
||||
logExporterFactory = nil
|
||||
metricExporterFactory = nil
|
||||
traceExporterFactory = nil
|
||||
}
|
||||
|
||||
// getTLSConfig creates a TLS configuration from the provided Config.
|
||||
func getTLSConfig(cfg *Config) *tls.Config {
|
||||
if cfg.CertificateProvider == nil {
|
||||
return nil
|
||||
}
|
||||
return &tls.Config{
|
||||
GetClientCertificate: cfg.CertificateProvider,
|
||||
RootCAs: cfg.RootCAs,
|
||||
}
|
||||
}
|
||||
|
||||
// getProtocol determines the OTLP protocol to use for the given signal type.
|
||||
// It follows OpenTelemetry environment variable precedence.
|
||||
func getProtocol(signalSpecificEnv string) string {
|
||||
proto := os.Getenv(signalSpecificEnv)
|
||||
if proto == "" {
|
||||
proto = os.Getenv(otelExporterOTLPProtoEnvKey)
|
||||
}
|
||||
// Fallback to default, http/protobuf.
|
||||
if proto == "" {
|
||||
proto = "http/protobuf"
|
||||
}
|
||||
return proto
|
||||
}
|
||||
|
||||
// CreateOTLPLogExporter creates an OTLP log exporter using the provided configuration.
|
||||
func CreateOTLPLogExporter(ctx context.Context, cfg *Config) (sdklog.Exporter, error) {
|
||||
tlsConfig := getTLSConfig(cfg)
|
||||
proto := getProtocol(otelExporterOTLPLogsProtoEnvKey)
|
||||
|
||||
switch proto {
|
||||
case "grpc":
|
||||
opts := []otlploggrpc.Option{
|
||||
otlploggrpc.WithCompressor("gzip"),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlploggrpc.WithTLSCredentials(credentials.NewTLS(tlsConfig)))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlploggrpc.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlploggrpc.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
return otlploggrpc.New(ctx, opts...)
|
||||
case "http/protobuf", "http/json":
|
||||
opts := []otlploghttp.Option{
|
||||
otlploghttp.WithCompression(otlploghttp.GzipCompression),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlploghttp.WithTLSClientConfig(tlsConfig))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlploghttp.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlploghttp.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
opts = append(opts, otlploghttp.WithRetry(otlploghttp.RetryConfig{
|
||||
Enabled: true,
|
||||
InitialInterval: 3 * time.Second,
|
||||
MaxInterval: 60 * time.Second,
|
||||
MaxElapsedTime: 5 * time.Minute,
|
||||
}))
|
||||
|
||||
return otlploghttp.New(ctx, opts...)
|
||||
default:
|
||||
return nil, newInvalidProtocolError(proto, "logs")
|
||||
}
|
||||
}
|
||||
|
||||
// CreateOTLPMetricExporter creates an OTLP metric exporter using the provided configuration.
|
||||
func CreateOTLPMetricExporter(ctx context.Context, cfg *Config) (sdkmetric.Exporter, error) {
|
||||
tlsConfig := getTLSConfig(cfg)
|
||||
proto := getProtocol(otelExporterOTLPMetricsProtoEnvKey)
|
||||
|
||||
switch proto {
|
||||
case "grpc":
|
||||
opts := []otlpmetricgrpc.Option{
|
||||
otlpmetricgrpc.WithCompressor("gzip"),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlpmetricgrpc.WithTLSCredentials(credentials.NewTLS(tlsConfig)))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlpmetricgrpc.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlpmetricgrpc.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
return otlpmetricgrpc.New(ctx, opts...)
|
||||
case "http/protobuf", "http/json":
|
||||
opts := []otlpmetrichttp.Option{
|
||||
otlpmetrichttp.WithCompression(otlpmetrichttp.GzipCompression),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlpmetrichttp.WithTLSClientConfig(tlsConfig))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlpmetrichttp.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlpmetrichttp.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
opts = append(opts, otlpmetrichttp.WithRetry(otlpmetrichttp.RetryConfig{
|
||||
Enabled: true,
|
||||
InitialInterval: 3 * time.Second,
|
||||
MaxInterval: 60 * time.Second,
|
||||
MaxElapsedTime: 5 * time.Minute,
|
||||
}))
|
||||
|
||||
return otlpmetrichttp.New(ctx, opts...)
|
||||
default:
|
||||
return nil, newInvalidProtocolError(proto, "metrics")
|
||||
}
|
||||
}
|
||||
|
||||
// CreateOTLPTraceExporter creates an OTLP trace exporter using the provided configuration.
|
||||
func CreateOTLPTraceExporter(ctx context.Context, cfg *Config) (sdktrace.SpanExporter, error) {
|
||||
tlsConfig := getTLSConfig(cfg)
|
||||
proto := getProtocol(otelExporterOTLPTracesProtoEnvKey)
|
||||
|
||||
var client otlptrace.Client
|
||||
|
||||
switch proto {
|
||||
case "grpc":
|
||||
opts := []otlptracegrpc.Option{
|
||||
otlptracegrpc.WithCompressor("gzip"),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlptracegrpc.WithTLSCredentials(credentials.NewTLS(tlsConfig)))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlptracegrpc.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlptracegrpc.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
client = otlptracegrpc.NewClient(opts...)
|
||||
case "http/protobuf", "http/json":
|
||||
opts := []otlptracehttp.Option{
|
||||
otlptracehttp.WithCompression(otlptracehttp.GzipCompression),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlptracehttp.WithTLSClientConfig(tlsConfig))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlptracehttp.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlptracehttp.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
opts = append(opts, otlptracehttp.WithRetry(otlptracehttp.RetryConfig{
|
||||
Enabled: true,
|
||||
InitialInterval: 3 * time.Second,
|
||||
MaxInterval: 60 * time.Second,
|
||||
MaxElapsedTime: 5 * time.Minute,
|
||||
}))
|
||||
|
||||
client = otlptracehttp.NewClient(opts...)
|
||||
default:
|
||||
return nil, newInvalidProtocolError(proto, "traces")
|
||||
}
|
||||
|
||||
return otlptrace.New(ctx, client)
|
||||
}
|
474
internal/tracerconfig/config_test.go
Normal file
474
internal/tracerconfig/config_test.go
Normal file
@ -0,0 +1,474 @@
|
||||
package tracerconfig
|
||||
|
||||
import (
|
||||
"context"
|
||||
"crypto/tls"
|
||||
"crypto/x509"
|
||||
"os"
|
||||
"strings"
|
||||
"sync"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
sdklog "go.opentelemetry.io/otel/sdk/log"
|
||||
sdkmetric "go.opentelemetry.io/otel/sdk/metric"
|
||||
sdktrace "go.opentelemetry.io/otel/sdk/trace"
|
||||
)
|
||||
|
||||
func TestStore_And_Retrieve(t *testing.T) {
|
||||
// Clear any existing configuration
|
||||
Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
config := &Config{
|
||||
ServiceName: "test-service",
|
||||
Environment: "test",
|
||||
Endpoint: "localhost:4317",
|
||||
}
|
||||
|
||||
// Create mock factories
|
||||
logFactory := func(context.Context, *Config) (sdklog.Exporter, error) { return nil, nil }
|
||||
metricFactory := func(context.Context, *Config) (sdkmetric.Exporter, error) { return nil, nil }
|
||||
traceFactory := func(context.Context, *Config) (sdktrace.SpanExporter, error) { return nil, nil }
|
||||
|
||||
// Store configuration
|
||||
Store(ctx, config, logFactory, metricFactory, traceFactory)
|
||||
|
||||
// Test IsConfigured
|
||||
if !IsConfigured() {
|
||||
t.Error("IsConfigured() should return true after Store()")
|
||||
}
|
||||
|
||||
// Test GetLogExporter
|
||||
cfg, ctx2, factory := GetLogExporter()
|
||||
if cfg == nil || ctx2 == nil || factory == nil {
|
||||
t.Error("GetLogExporter() should return non-nil values")
|
||||
}
|
||||
if cfg.ServiceName != "test-service" {
|
||||
t.Errorf("Expected ServiceName 'test-service', got '%s'", cfg.ServiceName)
|
||||
}
|
||||
|
||||
// Test GetMetricExporter
|
||||
cfg, ctx3, metricFact := GetMetricExporter()
|
||||
if cfg == nil || ctx3 == nil || metricFact == nil {
|
||||
t.Error("GetMetricExporter() should return non-nil values")
|
||||
}
|
||||
|
||||
// Test GetTraceExporter
|
||||
cfg, ctx4, traceFact := GetTraceExporter()
|
||||
if cfg == nil || ctx4 == nil || traceFact == nil {
|
||||
t.Error("GetTraceExporter() should return non-nil values")
|
||||
}
|
||||
|
||||
// Test backward compatibility Get()
|
||||
cfg, ctx5, logFact := Get()
|
||||
if cfg == nil || ctx5 == nil || logFact == nil {
|
||||
t.Error("Get() should return non-nil values for backward compatibility")
|
||||
}
|
||||
}
|
||||
|
||||
func TestClear(t *testing.T) {
|
||||
// Store some configuration first
|
||||
ctx := context.Background()
|
||||
config := &Config{ServiceName: "test"}
|
||||
Store(ctx, config, nil, nil, nil)
|
||||
|
||||
if !IsConfigured() {
|
||||
t.Error("Should be configured before Clear()")
|
||||
}
|
||||
|
||||
// Clear configuration
|
||||
Clear()
|
||||
|
||||
if IsConfigured() {
|
||||
t.Error("Should not be configured after Clear()")
|
||||
}
|
||||
|
||||
// All getters should return nil
|
||||
cfg, ctx2, factory := GetLogExporter()
|
||||
if cfg != nil || ctx2 != nil || factory != nil {
|
||||
t.Error("GetLogExporter() should return nil values after Clear()")
|
||||
}
|
||||
}
|
||||
|
||||
func TestConcurrentAccess(t *testing.T) {
|
||||
Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
config := &Config{ServiceName: "concurrent-test"}
|
||||
|
||||
var wg sync.WaitGroup
|
||||
const numGoroutines = 10
|
||||
|
||||
// Test concurrent Store and Get operations
|
||||
wg.Add(numGoroutines * 2)
|
||||
|
||||
// Concurrent Store operations
|
||||
for i := 0; i < numGoroutines; i++ {
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
Store(ctx, config, nil, nil, nil)
|
||||
}()
|
||||
}
|
||||
|
||||
// Concurrent Get operations
|
||||
for i := 0; i < numGoroutines; i++ {
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
IsConfigured()
|
||||
GetLogExporter()
|
||||
GetMetricExporter()
|
||||
GetTraceExporter()
|
||||
}()
|
||||
}
|
||||
|
||||
wg.Wait()
|
||||
|
||||
// Should be configured after all operations
|
||||
if !IsConfigured() {
|
||||
t.Error("Should be configured after concurrent operations")
|
||||
}
|
||||
}
|
||||
|
||||
func TestGetTLSConfig(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
config *Config
|
||||
expected bool // whether TLS config should be nil
|
||||
}{
|
||||
{
|
||||
name: "nil certificate provider",
|
||||
config: &Config{},
|
||||
expected: true, // should be nil
|
||||
},
|
||||
{
|
||||
name: "with certificate provider",
|
||||
config: &Config{
|
||||
CertificateProvider: func(*tls.CertificateRequestInfo) (*tls.Certificate, error) {
|
||||
return &tls.Certificate{}, nil
|
||||
},
|
||||
},
|
||||
expected: false, // should not be nil
|
||||
},
|
||||
{
|
||||
name: "with certificate provider and RootCAs",
|
||||
config: &Config{
|
||||
CertificateProvider: func(*tls.CertificateRequestInfo) (*tls.Certificate, error) {
|
||||
return &tls.Certificate{}, nil
|
||||
},
|
||||
RootCAs: x509.NewCertPool(),
|
||||
},
|
||||
expected: false, // should not be nil
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
tlsConfig := getTLSConfig(tt.config)
|
||||
if tt.expected && tlsConfig != nil {
|
||||
t.Errorf("Expected nil TLS config, got %v", tlsConfig)
|
||||
}
|
||||
if !tt.expected && tlsConfig == nil {
|
||||
t.Error("Expected non-nil TLS config, got nil")
|
||||
}
|
||||
if !tt.expected && tlsConfig != nil {
|
||||
if tlsConfig.GetClientCertificate == nil {
|
||||
t.Error("Expected GetClientCertificate to be set")
|
||||
}
|
||||
if tt.config.RootCAs != nil && tlsConfig.RootCAs != tt.config.RootCAs {
|
||||
t.Error("Expected RootCAs to be set correctly")
|
||||
}
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestGetProtocol(t *testing.T) {
|
||||
// Save original env vars
|
||||
originalGeneral := os.Getenv(otelExporterOTLPProtoEnvKey)
|
||||
originalLogs := os.Getenv(otelExporterOTLPLogsProtoEnvKey)
|
||||
|
||||
defer func() {
|
||||
// Restore original env vars
|
||||
if originalGeneral != "" {
|
||||
os.Setenv(otelExporterOTLPProtoEnvKey, originalGeneral)
|
||||
} else {
|
||||
os.Unsetenv(otelExporterOTLPProtoEnvKey)
|
||||
}
|
||||
if originalLogs != "" {
|
||||
os.Setenv(otelExporterOTLPLogsProtoEnvKey, originalLogs)
|
||||
} else {
|
||||
os.Unsetenv(otelExporterOTLPLogsProtoEnvKey)
|
||||
}
|
||||
}()
|
||||
|
||||
tests := []struct {
|
||||
name string
|
||||
signalSpecific string
|
||||
generalProto string
|
||||
specificProto string
|
||||
expectedResult string
|
||||
}{
|
||||
{
|
||||
name: "no env vars set - default",
|
||||
signalSpecific: otelExporterOTLPLogsProtoEnvKey,
|
||||
expectedResult: "http/protobuf",
|
||||
},
|
||||
{
|
||||
name: "general env var set",
|
||||
signalSpecific: otelExporterOTLPLogsProtoEnvKey,
|
||||
generalProto: "grpc",
|
||||
expectedResult: "grpc",
|
||||
},
|
||||
{
|
||||
name: "specific env var overrides general",
|
||||
signalSpecific: otelExporterOTLPLogsProtoEnvKey,
|
||||
generalProto: "grpc",
|
||||
specificProto: "http/protobuf",
|
||||
expectedResult: "http/protobuf",
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
// Clear env vars
|
||||
os.Unsetenv(otelExporterOTLPProtoEnvKey)
|
||||
os.Unsetenv(otelExporterOTLPLogsProtoEnvKey)
|
||||
|
||||
// Set test env vars
|
||||
if tt.generalProto != "" {
|
||||
os.Setenv(otelExporterOTLPProtoEnvKey, tt.generalProto)
|
||||
}
|
||||
if tt.specificProto != "" {
|
||||
os.Setenv(tt.signalSpecific, tt.specificProto)
|
||||
}
|
||||
|
||||
result := getProtocol(tt.signalSpecific)
|
||||
if result != tt.expectedResult {
|
||||
t.Errorf("Expected protocol '%s', got '%s'", tt.expectedResult, result)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestCreateExporterErrors(t *testing.T) {
|
||||
ctx := context.Background()
|
||||
config := &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "invalid-endpoint",
|
||||
}
|
||||
|
||||
// Test with invalid protocol for logs
|
||||
os.Setenv(otelExporterOTLPLogsProtoEnvKey, "invalid-protocol")
|
||||
defer os.Unsetenv(otelExporterOTLPLogsProtoEnvKey)
|
||||
|
||||
_, err := CreateOTLPLogExporter(ctx, config)
|
||||
if err == nil {
|
||||
t.Error("Expected error for invalid protocol")
|
||||
}
|
||||
// Check that it's a protocol error (the specific message will be different now)
|
||||
if !strings.Contains(err.Error(), "invalid OTLP protocol") {
|
||||
t.Errorf("Expected protocol error, got %v", err)
|
||||
}
|
||||
|
||||
// Test with invalid protocol for metrics
|
||||
os.Setenv(otelExporterOTLPMetricsProtoEnvKey, "invalid-protocol")
|
||||
defer os.Unsetenv(otelExporterOTLPMetricsProtoEnvKey)
|
||||
|
||||
_, err = CreateOTLPMetricExporter(ctx, config)
|
||||
if err == nil {
|
||||
t.Error("Expected error for invalid protocol")
|
||||
}
|
||||
if !strings.Contains(err.Error(), "invalid OTLP protocol") {
|
||||
t.Errorf("Expected protocol error, got %v", err)
|
||||
}
|
||||
|
||||
// Test with invalid protocol for traces
|
||||
os.Setenv(otelExporterOTLPTracesProtoEnvKey, "invalid-protocol")
|
||||
defer os.Unsetenv(otelExporterOTLPTracesProtoEnvKey)
|
||||
|
||||
_, err = CreateOTLPTraceExporter(ctx, config)
|
||||
if err == nil {
|
||||
t.Error("Expected error for invalid protocol")
|
||||
}
|
||||
if !strings.Contains(err.Error(), "invalid OTLP protocol") {
|
||||
t.Errorf("Expected protocol error, got %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestCreateExporterValidProtocols(t *testing.T) {
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second)
|
||||
defer cancel()
|
||||
|
||||
config := &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "localhost:4317", // This will likely fail to connect, but should create exporter
|
||||
}
|
||||
|
||||
protocols := []string{"grpc", "http/protobuf", "http/json"}
|
||||
|
||||
for _, proto := range protocols {
|
||||
t.Run("logs_"+proto, func(t *testing.T) {
|
||||
os.Setenv(otelExporterOTLPLogsProtoEnvKey, proto)
|
||||
defer os.Unsetenv(otelExporterOTLPLogsProtoEnvKey)
|
||||
|
||||
exporter, err := CreateOTLPLogExporter(ctx, config)
|
||||
if err != nil {
|
||||
// Connection errors are expected since we're not running a real OTLP server
|
||||
// but the exporter should be created successfully
|
||||
t.Logf("Connection error expected: %v", err)
|
||||
}
|
||||
if exporter != nil {
|
||||
exporter.Shutdown(ctx)
|
||||
}
|
||||
})
|
||||
|
||||
t.Run("metrics_"+proto, func(t *testing.T) {
|
||||
os.Setenv(otelExporterOTLPMetricsProtoEnvKey, proto)
|
||||
defer os.Unsetenv(otelExporterOTLPMetricsProtoEnvKey)
|
||||
|
||||
exporter, err := CreateOTLPMetricExporter(ctx, config)
|
||||
if err != nil {
|
||||
t.Logf("Connection error expected: %v", err)
|
||||
}
|
||||
if exporter != nil {
|
||||
exporter.Shutdown(ctx)
|
||||
}
|
||||
})
|
||||
|
||||
t.Run("traces_"+proto, func(t *testing.T) {
|
||||
os.Setenv(otelExporterOTLPTracesProtoEnvKey, proto)
|
||||
defer os.Unsetenv(otelExporterOTLPTracesProtoEnvKey)
|
||||
|
||||
exporter, err := CreateOTLPTraceExporter(ctx, config)
|
||||
if err != nil {
|
||||
t.Logf("Connection error expected: %v", err)
|
||||
}
|
||||
if exporter != nil {
|
||||
exporter.Shutdown(ctx)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestConfigValidation(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
config *Config
|
||||
shouldErr bool
|
||||
}{
|
||||
{
|
||||
name: "valid empty config",
|
||||
config: &Config{},
|
||||
shouldErr: false,
|
||||
},
|
||||
{
|
||||
name: "valid config with endpoint",
|
||||
config: &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "localhost:4317",
|
||||
},
|
||||
shouldErr: false,
|
||||
},
|
||||
{
|
||||
name: "valid config with endpoint URL",
|
||||
config: &Config{
|
||||
ServiceName: "test-service",
|
||||
EndpointURL: "https://otlp.example.com:4317/v1/traces",
|
||||
},
|
||||
shouldErr: false,
|
||||
},
|
||||
{
|
||||
name: "invalid - both endpoint and endpoint URL",
|
||||
config: &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "localhost:4317",
|
||||
EndpointURL: "https://otlp.example.com:4317/v1/traces",
|
||||
},
|
||||
shouldErr: true,
|
||||
},
|
||||
{
|
||||
name: "invalid - endpoint with protocol",
|
||||
config: &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "https://localhost:4317",
|
||||
},
|
||||
shouldErr: true,
|
||||
},
|
||||
{
|
||||
name: "invalid - empty endpoint",
|
||||
config: &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: " ",
|
||||
},
|
||||
shouldErr: true,
|
||||
},
|
||||
{
|
||||
name: "invalid - malformed endpoint URL",
|
||||
config: &Config{
|
||||
ServiceName: "test-service",
|
||||
EndpointURL: "://invalid-url-missing-scheme",
|
||||
},
|
||||
shouldErr: true,
|
||||
},
|
||||
{
|
||||
name: "invalid - empty service name",
|
||||
config: &Config{
|
||||
ServiceName: " ",
|
||||
Endpoint: "localhost:4317",
|
||||
},
|
||||
shouldErr: true,
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
err := tt.config.Validate()
|
||||
if tt.shouldErr && err == nil {
|
||||
t.Error("Expected validation error, got nil")
|
||||
}
|
||||
if !tt.shouldErr && err != nil {
|
||||
t.Errorf("Expected no validation error, got: %v", err)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestValidateAndStore(t *testing.T) {
|
||||
Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
|
||||
// Test with valid config
|
||||
validConfig := &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "localhost:4317",
|
||||
}
|
||||
|
||||
err := ValidateAndStore(ctx, validConfig, nil, nil, nil)
|
||||
if err != nil {
|
||||
t.Errorf("ValidateAndStore with valid config should not error: %v", err)
|
||||
}
|
||||
|
||||
if !IsConfigured() {
|
||||
t.Error("Should be configured after ValidateAndStore")
|
||||
}
|
||||
|
||||
Clear()
|
||||
|
||||
// Test with invalid config
|
||||
invalidConfig := &Config{
|
||||
ServiceName: "test-service",
|
||||
Endpoint: "localhost:4317",
|
||||
EndpointURL: "https://example.com:4317", // both specified - invalid
|
||||
}
|
||||
|
||||
err = ValidateAndStore(ctx, invalidConfig, nil, nil, nil)
|
||||
if err == nil {
|
||||
t.Error("ValidateAndStore with invalid config should return error")
|
||||
}
|
||||
|
||||
if IsConfigured() {
|
||||
t.Error("Should not be configured after failed ValidateAndStore")
|
||||
}
|
||||
}
|
@ -1,3 +1,32 @@
|
||||
// Package kafconn provides a Kafka client wrapper with TLS support for secure log streaming.
|
||||
//
|
||||
// This package handles Kafka connections with mutual TLS authentication for the NTP Pool
|
||||
// project's log streaming infrastructure. It provides factories for creating Kafka readers
|
||||
// and writers with automatic broker discovery, TLS configuration, and connection management.
|
||||
//
|
||||
// The package is designed specifically for the NTP Pool pipeline infrastructure and includes
|
||||
// hardcoded bootstrap servers and group configurations. It uses certman for automatic
|
||||
// certificate renewal and provides compression and batching optimizations.
|
||||
//
|
||||
// Key features:
|
||||
// - Mutual TLS authentication with automatic certificate renewal
|
||||
// - Broker discovery and connection pooling
|
||||
// - Reader and writer factory methods with optimized configurations
|
||||
// - LZ4 compression for efficient data transfer
|
||||
// - Configurable batch sizes and load balancing
|
||||
//
|
||||
// Example usage:
|
||||
//
|
||||
// tlsSetup := kafconn.TLSSetup{
|
||||
// CA: "/path/to/ca.pem",
|
||||
// Cert: "/path/to/client.pem",
|
||||
// Key: "/path/to/client.key",
|
||||
// }
|
||||
// kafka, err := kafconn.NewKafka(ctx, tlsSetup)
|
||||
// if err != nil {
|
||||
// log.Fatal(err)
|
||||
// }
|
||||
// writer, err := kafka.NewWriter("logs")
|
||||
package kafconn
|
||||
|
||||
import (
|
||||
@ -24,12 +53,17 @@ const (
|
||||
// kafkaMinBatchSize = 1000
|
||||
)
|
||||
|
||||
// TLSSetup contains file paths for TLS certificate configuration.
|
||||
// All fields are required for establishing secure Kafka connections.
|
||||
type TLSSetup struct {
|
||||
CA string
|
||||
Key string
|
||||
Cert string
|
||||
CA string // Path to CA certificate file for server verification
|
||||
Key string // Path to client private key file
|
||||
Cert string // Path to client certificate file
|
||||
}
|
||||
|
||||
// Kafka represents a configured Kafka client with TLS support.
|
||||
// It manages connections, brokers, and provides factory methods for readers and writers.
|
||||
// The client handles broker discovery, connection pooling, and TLS configuration automatically.
|
||||
type Kafka struct {
|
||||
tls TLSSetup
|
||||
|
||||
@ -42,11 +76,9 @@ type Kafka struct {
|
||||
l *log.Logger
|
||||
|
||||
// wr *kafka.Writer
|
||||
|
||||
}
|
||||
|
||||
func (k *Kafka) tlsConfig() (*tls.Config, error) {
|
||||
|
||||
cm, err := certman.New(k.tls.Cert, k.tls.Key)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
@ -118,6 +150,19 @@ func (k *Kafka) kafkaTransport(ctx context.Context) (*kafka.Transport, error) {
|
||||
return transport, nil
|
||||
}
|
||||
|
||||
// NewKafka creates a new Kafka client with TLS configuration and establishes initial connections.
|
||||
// It performs broker discovery, validates TLS certificates, and prepares the client for creating
|
||||
// readers and writers.
|
||||
//
|
||||
// The function validates TLS configuration, establishes a connection to the bootstrap server,
|
||||
// discovers all available brokers, and configures transport layers for optimal performance.
|
||||
//
|
||||
// Parameters:
|
||||
// - ctx: Context for connection establishment and timeouts
|
||||
// - tls: TLS configuration with paths to CA, certificate, and key files
|
||||
//
|
||||
// Returns a configured Kafka client ready for creating readers and writers, or an error
|
||||
// if TLS setup fails, connection cannot be established, or broker discovery fails.
|
||||
func NewKafka(ctx context.Context, tls TLSSetup) (*Kafka, error) {
|
||||
l := log.New(os.Stdout, "kafka: ", log.Ldate|log.Ltime|log.LUTC|log.Lmsgprefix|log.Lmicroseconds)
|
||||
|
||||
@ -173,6 +218,12 @@ func NewKafka(ctx context.Context, tls TLSSetup) (*Kafka, error) {
|
||||
return k, nil
|
||||
}
|
||||
|
||||
// NewReader creates a new Kafka reader with the client's broker list and TLS configuration.
|
||||
// The provided config is enhanced with the discovered brokers and configured dialer.
|
||||
// The reader supports automatic offset management, consumer group coordination, and reconnection.
|
||||
//
|
||||
// The caller should configure the reader's Topic, GroupID, and other consumer-specific settings
|
||||
// in the provided config. The client automatically sets Brokers and Dialer fields.
|
||||
func (k *Kafka) NewReader(config kafka.ReaderConfig) *kafka.Reader {
|
||||
config.Brokers = k.brokerAddrs()
|
||||
config.Dialer = k.dialer
|
||||
@ -188,8 +239,17 @@ func (k *Kafka) brokerAddrs() []string {
|
||||
return addrs
|
||||
}
|
||||
|
||||
// NewWriter creates a new Kafka writer for the specified topic with optimized configuration.
|
||||
// The writer uses LZ4 compression, least-bytes load balancing, and batching for performance.
|
||||
//
|
||||
// Configuration includes:
|
||||
// - Batch size: 2000 messages for efficient throughput
|
||||
// - Compression: LZ4 for fast compression with good ratios
|
||||
// - Balancer: LeastBytes for optimal partition distribution
|
||||
// - Transport: TLS-configured transport with connection pooling
|
||||
//
|
||||
// The writer is ready for immediate use and handles connection management automatically.
|
||||
func (k *Kafka) NewWriter(topic string) (*kafka.Writer, error) {
|
||||
|
||||
// https://pkg.go.dev/github.com/segmentio/kafka-go#Writer
|
||||
w := &kafka.Writer{
|
||||
Addr: kafka.TCP(k.brokerAddrs()...),
|
||||
@ -205,6 +265,12 @@ func (k *Kafka) NewWriter(topic string) (*kafka.Writer, error) {
|
||||
return w, nil
|
||||
}
|
||||
|
||||
// CheckPartitions verifies that the Kafka connection can read partition metadata.
|
||||
// This method is useful for health checks and connection validation.
|
||||
//
|
||||
// Returns an error if partition metadata cannot be retrieved, which typically
|
||||
// indicates connection problems, authentication failures, or broker unavailability.
|
||||
// Logs a warning if no partitions are available but does not return an error.
|
||||
func (k *Kafka) CheckPartitions() error {
|
||||
partitions, err := k.conn.ReadPartitions()
|
||||
if err != nil {
|
||||
|
204
logger/buffering_exporter.go
Normal file
204
logger/buffering_exporter.go
Normal file
@ -0,0 +1,204 @@
|
||||
package logger
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"fmt"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"go.ntppool.org/common/internal/tracerconfig"
|
||||
otellog "go.opentelemetry.io/otel/sdk/log"
|
||||
)
|
||||
|
||||
// bufferingExporter wraps an OTLP exporter and buffers logs until tracing is configured
|
||||
type bufferingExporter struct {
|
||||
mu sync.RWMutex
|
||||
|
||||
// Buffered records while waiting for tracing config
|
||||
buffer [][]otellog.Record
|
||||
bufferSize int
|
||||
maxBuffSize int
|
||||
|
||||
// Real exporter (created when tracing is configured)
|
||||
exporter otellog.Exporter
|
||||
|
||||
// Thread-safe initialization state (managed only by checkReadiness)
|
||||
initErr error
|
||||
|
||||
// Background checker
|
||||
stopChecker chan struct{}
|
||||
checkerDone chan struct{}
|
||||
}
|
||||
|
||||
// newBufferingExporter creates a new exporter that buffers logs until tracing is configured
|
||||
func newBufferingExporter() *bufferingExporter {
|
||||
e := &bufferingExporter{
|
||||
maxBuffSize: 1000, // Max number of batches to buffer
|
||||
stopChecker: make(chan struct{}),
|
||||
checkerDone: make(chan struct{}),
|
||||
}
|
||||
|
||||
// Start background readiness checker
|
||||
go e.checkReadiness()
|
||||
|
||||
return e
|
||||
}
|
||||
|
||||
// Export implements otellog.Exporter
|
||||
func (e *bufferingExporter) Export(ctx context.Context, records []otellog.Record) error {
|
||||
// Check if exporter is ready (initialization handled by checkReadiness goroutine)
|
||||
e.mu.RLock()
|
||||
exporter := e.exporter
|
||||
e.mu.RUnlock()
|
||||
|
||||
if exporter != nil {
|
||||
return exporter.Export(ctx, records)
|
||||
}
|
||||
|
||||
// Not ready yet, buffer the records
|
||||
return e.bufferRecords(records)
|
||||
}
|
||||
|
||||
// initialize attempts to create the real OTLP exporter using tracing config
|
||||
func (e *bufferingExporter) initialize() error {
|
||||
cfg, ctx, factory := tracerconfig.Get()
|
||||
if cfg == nil || ctx == nil || factory == nil {
|
||||
return errors.New("tracer not configured yet")
|
||||
}
|
||||
|
||||
// Add timeout for initialization
|
||||
initCtx, cancel := context.WithTimeout(ctx, 10*time.Second)
|
||||
defer cancel()
|
||||
|
||||
exporter, err := factory(initCtx, cfg)
|
||||
if err != nil {
|
||||
return fmt.Errorf("failed to create OTLP exporter: %w", err)
|
||||
}
|
||||
|
||||
e.mu.Lock()
|
||||
e.exporter = exporter
|
||||
flushErr := e.flushBuffer(initCtx)
|
||||
e.mu.Unlock()
|
||||
|
||||
if flushErr != nil {
|
||||
// Log but don't fail initialization
|
||||
Setup().Warn("buffer flush failed during initialization", "error", flushErr)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// bufferRecords adds records to the buffer for later processing
|
||||
func (e *bufferingExporter) bufferRecords(records []otellog.Record) error {
|
||||
e.mu.Lock()
|
||||
defer e.mu.Unlock()
|
||||
|
||||
// Buffer the batch if we have space
|
||||
if e.bufferSize < e.maxBuffSize {
|
||||
// Clone records to avoid retention issues
|
||||
cloned := make([]otellog.Record, len(records))
|
||||
for i, r := range records {
|
||||
cloned[i] = r.Clone()
|
||||
}
|
||||
e.buffer = append(e.buffer, cloned)
|
||||
e.bufferSize++
|
||||
}
|
||||
|
||||
// Always return success to BatchProcessor
|
||||
return nil
|
||||
}
|
||||
|
||||
// checkReadiness periodically attempts initialization until successful
|
||||
func (e *bufferingExporter) checkReadiness() {
|
||||
defer close(e.checkerDone)
|
||||
|
||||
ticker := time.NewTicker(1 * time.Second)
|
||||
defer ticker.Stop()
|
||||
|
||||
for {
|
||||
select {
|
||||
case <-ticker.C:
|
||||
// Check if we already have a working exporter
|
||||
e.mu.RLock()
|
||||
hasExporter := e.exporter != nil
|
||||
e.mu.RUnlock()
|
||||
|
||||
if hasExporter {
|
||||
return // Exporter ready, checker no longer needed
|
||||
}
|
||||
|
||||
// Try to initialize
|
||||
err := e.initialize()
|
||||
e.mu.Lock()
|
||||
e.initErr = err
|
||||
e.mu.Unlock()
|
||||
|
||||
case <-e.stopChecker:
|
||||
return
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// flushBuffer sends all buffered batches through the real exporter
|
||||
func (e *bufferingExporter) flushBuffer(ctx context.Context) error {
|
||||
if len(e.buffer) == 0 {
|
||||
return nil
|
||||
}
|
||||
|
||||
flushCtx, cancel := context.WithTimeout(ctx, 30*time.Second)
|
||||
defer cancel()
|
||||
|
||||
var lastErr error
|
||||
for _, batch := range e.buffer {
|
||||
if err := e.exporter.Export(flushCtx, batch); err != nil {
|
||||
lastErr = err
|
||||
}
|
||||
}
|
||||
|
||||
// Clear buffer after flush attempt
|
||||
e.buffer = nil
|
||||
e.bufferSize = 0
|
||||
|
||||
return lastErr
|
||||
}
|
||||
|
||||
// ForceFlush implements otellog.Exporter
|
||||
func (e *bufferingExporter) ForceFlush(ctx context.Context) error {
|
||||
e.mu.RLock()
|
||||
defer e.mu.RUnlock()
|
||||
|
||||
if e.exporter != nil {
|
||||
return e.exporter.ForceFlush(ctx)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Shutdown implements otellog.Exporter
|
||||
func (e *bufferingExporter) Shutdown(ctx context.Context) error {
|
||||
// Stop the readiness checker from continuing
|
||||
close(e.stopChecker)
|
||||
|
||||
// Wait for readiness checker goroutine to complete
|
||||
<-e.checkerDone
|
||||
|
||||
// Give one final chance for TLS/tracing to become ready for buffer flushing
|
||||
e.mu.RLock()
|
||||
hasExporter := e.exporter != nil
|
||||
e.mu.RUnlock()
|
||||
|
||||
if !hasExporter {
|
||||
err := e.initialize()
|
||||
e.mu.Lock()
|
||||
e.initErr = err
|
||||
e.mu.Unlock()
|
||||
}
|
||||
|
||||
e.mu.Lock()
|
||||
defer e.mu.Unlock()
|
||||
|
||||
if e.exporter != nil {
|
||||
return e.exporter.Shutdown(ctx)
|
||||
}
|
||||
return nil
|
||||
}
|
235
logger/level_test.go
Normal file
235
logger/level_test.go
Normal file
@ -0,0 +1,235 @@
|
||||
package logger
|
||||
|
||||
import (
|
||||
"context"
|
||||
"log/slog"
|
||||
"os"
|
||||
"testing"
|
||||
"time"
|
||||
)
|
||||
|
||||
func TestParseLevel(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
input string
|
||||
expected slog.Level
|
||||
expectError bool
|
||||
}{
|
||||
{"empty string", "", slog.LevelInfo, false},
|
||||
{"DEBUG upper", "DEBUG", slog.LevelDebug, false},
|
||||
{"debug lower", "debug", slog.LevelDebug, false},
|
||||
{"INFO upper", "INFO", slog.LevelInfo, false},
|
||||
{"info lower", "info", slog.LevelInfo, false},
|
||||
{"WARN upper", "WARN", slog.LevelWarn, false},
|
||||
{"warn lower", "warn", slog.LevelWarn, false},
|
||||
{"ERROR upper", "ERROR", slog.LevelError, false},
|
||||
{"error lower", "error", slog.LevelError, false},
|
||||
{"invalid level", "invalid", slog.LevelInfo, true},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
level, err := ParseLevel(tt.input)
|
||||
if tt.expectError {
|
||||
if err == nil {
|
||||
t.Errorf("expected error for input %q, got nil", tt.input)
|
||||
}
|
||||
} else {
|
||||
if err != nil {
|
||||
t.Errorf("unexpected error for input %q: %v", tt.input, err)
|
||||
}
|
||||
if level != tt.expected {
|
||||
t.Errorf("expected level %v for input %q, got %v", tt.expected, tt.input, level)
|
||||
}
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestSetLevel(t *testing.T) {
|
||||
// Store original level to restore later
|
||||
originalLevel := Level.Level()
|
||||
defer Level.Set(originalLevel)
|
||||
|
||||
SetLevel(slog.LevelDebug)
|
||||
if Level.Level() != slog.LevelDebug {
|
||||
t.Errorf("expected Level to be Debug, got %v", Level.Level())
|
||||
}
|
||||
|
||||
SetLevel(slog.LevelError)
|
||||
if Level.Level() != slog.LevelError {
|
||||
t.Errorf("expected Level to be Error, got %v", Level.Level())
|
||||
}
|
||||
}
|
||||
|
||||
func TestSetOTLPLevel(t *testing.T) {
|
||||
// Store original level to restore later
|
||||
originalLevel := OTLPLevel.Level()
|
||||
defer OTLPLevel.Set(originalLevel)
|
||||
|
||||
SetOTLPLevel(slog.LevelWarn)
|
||||
if OTLPLevel.Level() != slog.LevelWarn {
|
||||
t.Errorf("expected OTLPLevel to be Warn, got %v", OTLPLevel.Level())
|
||||
}
|
||||
|
||||
SetOTLPLevel(slog.LevelDebug)
|
||||
if OTLPLevel.Level() != slog.LevelDebug {
|
||||
t.Errorf("expected OTLPLevel to be Debug, got %v", OTLPLevel.Level())
|
||||
}
|
||||
}
|
||||
|
||||
func TestOTLPLevelHandler(t *testing.T) {
|
||||
// Create a mock handler that counts calls
|
||||
callCount := 0
|
||||
mockHandler := &mockHandler{
|
||||
handleFunc: func(ctx context.Context, r slog.Record) error {
|
||||
callCount++
|
||||
return nil
|
||||
},
|
||||
}
|
||||
|
||||
// Set OTLP level to Warn
|
||||
originalLevel := OTLPLevel.Level()
|
||||
defer OTLPLevel.Set(originalLevel)
|
||||
OTLPLevel.Set(slog.LevelWarn)
|
||||
|
||||
// Create OTLP level handler
|
||||
handler := newOTLPLevelHandler(mockHandler)
|
||||
|
||||
ctx := context.Background()
|
||||
|
||||
// Test that Debug and Info are filtered out
|
||||
if handler.Enabled(ctx, slog.LevelDebug) {
|
||||
t.Error("Debug level should be disabled when OTLP level is Warn")
|
||||
}
|
||||
if handler.Enabled(ctx, slog.LevelInfo) {
|
||||
t.Error("Info level should be disabled when OTLP level is Warn")
|
||||
}
|
||||
|
||||
// Test that Warn and Error are enabled
|
||||
if !handler.Enabled(ctx, slog.LevelWarn) {
|
||||
t.Error("Warn level should be enabled when OTLP level is Warn")
|
||||
}
|
||||
if !handler.Enabled(ctx, slog.LevelError) {
|
||||
t.Error("Error level should be enabled when OTLP level is Warn")
|
||||
}
|
||||
|
||||
// Test that Handle respects level filtering
|
||||
now := time.Now()
|
||||
debugRecord := slog.NewRecord(now, slog.LevelDebug, "debug message", 0)
|
||||
warnRecord := slog.NewRecord(now, slog.LevelWarn, "warn message", 0)
|
||||
|
||||
handler.Handle(ctx, debugRecord)
|
||||
if callCount != 0 {
|
||||
t.Error("Debug record should not be passed to underlying handler")
|
||||
}
|
||||
|
||||
handler.Handle(ctx, warnRecord)
|
||||
if callCount != 1 {
|
||||
t.Error("Warn record should be passed to underlying handler")
|
||||
}
|
||||
}
|
||||
|
||||
func TestEnvironmentVariables(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
envVar string
|
||||
envValue string
|
||||
configPrefix string
|
||||
testFunc func(t *testing.T)
|
||||
}{
|
||||
{
|
||||
name: "LOG_LEVEL sets stderr level",
|
||||
envVar: "LOG_LEVEL",
|
||||
envValue: "ERROR",
|
||||
testFunc: func(t *testing.T) {
|
||||
// Reset the setup state
|
||||
resetLoggerSetup()
|
||||
|
||||
// Call setupStdErrHandler which should read the env var
|
||||
handler := setupStdErrHandler()
|
||||
if handler == nil {
|
||||
t.Fatal("setupStdErrHandler returned nil")
|
||||
}
|
||||
|
||||
if Level.Level() != slog.LevelError {
|
||||
t.Errorf("expected Level to be Error after setting LOG_LEVEL=ERROR, got %v", Level.Level())
|
||||
}
|
||||
},
|
||||
},
|
||||
{
|
||||
name: "Prefixed LOG_LEVEL",
|
||||
envVar: "TEST_LOG_LEVEL",
|
||||
envValue: "DEBUG",
|
||||
configPrefix: "TEST",
|
||||
testFunc: func(t *testing.T) {
|
||||
ConfigPrefix = "TEST"
|
||||
defer func() { ConfigPrefix = "" }()
|
||||
|
||||
resetLoggerSetup()
|
||||
handler := setupStdErrHandler()
|
||||
if handler == nil {
|
||||
t.Fatal("setupStdErrHandler returned nil")
|
||||
}
|
||||
|
||||
if Level.Level() != slog.LevelDebug {
|
||||
t.Errorf("expected Level to be Debug after setting TEST_LOG_LEVEL=DEBUG, got %v", Level.Level())
|
||||
}
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
// Store original env value and level
|
||||
originalEnv := os.Getenv(tt.envVar)
|
||||
originalLevel := Level.Level()
|
||||
defer func() {
|
||||
os.Setenv(tt.envVar, originalEnv)
|
||||
Level.Set(originalLevel)
|
||||
}()
|
||||
|
||||
// Set test environment variable
|
||||
os.Setenv(tt.envVar, tt.envValue)
|
||||
|
||||
// Run the test
|
||||
tt.testFunc(t)
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
// mockHandler is a simple mock implementation of slog.Handler for testing
|
||||
type mockHandler struct {
|
||||
handleFunc func(ctx context.Context, r slog.Record) error
|
||||
}
|
||||
|
||||
func (m *mockHandler) Enabled(ctx context.Context, level slog.Level) bool {
|
||||
return true
|
||||
}
|
||||
|
||||
func (m *mockHandler) Handle(ctx context.Context, r slog.Record) error {
|
||||
if m.handleFunc != nil {
|
||||
return m.handleFunc(ctx, r)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *mockHandler) WithAttrs(attrs []slog.Attr) slog.Handler {
|
||||
return m
|
||||
}
|
||||
|
||||
func (m *mockHandler) WithGroup(name string) slog.Handler {
|
||||
return m
|
||||
}
|
||||
|
||||
// resetLoggerSetup resets the sync.Once instances for testing
|
||||
func resetLoggerSetup() {
|
||||
// Reset package-level variables
|
||||
textLogger = nil
|
||||
otlpLogger = nil
|
||||
multiLogger = nil
|
||||
|
||||
// Note: We can't easily reset sync.Once instances in tests,
|
||||
// but for the specific test we're doing (environment variable parsing)
|
||||
// we can test the setupStdErrHandler function directly
|
||||
}
|
@ -16,14 +16,9 @@ type logfmt struct {
|
||||
mu sync.Mutex
|
||||
}
|
||||
|
||||
func newLogFmtHandler(next slog.Handler) slog.Handler {
|
||||
|
||||
buf := bytes.NewBuffer([]byte{})
|
||||
|
||||
h := &logfmt{
|
||||
buf: buf,
|
||||
next: next,
|
||||
txt: slog.NewTextHandler(buf, &slog.HandlerOptions{
|
||||
// createTextHandlerOptions creates the common slog.HandlerOptions used by all logfmt handlers
|
||||
func createTextHandlerOptions() *slog.HandlerOptions {
|
||||
return &slog.HandlerOptions{
|
||||
ReplaceAttr: func(groups []string, a slog.Attr) slog.Attr {
|
||||
if a.Key == slog.TimeKey && len(groups) == 0 {
|
||||
return slog.Attr{}
|
||||
@ -33,7 +28,16 @@ func newLogFmtHandler(next slog.Handler) slog.Handler {
|
||||
}
|
||||
return a
|
||||
},
|
||||
}),
|
||||
}
|
||||
}
|
||||
|
||||
func newLogFmtHandler(next slog.Handler) slog.Handler {
|
||||
buf := bytes.NewBuffer([]byte{})
|
||||
|
||||
h := &logfmt{
|
||||
buf: buf,
|
||||
next: next,
|
||||
txt: slog.NewTextHandler(buf, createTextHandlerOptions()),
|
||||
}
|
||||
|
||||
return h
|
||||
@ -44,10 +48,11 @@ func (h *logfmt) Enabled(ctx context.Context, lvl slog.Level) bool {
|
||||
}
|
||||
|
||||
func (h *logfmt) WithAttrs(attrs []slog.Attr) slog.Handler {
|
||||
buf := bytes.NewBuffer([]byte{})
|
||||
return &logfmt{
|
||||
buf: bytes.NewBuffer([]byte{}),
|
||||
buf: buf,
|
||||
next: h.next.WithAttrs(slices.Clone(attrs)),
|
||||
txt: h.txt.WithAttrs(slices.Clone(attrs)),
|
||||
txt: slog.NewTextHandler(buf, createTextHandlerOptions()).WithAttrs(slices.Clone(attrs)),
|
||||
}
|
||||
}
|
||||
|
||||
@ -55,10 +60,11 @@ func (h *logfmt) WithGroup(g string) slog.Handler {
|
||||
if g == "" {
|
||||
return h
|
||||
}
|
||||
buf := bytes.NewBuffer([]byte{})
|
||||
return &logfmt{
|
||||
buf: bytes.NewBuffer([]byte{}),
|
||||
buf: buf,
|
||||
next: h.next.WithGroup(g),
|
||||
txt: h.txt.WithGroup(g),
|
||||
txt: slog.NewTextHandler(buf, createTextHandlerOptions()).WithGroup(g),
|
||||
}
|
||||
}
|
||||
|
||||
@ -70,10 +76,22 @@ func (h *logfmt) Handle(ctx context.Context, r slog.Record) error {
|
||||
panic("buffer wasn't empty")
|
||||
}
|
||||
|
||||
h.txt.Handle(ctx, r)
|
||||
r.Message = h.buf.String()
|
||||
r.Message = strings.TrimSuffix(r.Message, "\n")
|
||||
// Format using text handler to get the formatted message
|
||||
err := h.txt.Handle(ctx, r)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
formattedMessage := h.buf.String()
|
||||
formattedMessage = strings.TrimSuffix(formattedMessage, "\n")
|
||||
h.buf.Reset()
|
||||
|
||||
return h.next.Handle(ctx, r)
|
||||
// Create a new record with the formatted message
|
||||
newRecord := slog.NewRecord(r.Time, r.Level, formattedMessage, r.PC)
|
||||
r.Attrs(func(a slog.Attr) bool {
|
||||
newRecord.AddAttrs(a)
|
||||
return true
|
||||
})
|
||||
|
||||
return h.next.Handle(ctx, newRecord)
|
||||
}
|
||||
|
@ -9,7 +9,6 @@ import (
|
||||
)
|
||||
|
||||
func TestLogFmt(t *testing.T) {
|
||||
|
||||
var buf bytes.Buffer
|
||||
jsonh := slog.NewJSONHandler(&buf, nil)
|
||||
h := newLogFmtHandler(jsonh)
|
||||
@ -39,5 +38,4 @@ func TestLogFmt(t *testing.T) {
|
||||
t.Log("didn't find message in output")
|
||||
t.Fail()
|
||||
}
|
||||
|
||||
}
|
||||
|
235
logger/logger.go
235
logger/logger.go
@ -1,44 +1,132 @@
|
||||
// Package logger provides structured logging with OpenTelemetry trace integration.
|
||||
//
|
||||
// This package offers multiple logging configurations for different deployment scenarios:
|
||||
// - Text logging to stderr with optional timestamp removal for systemd
|
||||
// - OTLP (OpenTelemetry Protocol) logging for observability pipelines
|
||||
// - Multi-logger setup that outputs to both text and OTLP simultaneously
|
||||
// - Context-aware logging with trace ID correlation
|
||||
//
|
||||
// The package automatically detects systemd environments and adjusts timestamp handling
|
||||
// accordingly. It supports debug level configuration via environment variables and
|
||||
// provides compatibility bridges for legacy logging interfaces.
|
||||
//
|
||||
// Key features:
|
||||
// - Automatic OpenTelemetry trace and span ID inclusion in log entries
|
||||
// - Configurable log levels via DEBUG environment variable (with optional prefix)
|
||||
// - Systemd-compatible output (no timestamps when INVOCATION_ID is present)
|
||||
// - Thread-safe logger setup with sync.Once protection
|
||||
// - Context propagation for request-scoped logging
|
||||
//
|
||||
// Environment variables:
|
||||
// - LOG_LEVEL: Set stderr log level (DEBUG, INFO, WARN, ERROR) (configurable prefix via ConfigPrefix)
|
||||
// - OTLP_LOG_LEVEL: Set OTLP log level independently (configurable prefix via ConfigPrefix)
|
||||
// - DEBUG: Enable debug level logging for backward compatibility (configurable prefix via ConfigPrefix)
|
||||
// - INVOCATION_ID: Systemd detection for timestamp handling
|
||||
package logger
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"log"
|
||||
"log/slog"
|
||||
"os"
|
||||
"strconv"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
slogtraceid "github.com/remychantenay/slog-otel"
|
||||
slogmulti "github.com/samber/slog-multi"
|
||||
"go.opentelemetry.io/contrib/bridges/otelslog"
|
||||
"go.opentelemetry.io/otel/log/global"
|
||||
otellog "go.opentelemetry.io/otel/sdk/log"
|
||||
)
|
||||
|
||||
// ConfigPrefix allows customizing the environment variable prefix for configuration.
|
||||
// When set, environment variables like DEBUG become {ConfigPrefix}_DEBUG.
|
||||
// This enables multiple services to have independent logging configuration.
|
||||
var ConfigPrefix = ""
|
||||
|
||||
var textLogger *slog.Logger
|
||||
var otlpLogger *slog.Logger
|
||||
var multiLogger *slog.Logger
|
||||
var (
|
||||
// Level controls the log level for the default stderr logger.
|
||||
// Can be changed at runtime to adjust logging verbosity.
|
||||
Level = new(slog.LevelVar) // Info by default
|
||||
|
||||
var setupText sync.Once // this sets the default
|
||||
var setupOtlp sync.Once // this never sets the default
|
||||
var setupMulti sync.Once // this sets the default, and will always run after the others
|
||||
var mu sync.Mutex
|
||||
// OTLPLevel controls the log level for OTLP output.
|
||||
// Can be changed independently from the stderr logger level.
|
||||
OTLPLevel = new(slog.LevelVar) // Info by default
|
||||
)
|
||||
|
||||
var (
|
||||
textLogger *slog.Logger
|
||||
otlpLogger *slog.Logger
|
||||
multiLogger *slog.Logger
|
||||
)
|
||||
|
||||
var (
|
||||
setupText sync.Once // this sets the default
|
||||
setupOtlp sync.Once // this never sets the default
|
||||
setupMulti sync.Once // this sets the default, and will always run after the others
|
||||
mu sync.Mutex
|
||||
)
|
||||
|
||||
// SetLevel sets the log level for the default stderr logger.
|
||||
// This affects the primary application logger returned by Setup().
|
||||
func SetLevel(level slog.Level) {
|
||||
Level.Set(level)
|
||||
}
|
||||
|
||||
// SetOTLPLevel sets the log level for OTLP output.
|
||||
// This affects the logger returned by SetupOLTP() and the OTLP portion of SetupMultiLogger().
|
||||
func SetOTLPLevel(level slog.Level) {
|
||||
OTLPLevel.Set(level)
|
||||
}
|
||||
|
||||
// ParseLevel converts a string log level to slog.Level.
|
||||
// Supported levels: "DEBUG", "INFO", "WARN", "ERROR" (case insensitive).
|
||||
// Returns an error for unrecognized level strings.
|
||||
func ParseLevel(level string) (slog.Level, error) {
|
||||
switch {
|
||||
case level == "":
|
||||
return slog.LevelInfo, nil
|
||||
case level == "DEBUG" || level == "debug":
|
||||
return slog.LevelDebug, nil
|
||||
case level == "INFO" || level == "info":
|
||||
return slog.LevelInfo, nil
|
||||
case level == "WARN" || level == "warn":
|
||||
return slog.LevelWarn, nil
|
||||
case level == "ERROR" || level == "error":
|
||||
return slog.LevelError, nil
|
||||
default:
|
||||
return slog.LevelInfo, fmt.Errorf("unknown log level: %s", level)
|
||||
}
|
||||
}
|
||||
|
||||
func setupStdErrHandler() slog.Handler {
|
||||
var programLevel = new(slog.LevelVar) // Info by default
|
||||
|
||||
envVar := "DEBUG"
|
||||
// Parse LOG_LEVEL environment variable
|
||||
logLevelVar := "LOG_LEVEL"
|
||||
if len(ConfigPrefix) > 0 {
|
||||
envVar = ConfigPrefix + "_" + envVar
|
||||
logLevelVar = ConfigPrefix + "_" + logLevelVar
|
||||
}
|
||||
|
||||
if opt := os.Getenv(envVar); len(opt) > 0 {
|
||||
if levelStr := os.Getenv(logLevelVar); levelStr != "" {
|
||||
if level, err := ParseLevel(levelStr); err == nil {
|
||||
Level.Set(level)
|
||||
}
|
||||
}
|
||||
|
||||
// Maintain backward compatibility with DEBUG environment variable
|
||||
debugVar := "DEBUG"
|
||||
if len(ConfigPrefix) > 0 {
|
||||
debugVar = ConfigPrefix + "_" + debugVar
|
||||
}
|
||||
|
||||
if opt := os.Getenv(debugVar); len(opt) > 0 {
|
||||
if debug, _ := strconv.ParseBool(opt); debug {
|
||||
programLevel.Set(slog.LevelDebug)
|
||||
Level.Set(slog.LevelDebug)
|
||||
}
|
||||
}
|
||||
|
||||
logOptions := &slog.HandlerOptions{Level: programLevel}
|
||||
logOptions := &slog.HandlerOptions{Level: Level}
|
||||
|
||||
if len(os.Getenv("INVOCATION_ID")) > 0 {
|
||||
// don't add timestamps when running under systemd
|
||||
@ -56,17 +144,54 @@ func setupStdErrHandler() slog.Handler {
|
||||
|
||||
func setupOtlpLogger() *slog.Logger {
|
||||
setupOtlp.Do(func() {
|
||||
otlpLogger = slog.New(
|
||||
newLogFmtHandler(otelslog.NewHandler("common")),
|
||||
// Parse OTLP_LOG_LEVEL environment variable
|
||||
otlpLevelVar := "OTLP_LOG_LEVEL"
|
||||
if len(ConfigPrefix) > 0 {
|
||||
otlpLevelVar = ConfigPrefix + "_" + otlpLevelVar
|
||||
}
|
||||
|
||||
if levelStr := os.Getenv(otlpLevelVar); levelStr != "" {
|
||||
if level, err := ParseLevel(levelStr); err == nil {
|
||||
OTLPLevel.Set(level)
|
||||
}
|
||||
}
|
||||
|
||||
// Create our buffering exporter
|
||||
// It will buffer until tracing is configured
|
||||
bufferingExp := newBufferingExporter()
|
||||
|
||||
// Use BatchProcessor with our custom exporter
|
||||
processor := otellog.NewBatchProcessor(bufferingExp,
|
||||
otellog.WithExportInterval(10*time.Second),
|
||||
otellog.WithMaxQueueSize(2048),
|
||||
otellog.WithExportMaxBatchSize(512),
|
||||
)
|
||||
|
||||
// Create logger provider
|
||||
provider := otellog.NewLoggerProvider(
|
||||
otellog.WithProcessor(processor),
|
||||
)
|
||||
|
||||
// Set global provider
|
||||
global.SetLoggerProvider(provider)
|
||||
|
||||
// Create slog handler with level control
|
||||
baseHandler := newLogFmtHandler(otelslog.NewHandler("common"))
|
||||
handler := newOTLPLevelHandler(baseHandler)
|
||||
otlpLogger = slog.New(handler)
|
||||
})
|
||||
return otlpLogger
|
||||
}
|
||||
|
||||
// SetupMultiLogger will setup and make default a logger that
|
||||
// logs as described in Setup() as well as an OLTP logger.
|
||||
// The "multi logger" is made the default the first time
|
||||
// this function is called
|
||||
// SetupMultiLogger creates a logger that outputs to both text (stderr) and OTLP simultaneously.
|
||||
// This is useful for services that need both human-readable logs and structured observability data.
|
||||
//
|
||||
// The multi-logger combines:
|
||||
// - Text handler: Stderr output with OpenTelemetry trace integration
|
||||
// - OTLP handler: Structured logs sent via OpenTelemetry Protocol
|
||||
//
|
||||
// On first call, this logger becomes the default logger returned by Setup().
|
||||
// The function is thread-safe and uses sync.Once to ensure single initialization.
|
||||
func SetupMultiLogger() *slog.Logger {
|
||||
setupMulti.Do(func() {
|
||||
textHandler := Setup().Handler()
|
||||
@ -85,28 +210,38 @@ func SetupMultiLogger() *slog.Logger {
|
||||
return multiLogger
|
||||
}
|
||||
|
||||
// SetupOLTP configures and returns a logger sending logs
|
||||
// via OpenTelemetry (configured via the tracing package).
|
||||
// SetupOLTP creates a logger that sends structured logs via OpenTelemetry Protocol.
|
||||
// This logger is designed for observability pipelines and log aggregation systems.
|
||||
//
|
||||
// This was made to work with Loki + Grafana that makes it
|
||||
// hard to view the log attributes in the UI, so the log
|
||||
// message is formatted similarly to the text logger. The
|
||||
// attributes are duplicated as OLTP attributes in the
|
||||
// log messages. https://github.com/grafana/loki/issues/14788
|
||||
// The OTLP logger formats log messages similarly to the text logger for better
|
||||
// compatibility with Loki + Grafana, while still providing structured attributes.
|
||||
// Log attributes are available both in the message format and as OTLP attributes.
|
||||
//
|
||||
// This logger does not become the default logger and must be used explicitly.
|
||||
// It requires OpenTelemetry tracing configuration to be set up via the tracing package.
|
||||
//
|
||||
// See: https://github.com/grafana/loki/issues/14788 for formatting rationale.
|
||||
func SetupOLTP() *slog.Logger {
|
||||
return setupOtlpLogger()
|
||||
}
|
||||
|
||||
// Setup returns an slog.Logger configured for text formatting
|
||||
// to stderr.
|
||||
// OpenTelemetry trace_id and span_id's are logged as attributes
|
||||
// when available.
|
||||
// When the application is running under systemd timestamps are
|
||||
// omitted. On first call the slog default logger is set to this
|
||||
// logger as well.
|
||||
// Setup creates and returns the standard text logger for the application.
|
||||
// This is the primary logging function that most applications should use.
|
||||
//
|
||||
// If SetupMultiLogger has been called Setup() will return
|
||||
// the "multi logger"
|
||||
// Features:
|
||||
// - Text formatting to stderr with human-readable output
|
||||
// - Automatic OpenTelemetry trace_id and span_id inclusion when available
|
||||
// - Systemd compatibility: omits timestamps when INVOCATION_ID environment variable is present
|
||||
// - Debug level support via DEBUG environment variable (respects ConfigPrefix)
|
||||
// - Thread-safe initialization with sync.Once
|
||||
//
|
||||
// On first call, this logger becomes the slog default logger. If SetupMultiLogger()
|
||||
// has been called previously, Setup() returns the multi-logger instead of the text logger.
|
||||
//
|
||||
// The logger automatically detects execution context:
|
||||
// - Systemd: Removes timestamps (systemd adds its own)
|
||||
// - Debug mode: Enables debug level logging based on environment variables
|
||||
// - OpenTelemetry: Includes trace correlation when tracing is active
|
||||
func Setup() *slog.Logger {
|
||||
setupText.Do(func() {
|
||||
h := setupStdErrHandler()
|
||||
@ -125,15 +260,33 @@ func Setup() *slog.Logger {
|
||||
|
||||
type loggerKey struct{}
|
||||
|
||||
// NewContext adds the logger to the context. Use this
|
||||
// to for example make a request specific logger available
|
||||
// to other functions through the context
|
||||
// NewContext stores a logger in the context for request-scoped logging.
|
||||
// This enables passing request-specific loggers (e.g., with request IDs,
|
||||
// user context, or other correlation data) through the call stack.
|
||||
//
|
||||
// Use this to create context-aware logging where different parts of the
|
||||
// application can access the same enriched logger instance.
|
||||
//
|
||||
// Example:
|
||||
//
|
||||
// logger := slog.With("request_id", requestID)
|
||||
// ctx := logger.NewContext(ctx, logger)
|
||||
// // Pass ctx to downstream functions
|
||||
func NewContext(ctx context.Context, l *slog.Logger) context.Context {
|
||||
return context.WithValue(ctx, loggerKey{}, l)
|
||||
}
|
||||
|
||||
// FromContext retrieves a logger from the context. If there is none,
|
||||
// it returns the default logger
|
||||
// FromContext retrieves a logger from the context.
|
||||
// If no logger is stored in the context, it returns the default logger from Setup().
|
||||
//
|
||||
// This function provides a safe way to access context-scoped loggers without
|
||||
// needing to check for nil values. It ensures that logging is always available,
|
||||
// falling back to the application's standard logger configuration.
|
||||
//
|
||||
// Example:
|
||||
//
|
||||
// log := logger.FromContext(ctx)
|
||||
// log.Info("processing request") // Uses context logger or default
|
||||
func FromContext(ctx context.Context) *slog.Logger {
|
||||
if l, ok := ctx.Value(loggerKey{}).(*slog.Logger); ok {
|
||||
return l
|
||||
|
48
logger/otlp_handler.go
Normal file
48
logger/otlp_handler.go
Normal file
@ -0,0 +1,48 @@
|
||||
package logger
|
||||
|
||||
import (
|
||||
"context"
|
||||
"log/slog"
|
||||
)
|
||||
|
||||
// otlpLevelHandler is a wrapper that enforces level checking for OTLP handlers.
|
||||
// This allows independent level control for OTLP output separate from stderr logging.
|
||||
type otlpLevelHandler struct {
|
||||
next slog.Handler
|
||||
}
|
||||
|
||||
// newOTLPLevelHandler creates a new OTLP level wrapper handler.
|
||||
func newOTLPLevelHandler(next slog.Handler) slog.Handler {
|
||||
return &otlpLevelHandler{
|
||||
next: next,
|
||||
}
|
||||
}
|
||||
|
||||
// Enabled checks if the log level should be processed by the OTLP handler.
|
||||
// It uses the OTLPLevel variable to determine if the record should be processed.
|
||||
func (h *otlpLevelHandler) Enabled(ctx context.Context, level slog.Level) bool {
|
||||
return level >= OTLPLevel.Level()
|
||||
}
|
||||
|
||||
// Handle processes the log record if the level is enabled.
|
||||
// If disabled by level checking, the record is silently dropped.
|
||||
func (h *otlpLevelHandler) Handle(ctx context.Context, r slog.Record) error {
|
||||
if !h.Enabled(ctx, r.Level) {
|
||||
return nil
|
||||
}
|
||||
return h.next.Handle(ctx, r)
|
||||
}
|
||||
|
||||
// WithAttrs returns a new handler with the specified attributes added.
|
||||
func (h *otlpLevelHandler) WithAttrs(attrs []slog.Attr) slog.Handler {
|
||||
return &otlpLevelHandler{
|
||||
next: h.next.WithAttrs(attrs),
|
||||
}
|
||||
}
|
||||
|
||||
// WithGroup returns a new handler with the specified group name.
|
||||
func (h *otlpLevelHandler) WithGroup(name string) slog.Handler {
|
||||
return &otlpLevelHandler{
|
||||
next: h.next.WithGroup(name),
|
||||
}
|
||||
}
|
@ -5,12 +5,24 @@ import (
|
||||
"log/slog"
|
||||
)
|
||||
|
||||
// stdLoggerish provides a bridge between legacy log interfaces and slog.
|
||||
// It implements common logging methods (Println, Printf, Fatalf) that
|
||||
// delegate to structured logging with a consistent key prefix.
|
||||
type stdLoggerish struct {
|
||||
key string
|
||||
log *slog.Logger
|
||||
f func(string, ...any)
|
||||
key string // Prefix key for all log messages
|
||||
log *slog.Logger // Underlying structured logger
|
||||
f func(string, ...any) // Log function (Info or Debug level)
|
||||
}
|
||||
|
||||
// NewStdLog creates a legacy-compatible logger that bridges to structured logging.
|
||||
// This is useful for third-party libraries that expect a standard log.Logger interface.
|
||||
//
|
||||
// Parameters:
|
||||
// - key: Prefix added to all log messages for identification
|
||||
// - debug: If true, logs at debug level; otherwise logs at info level
|
||||
// - log: Underlying slog.Logger (uses Setup() if nil)
|
||||
//
|
||||
// The returned logger implements Println, Printf, and Fatalf methods.
|
||||
func NewStdLog(key string, debug bool, log *slog.Logger) *stdLoggerish {
|
||||
if log == nil {
|
||||
log = Setup()
|
||||
@ -27,15 +39,19 @@ func NewStdLog(key string, debug bool, log *slog.Logger) *stdLoggerish {
|
||||
return sl
|
||||
}
|
||||
|
||||
func (l stdLoggerish) Println(msg ...interface{}) {
|
||||
// Println logs the arguments using the configured log level with the instance key.
|
||||
func (l stdLoggerish) Println(msg ...any) {
|
||||
l.f(l.key, "msg", msg)
|
||||
}
|
||||
|
||||
func (l stdLoggerish) Printf(msg string, args ...interface{}) {
|
||||
// Printf logs a formatted message using the configured log level with the instance key.
|
||||
func (l stdLoggerish) Printf(msg string, args ...any) {
|
||||
l.f(l.key, "msg", fmt.Sprintf(msg, args...))
|
||||
}
|
||||
|
||||
func (l stdLoggerish) Fatalf(msg string, args ...interface{}) {
|
||||
// Fatalf logs a formatted error message and panics.
|
||||
// Note: This implementation panics instead of calling os.Exit for testability.
|
||||
func (l stdLoggerish) Fatalf(msg string, args ...any) {
|
||||
l.log.Error(l.key, "msg", fmt.Sprintf(msg, args...))
|
||||
panic("fatal error") // todo: does this make sense at all?
|
||||
}
|
||||
|
@ -1,17 +0,0 @@
|
||||
package logger
|
||||
|
||||
type Error struct {
|
||||
Msg string
|
||||
Data []any
|
||||
}
|
||||
|
||||
func NewError(msg string, data ...any) *Error {
|
||||
return &Error{
|
||||
Msg: msg,
|
||||
Data: data,
|
||||
}
|
||||
}
|
||||
|
||||
func (e *Error) Error() string {
|
||||
return "not implemented"
|
||||
}
|
122
metrics/metrics.go
Normal file
122
metrics/metrics.go
Normal file
@ -0,0 +1,122 @@
|
||||
// Package metrics provides OpenTelemetry-native metrics with OTLP export support.
|
||||
//
|
||||
// This package implements a metrics system using the OpenTelemetry metrics data model
|
||||
// with OTLP export capabilities. It's designed for new applications that want to use
|
||||
// structured metrics export to observability backends.
|
||||
//
|
||||
// Key features:
|
||||
// - OpenTelemetry native metric types (Counter, Histogram, Gauge, etc.)
|
||||
// - OTLP export for sending metrics to observability backends
|
||||
// - Resource detection and correlation with traces/logs
|
||||
// - Graceful handling when OTLP configuration is not available
|
||||
//
|
||||
// Example usage:
|
||||
//
|
||||
// // Initialize metrics along with tracing
|
||||
// shutdown, err := tracing.InitTracer(ctx, cfg)
|
||||
// if err != nil {
|
||||
// log.Fatal(err)
|
||||
// }
|
||||
// defer shutdown(ctx)
|
||||
//
|
||||
// // Get a meter and create instruments
|
||||
// meter := metrics.GetMeter("my-service")
|
||||
// counter, _ := meter.Int64Counter("requests_total")
|
||||
// counter.Add(ctx, 1, metric.WithAttributes(attribute.String("method", "GET")))
|
||||
package metrics
|
||||
|
||||
import (
|
||||
"context"
|
||||
"log/slog"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"go.ntppool.org/common/internal/tracerconfig"
|
||||
"go.opentelemetry.io/otel"
|
||||
"go.opentelemetry.io/otel/metric"
|
||||
sdkmetric "go.opentelemetry.io/otel/sdk/metric"
|
||||
)
|
||||
|
||||
var (
|
||||
meterProvider metric.MeterProvider
|
||||
setupOnce sync.Once
|
||||
setupErr error
|
||||
)
|
||||
|
||||
// Setup initializes the OpenTelemetry metrics provider with OTLP export.
|
||||
// This function uses the configuration stored by the tracing package and
|
||||
// creates a metrics provider that exports to the same OTLP endpoint.
|
||||
//
|
||||
// The function is safe to call multiple times - it will only initialize once.
|
||||
// If tracing configuration is not available, it returns a no-op provider that
|
||||
// doesn't export metrics.
|
||||
//
|
||||
// Returns an error only if there's a configuration problem. Missing tracing
|
||||
// configuration is handled gracefully with a warning log.
|
||||
func Setup(ctx context.Context) error {
|
||||
setupOnce.Do(func() {
|
||||
setupErr = initializeMetrics(ctx)
|
||||
})
|
||||
return setupErr
|
||||
}
|
||||
|
||||
// GetMeter returns a named meter for creating metric instruments.
|
||||
// The meter uses the configured metrics provider, or the global provider
|
||||
// if metrics haven't been set up yet.
|
||||
//
|
||||
// This is the primary entry point for creating metric instruments in your application.
|
||||
func GetMeter(name string, opts ...metric.MeterOption) metric.Meter {
|
||||
if meterProvider == nil {
|
||||
// Return the global provider as fallback (no-op if not configured)
|
||||
return otel.GetMeterProvider().Meter(name, opts...)
|
||||
}
|
||||
return meterProvider.Meter(name, opts...)
|
||||
}
|
||||
|
||||
// initializeMetrics sets up the OpenTelemetry metrics provider with OTLP export.
|
||||
func initializeMetrics(ctx context.Context) error {
|
||||
log := slog.Default()
|
||||
|
||||
// Check if tracing configuration is available
|
||||
cfg, configCtx, factory := tracerconfig.GetMetricExporter()
|
||||
if cfg == nil || configCtx == nil || factory == nil {
|
||||
log.Warn("metrics setup: tracing configuration not available, using no-op provider")
|
||||
// Set the global provider as fallback - metrics just won't be exported
|
||||
meterProvider = otel.GetMeterProvider()
|
||||
return nil
|
||||
}
|
||||
|
||||
// Create OTLP metrics exporter
|
||||
exporter, err := factory(ctx, cfg)
|
||||
if err != nil {
|
||||
log.Error("metrics setup: failed to create OTLP exporter", "error", err)
|
||||
// Fall back to global provider
|
||||
meterProvider = otel.GetMeterProvider()
|
||||
return nil
|
||||
}
|
||||
|
||||
// Create metrics provider with the exporter
|
||||
provider := sdkmetric.NewMeterProvider(
|
||||
sdkmetric.WithReader(sdkmetric.NewPeriodicReader(
|
||||
exporter,
|
||||
sdkmetric.WithInterval(15*time.Second),
|
||||
)),
|
||||
)
|
||||
|
||||
// Set the global provider
|
||||
otel.SetMeterProvider(provider)
|
||||
meterProvider = provider
|
||||
|
||||
log.Info("metrics setup: OTLP metrics provider initialized")
|
||||
return nil
|
||||
}
|
||||
|
||||
// Shutdown gracefully shuts down the metrics provider.
|
||||
// This should be called during application shutdown to ensure all metrics
|
||||
// are properly flushed and exported.
|
||||
func Shutdown(ctx context.Context) error {
|
||||
if provider, ok := meterProvider.(*sdkmetric.MeterProvider); ok {
|
||||
return provider.Shutdown(ctx)
|
||||
}
|
||||
return nil
|
||||
}
|
296
metrics/metrics_test.go
Normal file
296
metrics/metrics_test.go
Normal file
@ -0,0 +1,296 @@
|
||||
package metrics
|
||||
|
||||
import (
|
||||
"context"
|
||||
"os"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"go.ntppool.org/common/internal/tracerconfig"
|
||||
"go.opentelemetry.io/otel/attribute"
|
||||
"go.opentelemetry.io/otel/metric"
|
||||
sdkmetric "go.opentelemetry.io/otel/sdk/metric"
|
||||
"go.opentelemetry.io/otel/sdk/metric/metricdata"
|
||||
)
|
||||
|
||||
func TestSetup_NoConfiguration(t *testing.T) {
|
||||
// Clear any existing configuration
|
||||
tracerconfig.Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
err := Setup(ctx)
|
||||
// Should not return an error even when no configuration is available
|
||||
if err != nil {
|
||||
t.Errorf("Setup() returned unexpected error: %v", err)
|
||||
}
|
||||
|
||||
// Should be able to get a meter (even if it's a no-op)
|
||||
meter := GetMeter("test-meter")
|
||||
if meter == nil {
|
||||
t.Error("GetMeter() returned nil")
|
||||
}
|
||||
}
|
||||
|
||||
func TestGetMeter(t *testing.T) {
|
||||
// Clear any existing configuration
|
||||
tracerconfig.Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
_ = Setup(ctx)
|
||||
|
||||
meter := GetMeter("test-service")
|
||||
if meter == nil {
|
||||
t.Fatal("GetMeter() returned nil")
|
||||
}
|
||||
|
||||
// Test creating a counter instrument
|
||||
counter, err := meter.Int64Counter("test_counter")
|
||||
if err != nil {
|
||||
t.Errorf("Failed to create counter: %v", err)
|
||||
}
|
||||
|
||||
// Test using the counter (should not error even with no-op provider)
|
||||
counter.Add(ctx, 1, metric.WithAttributes(attribute.String("test", "value")))
|
||||
}
|
||||
|
||||
func TestSetup_MultipleCallsSafe(t *testing.T) {
|
||||
// Clear any existing configuration
|
||||
tracerconfig.Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
|
||||
// Call Setup multiple times
|
||||
err1 := Setup(ctx)
|
||||
err2 := Setup(ctx)
|
||||
err3 := Setup(ctx)
|
||||
|
||||
if err1 != nil {
|
||||
t.Errorf("First Setup() call returned error: %v", err1)
|
||||
}
|
||||
if err2 != nil {
|
||||
t.Errorf("Second Setup() call returned error: %v", err2)
|
||||
}
|
||||
if err3 != nil {
|
||||
t.Errorf("Third Setup() call returned error: %v", err3)
|
||||
}
|
||||
|
||||
// Should still be able to get meters
|
||||
meter := GetMeter("test-meter")
|
||||
if meter == nil {
|
||||
t.Error("GetMeter() returned nil after multiple Setup() calls")
|
||||
}
|
||||
}
|
||||
|
||||
func TestSetup_WithConfiguration(t *testing.T) {
|
||||
// Clear any existing configuration
|
||||
tracerconfig.Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
config := &tracerconfig.Config{
|
||||
ServiceName: "test-metrics-service",
|
||||
Environment: "test",
|
||||
Endpoint: "localhost:4317", // Will likely fail to connect, but should set up provider
|
||||
}
|
||||
|
||||
// Create a mock exporter factory that returns a working exporter
|
||||
mockFactory := func(ctx context.Context, cfg *tracerconfig.Config) (sdkmetric.Exporter, error) {
|
||||
// Create a simple in-memory exporter for testing
|
||||
return &mockMetricExporter{}, nil
|
||||
}
|
||||
|
||||
// Store configuration with mock factory
|
||||
tracerconfig.Store(ctx, config, nil, mockFactory, nil)
|
||||
|
||||
// Setup metrics
|
||||
err := Setup(ctx)
|
||||
if err != nil {
|
||||
t.Errorf("Setup() returned error: %v", err)
|
||||
}
|
||||
|
||||
// Should be able to get a meter
|
||||
meter := GetMeter("test-service")
|
||||
if meter == nil {
|
||||
t.Fatal("GetMeter() returned nil")
|
||||
}
|
||||
|
||||
// Test creating and using instruments
|
||||
counter, err := meter.Int64Counter("test_counter")
|
||||
if err != nil {
|
||||
t.Errorf("Failed to create counter: %v", err)
|
||||
}
|
||||
|
||||
histogram, err := meter.Float64Histogram("test_histogram")
|
||||
if err != nil {
|
||||
t.Errorf("Failed to create histogram: %v", err)
|
||||
}
|
||||
|
||||
gauge, err := meter.Int64UpDownCounter("test_gauge")
|
||||
if err != nil {
|
||||
t.Errorf("Failed to create gauge: %v", err)
|
||||
}
|
||||
|
||||
// Use the instruments
|
||||
counter.Add(ctx, 1, metric.WithAttributes(attribute.String("test", "value")))
|
||||
histogram.Record(ctx, 1.5, metric.WithAttributes(attribute.String("test", "value")))
|
||||
gauge.Add(ctx, 10, metric.WithAttributes(attribute.String("test", "value")))
|
||||
|
||||
// Test shutdown
|
||||
err = Shutdown(ctx)
|
||||
if err != nil {
|
||||
t.Errorf("Shutdown() returned error: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestSetup_WithRealOTLPConfig(t *testing.T) {
|
||||
// Skip this test in short mode since it may try to make network connections
|
||||
if testing.Short() {
|
||||
t.Skip("Skipping integration test in short mode")
|
||||
}
|
||||
|
||||
// Clear any existing configuration
|
||||
tracerconfig.Clear()
|
||||
|
||||
// Set environment variables for OTLP configuration
|
||||
originalEndpoint := os.Getenv("OTEL_EXPORTER_OTLP_ENDPOINT")
|
||||
originalProtocol := os.Getenv("OTEL_EXPORTER_OTLP_PROTOCOL")
|
||||
|
||||
defer func() {
|
||||
if originalEndpoint != "" {
|
||||
os.Setenv("OTEL_EXPORTER_OTLP_ENDPOINT", originalEndpoint)
|
||||
} else {
|
||||
os.Unsetenv("OTEL_EXPORTER_OTLP_ENDPOINT")
|
||||
}
|
||||
if originalProtocol != "" {
|
||||
os.Setenv("OTEL_EXPORTER_OTLP_PROTOCOL", originalProtocol)
|
||||
} else {
|
||||
os.Unsetenv("OTEL_EXPORTER_OTLP_PROTOCOL")
|
||||
}
|
||||
}()
|
||||
|
||||
os.Setenv("OTEL_EXPORTER_OTLP_ENDPOINT", "http://localhost:4318") // HTTP endpoint
|
||||
os.Setenv("OTEL_EXPORTER_OTLP_PROTOCOL", "http/protobuf")
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
|
||||
defer cancel()
|
||||
|
||||
config := &tracerconfig.Config{
|
||||
ServiceName: "test-metrics-e2e",
|
||||
Environment: "test",
|
||||
Endpoint: "localhost:4318",
|
||||
}
|
||||
|
||||
// Store configuration with real factory
|
||||
tracerconfig.Store(ctx, config, nil, tracerconfig.CreateOTLPMetricExporter, nil)
|
||||
|
||||
// Setup metrics - this may fail if no OTLP collector is running, which is okay
|
||||
err := Setup(ctx)
|
||||
if err != nil {
|
||||
t.Logf("Setup() returned error (expected if no OTLP collector): %v", err)
|
||||
}
|
||||
|
||||
// Should still be able to get a meter
|
||||
meter := GetMeter("test-service-e2e")
|
||||
if meter == nil {
|
||||
t.Fatal("GetMeter() returned nil")
|
||||
}
|
||||
|
||||
// Create and use instruments
|
||||
counter, err := meter.Int64Counter("e2e_test_counter")
|
||||
if err != nil {
|
||||
t.Errorf("Failed to create counter: %v", err)
|
||||
}
|
||||
|
||||
// Add some metrics
|
||||
for i := 0; i < 5; i++ {
|
||||
counter.Add(ctx, 1, metric.WithAttributes(
|
||||
attribute.String("iteration", string(rune('0'+i))),
|
||||
attribute.String("test_type", "e2e"),
|
||||
))
|
||||
}
|
||||
|
||||
// Give some time for export (if collector is running)
|
||||
time.Sleep(100 * time.Millisecond)
|
||||
|
||||
// Test shutdown
|
||||
err = Shutdown(ctx)
|
||||
if err != nil {
|
||||
t.Logf("Shutdown() returned error (may be expected): %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestConcurrentMetricUsage(t *testing.T) {
|
||||
// Clear any existing configuration
|
||||
tracerconfig.Clear()
|
||||
|
||||
ctx := context.Background()
|
||||
config := &tracerconfig.Config{
|
||||
ServiceName: "concurrent-test",
|
||||
}
|
||||
|
||||
// Use mock factory
|
||||
mockFactory := func(ctx context.Context, cfg *tracerconfig.Config) (sdkmetric.Exporter, error) {
|
||||
return &mockMetricExporter{}, nil
|
||||
}
|
||||
|
||||
tracerconfig.Store(ctx, config, nil, mockFactory, nil)
|
||||
Setup(ctx)
|
||||
|
||||
meter := GetMeter("concurrent-test")
|
||||
counter, err := meter.Int64Counter("concurrent_counter")
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to create counter: %v", err)
|
||||
}
|
||||
|
||||
// Test concurrent metric usage
|
||||
const numGoroutines = 10
|
||||
const metricsPerGoroutine = 100
|
||||
|
||||
done := make(chan bool, numGoroutines)
|
||||
|
||||
for i := 0; i < numGoroutines; i++ {
|
||||
go func(goroutineID int) {
|
||||
for j := 0; j < metricsPerGoroutine; j++ {
|
||||
counter.Add(ctx, 1, metric.WithAttributes(
|
||||
attribute.Int("goroutine", goroutineID),
|
||||
attribute.Int("iteration", j),
|
||||
))
|
||||
}
|
||||
done <- true
|
||||
}(i)
|
||||
}
|
||||
|
||||
// Wait for all goroutines to complete
|
||||
for i := 0; i < numGoroutines; i++ {
|
||||
<-done
|
||||
}
|
||||
|
||||
// Shutdown
|
||||
err = Shutdown(ctx)
|
||||
if err != nil {
|
||||
t.Errorf("Shutdown() returned error: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
// mockMetricExporter is a simple mock exporter for testing
|
||||
type mockMetricExporter struct{}
|
||||
|
||||
func (m *mockMetricExporter) Export(ctx context.Context, rm *metricdata.ResourceMetrics) error {
|
||||
// Just pretend to export
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *mockMetricExporter) ForceFlush(ctx context.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *mockMetricExporter) Shutdown(ctx context.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *mockMetricExporter) Temporality(kind sdkmetric.InstrumentKind) metricdata.Temporality {
|
||||
return metricdata.CumulativeTemporality
|
||||
}
|
||||
|
||||
func (m *mockMetricExporter) Aggregation(kind sdkmetric.InstrumentKind) sdkmetric.Aggregation {
|
||||
return sdkmetric.DefaultAggregationSelector(kind)
|
||||
}
|
@ -1,3 +1,8 @@
|
||||
// Package metricsserver provides a standalone HTTP server for exposing Prometheus metrics.
|
||||
//
|
||||
// This package implements a dedicated metrics server that exposes application metrics
|
||||
// via HTTP. It uses a custom Prometheus registry to avoid conflicts with other metric
|
||||
// collectors and provides graceful shutdown capabilities.
|
||||
package metricsserver
|
||||
|
||||
import (
|
||||
@ -13,10 +18,13 @@ import (
|
||||
"go.ntppool.org/common/logger"
|
||||
)
|
||||
|
||||
// Metrics provides a custom Prometheus registry and HTTP handlers for metrics exposure.
|
||||
// It isolates application metrics from the default global registry.
|
||||
type Metrics struct {
|
||||
r *prometheus.Registry
|
||||
}
|
||||
|
||||
// New creates a new Metrics instance with a custom Prometheus registry.
|
||||
func New() *Metrics {
|
||||
r := prometheus.NewRegistry()
|
||||
|
||||
@ -27,12 +35,14 @@ func New() *Metrics {
|
||||
return m
|
||||
}
|
||||
|
||||
// Registry returns the custom Prometheus registry.
|
||||
// Use this to register your application's metrics collectors.
|
||||
func (m *Metrics) Registry() *prometheus.Registry {
|
||||
return m.r
|
||||
}
|
||||
|
||||
// Handler returns an HTTP handler for the /metrics endpoint with OpenMetrics support.
|
||||
func (m *Metrics) Handler() http.Handler {
|
||||
|
||||
log := logger.NewStdLog("prom http", false, nil)
|
||||
|
||||
return promhttp.HandlerFor(m.r, promhttp.HandlerOpts{
|
||||
@ -42,11 +52,9 @@ func (m *Metrics) Handler() http.Handler {
|
||||
})
|
||||
}
|
||||
|
||||
// ListenAndServe starts a goroutine with a server running on
|
||||
// the specified port. The server will shutdown and return when
|
||||
// the provided context is done
|
||||
// ListenAndServe starts a metrics server on the specified port and blocks until ctx is done.
|
||||
// The server exposes the metrics handler and shuts down gracefully when the context is cancelled.
|
||||
func (m *Metrics) ListenAndServe(ctx context.Context, port int) error {
|
||||
|
||||
log := logger.Setup()
|
||||
|
||||
srv := &http.Server{
|
||||
|
242
metricsserver/metrics_test.go
Normal file
242
metricsserver/metrics_test.go
Normal file
@ -0,0 +1,242 @@
|
||||
package metricsserver
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
"net/http/httptest"
|
||||
"strings"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
func TestNew(t *testing.T) {
|
||||
metrics := New()
|
||||
|
||||
if metrics == nil {
|
||||
t.Fatal("New returned nil")
|
||||
}
|
||||
|
||||
if metrics.r == nil {
|
||||
t.Error("metrics registry is nil")
|
||||
}
|
||||
}
|
||||
|
||||
func TestRegistry(t *testing.T) {
|
||||
metrics := New()
|
||||
registry := metrics.Registry()
|
||||
|
||||
if registry == nil {
|
||||
t.Fatal("Registry() returned nil")
|
||||
}
|
||||
|
||||
if registry != metrics.r {
|
||||
t.Error("Registry() did not return the metrics registry")
|
||||
}
|
||||
|
||||
// Test that we can register a metric
|
||||
counter := prometheus.NewCounter(prometheus.CounterOpts{
|
||||
Name: "test_counter",
|
||||
Help: "A test counter",
|
||||
})
|
||||
|
||||
err := registry.Register(counter)
|
||||
if err != nil {
|
||||
t.Errorf("failed to register metric: %v", err)
|
||||
}
|
||||
|
||||
// Test that the metric is registered
|
||||
metricFamilies, err := registry.Gather()
|
||||
if err != nil {
|
||||
t.Errorf("failed to gather metrics: %v", err)
|
||||
}
|
||||
|
||||
found := false
|
||||
for _, mf := range metricFamilies {
|
||||
if mf.GetName() == "test_counter" {
|
||||
found = true
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
if !found {
|
||||
t.Error("registered metric not found in registry")
|
||||
}
|
||||
}
|
||||
|
||||
func TestHandler(t *testing.T) {
|
||||
metrics := New()
|
||||
|
||||
// Register a test metric
|
||||
counter := prometheus.NewCounterVec(
|
||||
prometheus.CounterOpts{
|
||||
Name: "test_requests_total",
|
||||
Help: "Total number of test requests",
|
||||
},
|
||||
[]string{"method"},
|
||||
)
|
||||
metrics.Registry().MustRegister(counter)
|
||||
counter.WithLabelValues("GET").Inc()
|
||||
|
||||
// Test the handler
|
||||
handler := metrics.Handler()
|
||||
if handler == nil {
|
||||
t.Fatal("Handler() returned nil")
|
||||
}
|
||||
|
||||
// Create a test request
|
||||
req := httptest.NewRequest("GET", "/metrics", nil)
|
||||
recorder := httptest.NewRecorder()
|
||||
|
||||
// Call the handler
|
||||
handler.ServeHTTP(recorder, req)
|
||||
|
||||
// Check response
|
||||
resp := recorder.Result()
|
||||
defer resp.Body.Close()
|
||||
|
||||
if resp.StatusCode != http.StatusOK {
|
||||
t.Errorf("expected status 200, got %d", resp.StatusCode)
|
||||
}
|
||||
|
||||
body, err := io.ReadAll(resp.Body)
|
||||
if err != nil {
|
||||
t.Fatalf("failed to read response body: %v", err)
|
||||
}
|
||||
|
||||
bodyStr := string(body)
|
||||
|
||||
// Check for our test metric
|
||||
if !strings.Contains(bodyStr, "test_requests_total") {
|
||||
t.Error("test metric not found in metrics output")
|
||||
}
|
||||
|
||||
// Check for OpenMetrics format indicators
|
||||
if !strings.Contains(bodyStr, "# TYPE") {
|
||||
t.Error("metrics output missing TYPE comments")
|
||||
}
|
||||
}
|
||||
|
||||
func TestListenAndServe(t *testing.T) {
|
||||
metrics := New()
|
||||
|
||||
// Register a test metric
|
||||
counter := prometheus.NewCounterVec(
|
||||
prometheus.CounterOpts{
|
||||
Name: "test_requests_total",
|
||||
Help: "Total number of test requests",
|
||||
},
|
||||
[]string{"method"},
|
||||
)
|
||||
metrics.Registry().MustRegister(counter)
|
||||
counter.WithLabelValues("GET").Inc()
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
|
||||
// Start server in a goroutine
|
||||
errCh := make(chan error, 1)
|
||||
go func() {
|
||||
// Use a high port number to avoid conflicts
|
||||
errCh <- metrics.ListenAndServe(ctx, 9999)
|
||||
}()
|
||||
|
||||
// Give the server a moment to start
|
||||
time.Sleep(100 * time.Millisecond)
|
||||
|
||||
// Test metrics endpoint
|
||||
resp, err := http.Get("http://localhost:9999/metrics")
|
||||
if err != nil {
|
||||
t.Fatalf("failed to GET /metrics: %v", err)
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
|
||||
if resp.StatusCode != http.StatusOK {
|
||||
t.Errorf("expected status 200, got %d", resp.StatusCode)
|
||||
}
|
||||
|
||||
body, err := io.ReadAll(resp.Body)
|
||||
if err != nil {
|
||||
t.Fatalf("failed to read response body: %v", err)
|
||||
}
|
||||
|
||||
bodyStr := string(body)
|
||||
|
||||
// Check for our test metric
|
||||
if !strings.Contains(bodyStr, "test_requests_total") {
|
||||
t.Error("test metric not found in metrics output")
|
||||
}
|
||||
|
||||
// Cancel context to stop server
|
||||
cancel()
|
||||
|
||||
// Wait for server to stop
|
||||
select {
|
||||
case err := <-errCh:
|
||||
if err != nil {
|
||||
t.Errorf("server returned error: %v", err)
|
||||
}
|
||||
case <-time.After(5 * time.Second):
|
||||
t.Error("server did not stop within timeout")
|
||||
}
|
||||
}
|
||||
|
||||
func TestListenAndServeContextCancellation(t *testing.T) {
|
||||
metrics := New()
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
|
||||
// Start server
|
||||
errCh := make(chan error, 1)
|
||||
go func() {
|
||||
errCh <- metrics.ListenAndServe(ctx, 9998)
|
||||
}()
|
||||
|
||||
// Give server time to start
|
||||
time.Sleep(100 * time.Millisecond)
|
||||
|
||||
// Cancel context
|
||||
cancel()
|
||||
|
||||
// Server should stop gracefully
|
||||
select {
|
||||
case err := <-errCh:
|
||||
if err != nil {
|
||||
t.Errorf("server returned error on graceful shutdown: %v", err)
|
||||
}
|
||||
case <-time.After(5 * time.Second):
|
||||
t.Error("server did not stop within timeout after context cancellation")
|
||||
}
|
||||
}
|
||||
|
||||
// Benchmark the metrics handler response time
|
||||
func BenchmarkMetricsHandler(b *testing.B) {
|
||||
metrics := New()
|
||||
|
||||
// Register some test metrics
|
||||
for i := 0; i < 10; i++ {
|
||||
counter := prometheus.NewCounter(prometheus.CounterOpts{
|
||||
Name: fmt.Sprintf("bench_counter_%d", i),
|
||||
Help: "A benchmark counter",
|
||||
})
|
||||
metrics.Registry().MustRegister(counter)
|
||||
counter.Add(float64(i * 100))
|
||||
}
|
||||
|
||||
handler := metrics.Handler()
|
||||
|
||||
b.ResetTimer()
|
||||
|
||||
for i := 0; i < b.N; i++ {
|
||||
req := httptest.NewRequest("GET", "/metrics", nil)
|
||||
recorder := httptest.NewRecorder()
|
||||
handler.ServeHTTP(recorder, req)
|
||||
|
||||
if recorder.Code != http.StatusOK {
|
||||
b.Fatalf("unexpected status code: %d", recorder.Code)
|
||||
}
|
||||
}
|
||||
}
|
@ -15,7 +15,7 @@ mkdir -p $DIR
|
||||
|
||||
BASE=https://geodns.bitnames.com/${BASE}/builds/${BUILD}
|
||||
|
||||
files=`curl -sSf ${BASE}/checksums.txt | awk '{print $2}'`
|
||||
files=`curl -sSf ${BASE}/checksums.txt | sed 's/^[a-f0-9]*[[:space:]]*//'`
|
||||
metafiles="checksums.txt metadata.json CHANGELOG.md artifacts.json"
|
||||
|
||||
for f in $metafiles; do
|
||||
|
@ -2,7 +2,7 @@
|
||||
|
||||
set -euo pipefail
|
||||
|
||||
go install github.com/goreleaser/goreleaser/v2@v2.5.0
|
||||
go install github.com/goreleaser/goreleaser/v2@v2.11.0
|
||||
|
||||
if [ ! -z "${harbor_username:-}" ]; then
|
||||
DOCKER_FILE=~/.docker/config.json
|
||||
|
@ -1,3 +1,4 @@
|
||||
// Package timeutil provides JSON-serializable time utilities.
|
||||
package timeutil
|
||||
|
||||
import (
|
||||
@ -6,16 +7,39 @@ import (
|
||||
"time"
|
||||
)
|
||||
|
||||
// Duration is a wrapper around time.Duration that supports JSON marshaling/unmarshaling.
|
||||
//
|
||||
// When marshaling to JSON, it outputs the duration as a string using time.Duration.String().
|
||||
// When unmarshaling from JSON, it accepts both:
|
||||
// - String values that can be parsed by time.ParseDuration (e.g., "30s", "5m", "1h30m")
|
||||
// - Numeric values that represent nanoseconds as a float64
|
||||
//
|
||||
// This makes it compatible with configuration files and APIs that need to represent
|
||||
// durations in a human-readable format.
|
||||
//
|
||||
// Example usage:
|
||||
//
|
||||
// type Config struct {
|
||||
// Timeout timeutil.Duration `json:"timeout"`
|
||||
// }
|
||||
//
|
||||
// // JSON: {"timeout": "30s"}
|
||||
// // or: {"timeout": 30000000000}
|
||||
type Duration struct {
|
||||
time.Duration
|
||||
}
|
||||
|
||||
// MarshalJSON implements json.Marshaler.
|
||||
// It marshals the duration as a string using time.Duration.String().
|
||||
func (d Duration) MarshalJSON() ([]byte, error) {
|
||||
return json.Marshal(time.Duration(d.Duration).String())
|
||||
}
|
||||
|
||||
// UnmarshalJSON implements json.Unmarshaler.
|
||||
// It accepts both string values (parsed via time.ParseDuration) and
|
||||
// numeric values (interpreted as nanoseconds).
|
||||
func (d *Duration) UnmarshalJSON(b []byte) error {
|
||||
var v interface{}
|
||||
var v any
|
||||
if err := json.Unmarshal(b, &v); err != nil {
|
||||
return err
|
||||
}
|
||||
|
@ -18,5 +18,4 @@ func TestDuration(t *testing.T) {
|
||||
if foo.Foo.Seconds() != 30 {
|
||||
t.Fatalf("parsed time.Duration wasn't 30 seconds: %s", foo.Foo)
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -1,3 +1,36 @@
|
||||
// Package tracing provides OpenTelemetry distributed tracing setup with OTLP export support.
|
||||
//
|
||||
// This package handles the complete OpenTelemetry SDK initialization including:
|
||||
// - Trace provider configuration with batching and resource detection
|
||||
// - Log provider setup for structured log export via OTLP
|
||||
// - Automatic resource discovery (service name, version, host, container, process info)
|
||||
// - Support for both gRPC and HTTP OTLP exporters with TLS configuration
|
||||
// - Propagation context setup for distributed tracing across services
|
||||
// - Graceful shutdown handling for all telemetry components
|
||||
//
|
||||
// The package supports various deployment scenarios:
|
||||
// - Development: Local OTLP collectors or observability backends
|
||||
// - Production: Secure OTLP export with mutual TLS authentication
|
||||
// - Container environments: Automatic container and Kubernetes resource detection
|
||||
//
|
||||
// Configuration is primarily handled via standard OpenTelemetry environment variables:
|
||||
// - OTEL_SERVICE_NAME: Service identification
|
||||
// - OTEL_EXPORTER_OTLP_PROTOCOL: Protocol selection (grpc, http/protobuf)
|
||||
// - OTEL_TRACES_EXPORTER: Exporter type (otlp, autoexport)
|
||||
// - OTEL_RESOURCE_ATTRIBUTES: Additional resource attributes
|
||||
//
|
||||
// Example usage:
|
||||
//
|
||||
// cfg := &tracing.TracerConfig{
|
||||
// ServiceName: "my-service",
|
||||
// Environment: "production",
|
||||
// Endpoint: "https://otlp.example.com:4317",
|
||||
// }
|
||||
// shutdown, err := tracing.InitTracer(ctx, cfg)
|
||||
// if err != nil {
|
||||
// log.Fatal(err)
|
||||
// }
|
||||
// defer shutdown(ctx)
|
||||
package tracing
|
||||
|
||||
// todo, review:
|
||||
@ -5,26 +38,23 @@ package tracing
|
||||
|
||||
import (
|
||||
"context"
|
||||
"crypto/tls"
|
||||
"crypto/x509"
|
||||
"errors"
|
||||
"log/slog"
|
||||
"os"
|
||||
"slices"
|
||||
"time"
|
||||
|
||||
"go.ntppool.org/common/logger"
|
||||
"go.ntppool.org/common/internal/tracerconfig"
|
||||
"go.ntppool.org/common/version"
|
||||
"google.golang.org/grpc/credentials"
|
||||
|
||||
"go.opentelemetry.io/contrib/exporters/autoexport"
|
||||
"go.opentelemetry.io/otel"
|
||||
"go.opentelemetry.io/otel/attribute"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlptrace"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracegrpc"
|
||||
"go.opentelemetry.io/otel/exporters/otlp/otlptrace/otlptracehttp"
|
||||
logglobal "go.opentelemetry.io/otel/log/global"
|
||||
"go.opentelemetry.io/otel/log/global"
|
||||
"go.opentelemetry.io/otel/propagation"
|
||||
sdklog "go.opentelemetry.io/otel/sdk/log"
|
||||
sdkmetric "go.opentelemetry.io/otel/sdk/metric"
|
||||
"go.opentelemetry.io/otel/sdk/resource"
|
||||
sdktrace "go.opentelemetry.io/otel/sdk/trace"
|
||||
semconv "go.opentelemetry.io/otel/semconv/v1.26.0"
|
||||
@ -34,49 +64,106 @@ import (
|
||||
const (
|
||||
// svcNameKey is the environment variable name that Service Name information will be read from.
|
||||
svcNameKey = "OTEL_SERVICE_NAME"
|
||||
|
||||
otelExporterOTLPProtoEnvKey = "OTEL_EXPORTER_OTLP_PROTOCOL"
|
||||
otelExporterOTLPTracesProtoEnvKey = "OTEL_EXPORTER_OTLP_TRACES_PROTOCOL"
|
||||
)
|
||||
|
||||
var errInvalidOTLPProtocol = errors.New("invalid OTLP protocol - should be one of ['grpc', 'http/protobuf']")
|
||||
// createOTLPLogExporter creates an OTLP log exporter using the provided configuration.
|
||||
// This function is used as the LogExporterFactory for the tracerconfig bridge.
|
||||
func createOTLPLogExporter(ctx context.Context, cfg *tracerconfig.Config) (sdklog.Exporter, error) {
|
||||
return tracerconfig.CreateOTLPLogExporter(ctx, cfg)
|
||||
}
|
||||
|
||||
// createOTLPMetricExporter creates an OTLP metric exporter using the provided configuration.
|
||||
// This function is used as the MetricExporterFactory for the tracerconfig bridge.
|
||||
func createOTLPMetricExporter(ctx context.Context, cfg *tracerconfig.Config) (sdkmetric.Exporter, error) {
|
||||
return tracerconfig.CreateOTLPMetricExporter(ctx, cfg)
|
||||
}
|
||||
|
||||
// createOTLPTraceExporter creates an OTLP trace exporter using the provided configuration.
|
||||
// This function is used as the TraceExporterFactory for the tracerconfig bridge.
|
||||
func createOTLPTraceExporter(ctx context.Context, cfg *tracerconfig.Config) (sdktrace.SpanExporter, error) {
|
||||
return tracerconfig.CreateOTLPTraceExporter(ctx, cfg)
|
||||
}
|
||||
|
||||
// https://github.com/open-telemetry/opentelemetry-go/blob/main/exporters/otlp/otlptrace/otlptracehttp/example_test.go
|
||||
|
||||
// TpShutdownFunc represents a function that gracefully shuts down telemetry providers.
|
||||
// It should be called during application shutdown to ensure all telemetry data is flushed
|
||||
// and exporters are properly closed. The context can be used to set shutdown timeouts.
|
||||
type TpShutdownFunc func(ctx context.Context) error
|
||||
|
||||
// Tracer returns the configured OpenTelemetry tracer for the NTP Pool project.
|
||||
// This tracer should be used for creating spans and distributed tracing throughout
|
||||
// the application. It uses the global tracer provider set up by InitTracer/SetupSDK.
|
||||
func Tracer() trace.Tracer {
|
||||
traceProvider := otel.GetTracerProvider()
|
||||
return traceProvider.Tracer("ntppool-tracer")
|
||||
}
|
||||
|
||||
// Start creates a new span with the given name and options using the configured tracer.
|
||||
// This is a convenience function that wraps the standard OpenTelemetry span creation.
|
||||
// It returns a new context containing the span and the span itself for further configuration.
|
||||
//
|
||||
// The returned context should be used for downstream operations to maintain trace correlation.
|
||||
func Start(ctx context.Context, spanName string, opts ...trace.SpanStartOption) (context.Context, trace.Span) {
|
||||
return Tracer().Start(ctx, spanName, opts...)
|
||||
}
|
||||
|
||||
type GetClientCertificate func(*tls.CertificateRequestInfo) (*tls.Certificate, error)
|
||||
// GetClientCertificate is an alias for the type defined in tracerconfig.
|
||||
// This maintains backward compatibility for existing code.
|
||||
type GetClientCertificate = tracerconfig.GetClientCertificate
|
||||
|
||||
// TracerConfig provides configuration options for OpenTelemetry tracing setup.
|
||||
// It supplements standard OpenTelemetry environment variables with additional
|
||||
// NTP Pool-specific configuration including TLS settings for secure OTLP export.
|
||||
type TracerConfig struct {
|
||||
ServiceName string
|
||||
Environment string
|
||||
Endpoint string
|
||||
EndpointURL string
|
||||
ServiceName string // Service name for resource identification (overrides OTEL_SERVICE_NAME)
|
||||
Environment string // Deployment environment (development, staging, production)
|
||||
Endpoint string // OTLP endpoint hostname/port (e.g., "otlp.example.com:4317")
|
||||
EndpointURL string // Complete OTLP endpoint URL (e.g., "https://otlp.example.com:4317/v1/traces")
|
||||
|
||||
CertificateProvider GetClientCertificate
|
||||
RootCAs *x509.CertPool
|
||||
CertificateProvider GetClientCertificate // Client certificate provider for mutual TLS
|
||||
RootCAs *x509.CertPool // CA certificate pool for server verification
|
||||
}
|
||||
|
||||
// InitTracer initializes the OpenTelemetry SDK with the provided configuration.
|
||||
// This is the main entry point for setting up distributed tracing in applications.
|
||||
//
|
||||
// The function configures trace and log providers, sets up OTLP exporters,
|
||||
// and returns a shutdown function that must be called during application termination.
|
||||
//
|
||||
// Returns a shutdown function and an error. The shutdown function should be called
|
||||
// with a context that has an appropriate timeout for graceful shutdown.
|
||||
func InitTracer(ctx context.Context, cfg *TracerConfig) (TpShutdownFunc, error) {
|
||||
// todo: setup environment from cfg
|
||||
return SetupSDK(ctx, cfg)
|
||||
}
|
||||
|
||||
// SetupSDK performs the complete OpenTelemetry SDK initialization including resource
|
||||
// discovery, exporter configuration, provider setup, and shutdown function creation.
|
||||
//
|
||||
// The function automatically discovers system resources (service info, host, container,
|
||||
// process details) and configures both trace and log exporters. It supports multiple
|
||||
// OTLP protocols (gRPC, HTTP) and handles TLS configuration for secure deployments.
|
||||
//
|
||||
// The returned shutdown function coordinates graceful shutdown of all telemetry
|
||||
// components in the reverse order of their initialization.
|
||||
func SetupSDK(ctx context.Context, cfg *TracerConfig) (shutdown TpShutdownFunc, err error) {
|
||||
if cfg == nil {
|
||||
cfg = &TracerConfig{}
|
||||
}
|
||||
|
||||
log := logger.Setup()
|
||||
// Store configuration for use by logger and metrics packages via bridge
|
||||
bridgeConfig := &tracerconfig.Config{
|
||||
ServiceName: cfg.ServiceName,
|
||||
Environment: cfg.Environment,
|
||||
Endpoint: cfg.Endpoint,
|
||||
EndpointURL: cfg.EndpointURL,
|
||||
CertificateProvider: cfg.CertificateProvider,
|
||||
RootCAs: cfg.RootCAs,
|
||||
}
|
||||
tracerconfig.Store(ctx, bridgeConfig, createOTLPLogExporter, createOTLPMetricExporter, createOTLPTraceExporter)
|
||||
|
||||
log := slog.Default()
|
||||
|
||||
if serviceName := os.Getenv(svcNameKey); len(serviceName) == 0 {
|
||||
if len(cfg.ServiceName) > 0 {
|
||||
@ -117,13 +204,21 @@ func SetupSDK(ctx context.Context, cfg *TracerConfig) (shutdown TpShutdownFunc,
|
||||
|
||||
var shutdownFuncs []func(context.Context) error
|
||||
shutdown = func(ctx context.Context) error {
|
||||
// Force flush the global logger provider before shutting down anything else
|
||||
if loggerProvider := global.GetLoggerProvider(); loggerProvider != nil {
|
||||
if sdkProvider, ok := loggerProvider.(*sdklog.LoggerProvider); ok {
|
||||
if flushErr := sdkProvider.ForceFlush(ctx); flushErr != nil {
|
||||
log.Warn("logger provider force flush failed", "err", flushErr)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
var err error
|
||||
// need to shutdown the providers first,
|
||||
// exporters after which is the opposite
|
||||
// order they are setup.
|
||||
slices.Reverse(shutdownFuncs)
|
||||
for _, fn := range shutdownFuncs {
|
||||
// log.Warn("shutting down", "fn", fn)
|
||||
err = errors.Join(err, fn(ctx))
|
||||
}
|
||||
shutdownFuncs = nil
|
||||
@ -145,9 +240,9 @@ func SetupSDK(ctx context.Context, cfg *TracerConfig) (shutdown TpShutdownFunc,
|
||||
|
||||
switch os.Getenv("OTEL_TRACES_EXPORTER") {
|
||||
case "":
|
||||
spanExporter, err = newOLTPExporter(ctx, cfg)
|
||||
spanExporter, err = createOTLPTraceExporter(ctx, bridgeConfig)
|
||||
case "otlp":
|
||||
spanExporter, err = newOLTPExporter(ctx, cfg)
|
||||
spanExporter, err = createOTLPTraceExporter(ctx, bridgeConfig)
|
||||
default:
|
||||
// log.Debug("OTEL_TRACES_EXPORTER", "fallback", os.Getenv("OTEL_TRACES_EXPORTER"))
|
||||
spanExporter, err = autoexport.NewSpanExporter(ctx)
|
||||
@ -158,13 +253,6 @@ func SetupSDK(ctx context.Context, cfg *TracerConfig) (shutdown TpShutdownFunc,
|
||||
}
|
||||
shutdownFuncs = append(shutdownFuncs, spanExporter.Shutdown)
|
||||
|
||||
logExporter, err := autoexport.NewLogExporter(ctx)
|
||||
if err != nil {
|
||||
handleErr(err)
|
||||
return
|
||||
}
|
||||
shutdownFuncs = append(shutdownFuncs, logExporter.Shutdown)
|
||||
|
||||
// Set up trace provider.
|
||||
tracerProvider, err := newTraceProvider(spanExporter, res)
|
||||
if err != nil {
|
||||
@ -174,19 +262,6 @@ func SetupSDK(ctx context.Context, cfg *TracerConfig) (shutdown TpShutdownFunc,
|
||||
shutdownFuncs = append(shutdownFuncs, tracerProvider.Shutdown)
|
||||
otel.SetTracerProvider(tracerProvider)
|
||||
|
||||
logProvider := sdklog.NewLoggerProvider(sdklog.WithResource(res),
|
||||
sdklog.WithProcessor(
|
||||
sdklog.NewBatchProcessor(logExporter, sdklog.WithExportBufferSize(10)),
|
||||
),
|
||||
)
|
||||
|
||||
logglobal.SetLoggerProvider(logProvider)
|
||||
shutdownFuncs = append(shutdownFuncs, func(ctx context.Context) error {
|
||||
logProvider.ForceFlush(ctx)
|
||||
return logProvider.Shutdown(ctx)
|
||||
},
|
||||
)
|
||||
|
||||
if err != nil {
|
||||
handleErr(err)
|
||||
return
|
||||
@ -195,75 +270,6 @@ func SetupSDK(ctx context.Context, cfg *TracerConfig) (shutdown TpShutdownFunc,
|
||||
return
|
||||
}
|
||||
|
||||
func newOLTPExporter(ctx context.Context, cfg *TracerConfig) (sdktrace.SpanExporter, error) {
|
||||
|
||||
log := logger.Setup()
|
||||
|
||||
var tlsConfig *tls.Config
|
||||
|
||||
if cfg.CertificateProvider != nil {
|
||||
tlsConfig = &tls.Config{
|
||||
GetClientCertificate: cfg.CertificateProvider,
|
||||
RootCAs: cfg.RootCAs,
|
||||
}
|
||||
}
|
||||
|
||||
proto := os.Getenv(otelExporterOTLPTracesProtoEnvKey)
|
||||
if proto == "" {
|
||||
proto = os.Getenv(otelExporterOTLPProtoEnvKey)
|
||||
}
|
||||
|
||||
// Fallback to default, http/protobuf.
|
||||
if proto == "" {
|
||||
proto = "http/protobuf"
|
||||
}
|
||||
|
||||
var client otlptrace.Client
|
||||
|
||||
switch proto {
|
||||
case "grpc":
|
||||
opts := []otlptracegrpc.Option{
|
||||
otlptracegrpc.WithCompressor("gzip"),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlptracegrpc.WithTLSCredentials(credentials.NewTLS(tlsConfig)))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
log.Info("adding option", "Endpoint", cfg.Endpoint)
|
||||
opts = append(opts, otlptracegrpc.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
log.Info("adding option", "EndpointURL", cfg.EndpointURL)
|
||||
opts = append(opts, otlptracegrpc.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
client = otlptracegrpc.NewClient(opts...)
|
||||
case "http/protobuf":
|
||||
opts := []otlptracehttp.Option{
|
||||
otlptracehttp.WithCompression(otlptracehttp.GzipCompression),
|
||||
}
|
||||
if tlsConfig != nil {
|
||||
opts = append(opts, otlptracehttp.WithTLSClientConfig(tlsConfig))
|
||||
}
|
||||
if len(cfg.Endpoint) > 0 {
|
||||
opts = append(opts, otlptracehttp.WithEndpoint(cfg.Endpoint))
|
||||
}
|
||||
if len(cfg.EndpointURL) > 0 {
|
||||
opts = append(opts, otlptracehttp.WithEndpointURL(cfg.EndpointURL))
|
||||
}
|
||||
|
||||
client = otlptracehttp.NewClient(opts...)
|
||||
default:
|
||||
return nil, errInvalidOTLPProtocol
|
||||
}
|
||||
|
||||
exporter, err := otlptrace.New(ctx, client)
|
||||
if err != nil {
|
||||
log.ErrorContext(ctx, "creating OTLP trace exporter", "err", err)
|
||||
}
|
||||
return exporter, err
|
||||
}
|
||||
|
||||
func newTraceProvider(traceExporter sdktrace.SpanExporter, res *resource.Resource) (*sdktrace.TracerProvider, error) {
|
||||
traceProvider := sdktrace.NewTracerProvider(
|
||||
sdktrace.WithResource(res),
|
||||
|
@ -7,7 +7,6 @@ import (
|
||||
)
|
||||
|
||||
func TestInit(t *testing.T) {
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
defer cancel()
|
||||
|
||||
@ -18,5 +17,4 @@ func TestInit(t *testing.T) {
|
||||
t.FailNow()
|
||||
}
|
||||
defer shutdownFn(ctx)
|
||||
|
||||
}
|
||||
|
@ -1,3 +1,17 @@
|
||||
// Package types provides shared data structures for the NTP Pool project.
|
||||
//
|
||||
// This package contains common types used across different NTP Pool services
|
||||
// for data exchange, logging, and database operations. The types are designed
|
||||
// to support JSON serialization for API responses and SQL database storage
|
||||
// with automatic marshaling/unmarshaling.
|
||||
//
|
||||
// Current types include:
|
||||
// - LogScoreAttributes: NTP server scoring metadata for monitoring and analysis
|
||||
//
|
||||
// All types implement appropriate interfaces for:
|
||||
// - JSON serialization (json.Marshaler/json.Unmarshaler)
|
||||
// - SQL database storage (database/sql/driver.Valuer/sql.Scanner)
|
||||
// - String representation for logging and debugging
|
||||
package types
|
||||
|
||||
import (
|
||||
@ -6,17 +20,26 @@ import (
|
||||
"errors"
|
||||
)
|
||||
|
||||
// LogScoreAttributes contains metadata about NTP server scoring and monitoring results.
|
||||
// This structure captures both NTP protocol-specific information (leap, stratum) and
|
||||
// operational data (errors, warnings, response status) for analysis and alerting.
|
||||
//
|
||||
// The type supports JSON serialization for API responses and database storage
|
||||
// via the database/sql/driver interfaces. Fields use omitempty tags to minimize
|
||||
// JSON payload size when values are at their zero state.
|
||||
type LogScoreAttributes struct {
|
||||
Leap int8 `json:"leap,omitempty"`
|
||||
Stratum int8 `json:"stratum,omitempty"`
|
||||
NoResponse bool `json:"no_response,omitempty"`
|
||||
Error string `json:"error,omitempty"`
|
||||
Warning string `json:"warning,omitempty"`
|
||||
Leap int8 `json:"leap,omitempty"` // NTP leap indicator (0=no warning, 1=+1s, 2=-1s, 3=unsynchronized)
|
||||
Stratum int8 `json:"stratum,omitempty"` // NTP stratum level (1=primary, 2-15=secondary, 16=unsynchronized)
|
||||
NoResponse bool `json:"no_response,omitempty"` // True if server failed to respond to NTP queries
|
||||
Error string `json:"error,omitempty"` // Error message if scoring failed
|
||||
Warning string `json:"warning,omitempty"` // Warning message for non-fatal issues
|
||||
|
||||
FromLSID int `json:"from_ls_id,omitempty"`
|
||||
FromSSID int `json:"from_ss_id,omitempty"`
|
||||
FromLSID int `json:"from_ls_id,omitempty"` // Source log server ID for traceability
|
||||
FromSSID int `json:"from_ss_id,omitempty"` // Source scoring system ID for traceability
|
||||
}
|
||||
|
||||
// String returns a JSON representation of the LogScoreAttributes for logging and debugging.
|
||||
// Returns an empty string if JSON marshaling fails.
|
||||
func (lsa *LogScoreAttributes) String() string {
|
||||
b, err := json.Marshal(lsa)
|
||||
if err != nil {
|
||||
@ -25,11 +48,18 @@ func (lsa *LogScoreAttributes) String() string {
|
||||
return string(b)
|
||||
}
|
||||
|
||||
// Value implements the database/sql/driver.Valuer interface for database storage.
|
||||
// It serializes the LogScoreAttributes to JSON for storage in SQL databases.
|
||||
// Returns the JSON bytes or an error if marshaling fails.
|
||||
func (lsa *LogScoreAttributes) Value() (driver.Value, error) {
|
||||
return json.Marshal(lsa)
|
||||
}
|
||||
|
||||
func (lsa *LogScoreAttributes) Scan(value interface{}) error {
|
||||
// Scan implements the database/sql.Scanner interface for reading from SQL databases.
|
||||
// It deserializes JSON data from the database back into LogScoreAttributes.
|
||||
// Supports both []byte and string input types, with nil values treated as no-op.
|
||||
// Returns an error if the input type is unsupported or JSON unmarshaling fails.
|
||||
func (lsa *LogScoreAttributes) Scan(value any) error {
|
||||
var source []byte
|
||||
_t := LogScoreAttributes{}
|
||||
|
||||
|
66
ulid/ulid.go
66
ulid/ulid.go
@ -1,48 +1,44 @@
|
||||
// Package ulid provides thread-safe ULID (Universally Unique Lexicographically Sortable Identifier) generation.
|
||||
//
|
||||
// ULIDs are 128-bit identifiers that are lexicographically sortable and contain
|
||||
// a timestamp component. This package uses cryptographically secure random
|
||||
// generation optimized for simplicity and performance in concurrent environments.
|
||||
package ulid
|
||||
|
||||
import (
|
||||
cryptorand "crypto/rand"
|
||||
"encoding/binary"
|
||||
"io"
|
||||
mathrand "math/rand"
|
||||
"os"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
oklid "github.com/oklog/ulid/v2"
|
||||
"go.ntppool.org/common/logger"
|
||||
)
|
||||
|
||||
var monotonicPool = sync.Pool{
|
||||
New: func() interface{} {
|
||||
|
||||
log := logger.Setup()
|
||||
|
||||
var seed int64
|
||||
err := binary.Read(cryptorand.Reader, binary.BigEndian, &seed)
|
||||
if err != nil {
|
||||
log.Error("crypto/rand error", "err", err)
|
||||
os.Exit(10)
|
||||
}
|
||||
|
||||
rand := mathrand.New(mathrand.NewSource(seed))
|
||||
|
||||
inc := uint64(mathrand.Int63())
|
||||
|
||||
// log.Printf("seed: %d", seed)
|
||||
// log.Printf("inc: %d", inc)
|
||||
|
||||
// inc = inc & ^uint64(1<<63) // only want 63 bits
|
||||
mono := oklid.Monotonic(rand, inc)
|
||||
return mono
|
||||
},
|
||||
}
|
||||
|
||||
// MakeULID generates a new ULID with the specified timestamp using cryptographically secure randomness.
|
||||
// The function is thread-safe and optimized for high-concurrency environments.
|
||||
//
|
||||
// This implementation prioritizes simplicity and performance over strict monotonicity within
|
||||
// the same millisecond. Each ULID is guaranteed to be unique and lexicographically sortable
|
||||
// across different timestamps.
|
||||
//
|
||||
// Returns a pointer to the generated ULID or an error if generation fails.
|
||||
// Generation should only fail under extreme circumstances (entropy exhaustion).
|
||||
func MakeULID(t time.Time) (*oklid.ULID, error) {
|
||||
|
||||
mono := monotonicPool.Get().(io.Reader)
|
||||
|
||||
id, err := oklid.New(oklid.Timestamp(t), mono)
|
||||
id, err := oklid.New(oklid.Timestamp(t), cryptorand.Reader)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
return &id, nil
|
||||
}
|
||||
|
||||
// Make generates a new ULID with the current timestamp using cryptographically secure randomness.
|
||||
// This is a convenience function equivalent to MakeULID(time.Now()).
|
||||
//
|
||||
// The function is thread-safe and optimized for high-concurrency environments.
|
||||
//
|
||||
// Returns a pointer to the generated ULID or an error if generation fails.
|
||||
// Generation should only fail under extreme circumstances (entropy exhaustion).
|
||||
func Make() (*oklid.ULID, error) {
|
||||
id, err := oklid.New(oklid.Now(), cryptorand.Reader)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
@ -1,25 +1,336 @@
|
||||
package ulid
|
||||
|
||||
import (
|
||||
cryptorand "crypto/rand"
|
||||
"sort"
|
||||
"sync"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
oklid "github.com/oklog/ulid/v2"
|
||||
)
|
||||
|
||||
func TestULID(t *testing.T) {
|
||||
func TestMakeULID(t *testing.T) {
|
||||
tm := time.Now()
|
||||
ul1, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
t.Logf("makeULID failed: %s", err)
|
||||
t.Fail()
|
||||
t.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
ul2, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
t.Logf("MakeULID failed: %s", err)
|
||||
t.Fail()
|
||||
t.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
|
||||
if ul1 == nil || ul2 == nil {
|
||||
t.Fatal("MakeULID returned nil ULID")
|
||||
}
|
||||
|
||||
if ul1.String() == ul2.String() {
|
||||
t.Logf("ul1 and ul2 got the same string: %s", ul1.String())
|
||||
t.Fail()
|
||||
t.Errorf("ul1 and ul2 should be different: %s", ul1.String())
|
||||
}
|
||||
|
||||
// Verify they have the same timestamp
|
||||
if ul1.Time() != ul2.Time() {
|
||||
t.Errorf("ULIDs with same input time should have same timestamp: %d != %d", ul1.Time(), ul2.Time())
|
||||
}
|
||||
|
||||
t.Logf("ulid string 1 and 2: %s | %s", ul1.String(), ul2.String())
|
||||
}
|
||||
|
||||
func TestMake(t *testing.T) {
|
||||
// Test Make() function (uses current time)
|
||||
ul1, err := Make()
|
||||
if err != nil {
|
||||
t.Fatalf("Make failed: %s", err)
|
||||
}
|
||||
|
||||
if ul1 == nil {
|
||||
t.Fatal("Make returned nil ULID")
|
||||
}
|
||||
|
||||
// Sleep a bit and generate another
|
||||
time.Sleep(2 * time.Millisecond)
|
||||
|
||||
ul2, err := Make()
|
||||
if err != nil {
|
||||
t.Fatalf("Make failed: %s", err)
|
||||
}
|
||||
|
||||
// Should be different ULIDs
|
||||
if ul1.String() == ul2.String() {
|
||||
t.Errorf("ULIDs from Make() should be different: %s", ul1.String())
|
||||
}
|
||||
|
||||
// Second should be later (or at least not earlier)
|
||||
if ul1.Time() > ul2.Time() {
|
||||
t.Errorf("second ULID should not have earlier timestamp: %d > %d", ul1.Time(), ul2.Time())
|
||||
}
|
||||
|
||||
t.Logf("Make() ULIDs: %s | %s", ul1.String(), ul2.String())
|
||||
}
|
||||
|
||||
func TestMakeULIDUniqueness(t *testing.T) {
|
||||
tm := time.Now()
|
||||
seen := make(map[string]bool)
|
||||
|
||||
for i := 0; i < 1000; i++ {
|
||||
ul, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeULID failed on iteration %d: %s", i, err)
|
||||
}
|
||||
|
||||
str := ul.String()
|
||||
if seen[str] {
|
||||
t.Errorf("duplicate ULID generated: %s", str)
|
||||
}
|
||||
seen[str] = true
|
||||
}
|
||||
}
|
||||
|
||||
func TestMakeUniqueness(t *testing.T) {
|
||||
seen := make(map[string]bool)
|
||||
|
||||
for i := 0; i < 1000; i++ {
|
||||
ul, err := Make()
|
||||
if err != nil {
|
||||
t.Fatalf("Make failed on iteration %d: %s", i, err)
|
||||
}
|
||||
|
||||
str := ul.String()
|
||||
if seen[str] {
|
||||
t.Errorf("duplicate ULID generated: %s", str)
|
||||
}
|
||||
seen[str] = true
|
||||
}
|
||||
}
|
||||
|
||||
func TestMakeULIDTimestampProgression(t *testing.T) {
|
||||
t1 := time.Now()
|
||||
ul1, err := MakeULID(t1)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
|
||||
// Wait to ensure different timestamp
|
||||
time.Sleep(2 * time.Millisecond)
|
||||
|
||||
t2 := time.Now()
|
||||
ul2, err := MakeULID(t2)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
|
||||
if ul1.Time() >= ul2.Time() {
|
||||
t.Errorf("second ULID should have later timestamp: %d >= %d", ul1.Time(), ul2.Time())
|
||||
}
|
||||
|
||||
if ul1.Compare(*ul2) >= 0 {
|
||||
t.Errorf("second ULID should be greater: %s >= %s", ul1.String(), ul2.String())
|
||||
}
|
||||
}
|
||||
|
||||
func TestMakeULIDConcurrency(t *testing.T) {
|
||||
const numGoroutines = 10
|
||||
const numULIDsPerGoroutine = 100
|
||||
|
||||
var wg sync.WaitGroup
|
||||
ulidChan := make(chan *oklid.ULID, numGoroutines*numULIDsPerGoroutine)
|
||||
tm := time.Now()
|
||||
|
||||
// Start multiple goroutines generating ULIDs concurrently
|
||||
for i := 0; i < numGoroutines; i++ {
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
for j := 0; j < numULIDsPerGoroutine; j++ {
|
||||
ul, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
t.Errorf("MakeULID failed: %s", err)
|
||||
return
|
||||
}
|
||||
ulidChan <- ul
|
||||
}
|
||||
}()
|
||||
}
|
||||
|
||||
wg.Wait()
|
||||
close(ulidChan)
|
||||
|
||||
// Collect all ULIDs and check uniqueness
|
||||
seen := make(map[string]bool)
|
||||
count := 0
|
||||
|
||||
for ul := range ulidChan {
|
||||
str := ul.String()
|
||||
if seen[str] {
|
||||
t.Errorf("duplicate ULID generated in concurrent test: %s", str)
|
||||
}
|
||||
seen[str] = true
|
||||
count++
|
||||
}
|
||||
|
||||
if count != numGoroutines*numULIDsPerGoroutine {
|
||||
t.Errorf("expected %d ULIDs, got %d", numGoroutines*numULIDsPerGoroutine, count)
|
||||
}
|
||||
}
|
||||
|
||||
func TestMakeConcurrency(t *testing.T) {
|
||||
const numGoroutines = 10
|
||||
const numULIDsPerGoroutine = 100
|
||||
|
||||
var wg sync.WaitGroup
|
||||
ulidChan := make(chan *oklid.ULID, numGoroutines*numULIDsPerGoroutine)
|
||||
|
||||
// Start multiple goroutines generating ULIDs concurrently
|
||||
for i := 0; i < numGoroutines; i++ {
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
for j := 0; j < numULIDsPerGoroutine; j++ {
|
||||
ul, err := Make()
|
||||
if err != nil {
|
||||
t.Errorf("Make failed: %s", err)
|
||||
return
|
||||
}
|
||||
ulidChan <- ul
|
||||
}
|
||||
}()
|
||||
}
|
||||
|
||||
wg.Wait()
|
||||
close(ulidChan)
|
||||
|
||||
// Collect all ULIDs and check uniqueness
|
||||
seen := make(map[string]bool)
|
||||
count := 0
|
||||
|
||||
for ul := range ulidChan {
|
||||
str := ul.String()
|
||||
if seen[str] {
|
||||
t.Errorf("duplicate ULID generated in concurrent test: %s", str)
|
||||
}
|
||||
seen[str] = true
|
||||
count++
|
||||
}
|
||||
|
||||
if count != numGoroutines*numULIDsPerGoroutine {
|
||||
t.Errorf("expected %d ULIDs, got %d", numGoroutines*numULIDsPerGoroutine, count)
|
||||
}
|
||||
}
|
||||
|
||||
func TestMakeULIDErrorHandling(t *testing.T) {
|
||||
// Test with various timestamps
|
||||
timestamps := []time.Time{
|
||||
time.Unix(0, 0), // Unix epoch
|
||||
time.Now(), // Current time
|
||||
time.Now().Add(time.Hour), // Future time
|
||||
}
|
||||
|
||||
for i, tm := range timestamps {
|
||||
ul, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
t.Errorf("MakeULID failed with timestamp %d: %s", i, err)
|
||||
}
|
||||
if ul == nil {
|
||||
t.Errorf("MakeULID returned nil ULID with timestamp %d", i)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestMakeULIDLexicographicOrdering(t *testing.T) {
|
||||
var ulids []*oklid.ULID
|
||||
var timestamps []time.Time
|
||||
|
||||
// Generate ULIDs with increasing timestamps
|
||||
for i := 0; i < 10; i++ {
|
||||
tm := time.Now().Add(time.Duration(i) * time.Millisecond)
|
||||
timestamps = append(timestamps, tm)
|
||||
|
||||
ul, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
ulids = append(ulids, ul)
|
||||
|
||||
// Small delay to ensure different timestamps
|
||||
time.Sleep(time.Millisecond)
|
||||
}
|
||||
|
||||
// Sort ULID strings lexicographically
|
||||
ulidStrings := make([]string, len(ulids))
|
||||
for i, ul := range ulids {
|
||||
ulidStrings[i] = ul.String()
|
||||
}
|
||||
|
||||
originalOrder := make([]string, len(ulidStrings))
|
||||
copy(originalOrder, ulidStrings)
|
||||
|
||||
sort.Strings(ulidStrings)
|
||||
|
||||
// Verify lexicographic order matches chronological order
|
||||
for i := 0; i < len(originalOrder); i++ {
|
||||
if originalOrder[i] != ulidStrings[i] {
|
||||
t.Errorf("lexicographic order doesn't match chronological order at index %d: %s != %s",
|
||||
i, originalOrder[i], ulidStrings[i])
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Benchmark ULID generation performance
|
||||
func BenchmarkMakeULID(b *testing.B) {
|
||||
tm := time.Now()
|
||||
|
||||
b.ResetTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
_, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
b.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Benchmark Make function
|
||||
func BenchmarkMake(b *testing.B) {
|
||||
b.ResetTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
_, err := Make()
|
||||
if err != nil {
|
||||
b.Fatalf("Make failed: %s", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Benchmark concurrent ULID generation
|
||||
func BenchmarkMakeULIDConcurrent(b *testing.B) {
|
||||
tm := time.Now()
|
||||
|
||||
b.RunParallel(func(pb *testing.PB) {
|
||||
for pb.Next() {
|
||||
_, err := MakeULID(tm)
|
||||
if err != nil {
|
||||
b.Fatalf("MakeULID failed: %s", err)
|
||||
}
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
// Benchmark concurrent Make function
|
||||
func BenchmarkMakeConcurrent(b *testing.B) {
|
||||
b.RunParallel(func(pb *testing.PB) {
|
||||
for pb.Next() {
|
||||
_, err := Make()
|
||||
if err != nil {
|
||||
b.Fatalf("Make failed: %s", err)
|
||||
}
|
||||
}
|
||||
})
|
||||
}
|
||||
|
||||
// Benchmark random number generation
|
||||
func BenchmarkCryptoRand(b *testing.B) {
|
||||
buf := make([]byte, 10) // ULID entropy size
|
||||
b.ResetTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
cryptorand.Read(buf)
|
||||
}
|
||||
}
|
||||
|
@ -1,3 +1,27 @@
|
||||
// Package version provides build metadata and version information management.
|
||||
//
|
||||
// This package manages application version information including semantic version,
|
||||
// Git revision, build time, and provides integration with CLI frameworks (Cobra, Kong)
|
||||
// and Prometheus metrics for operational visibility.
|
||||
//
|
||||
// Version information can be injected at build time using ldflags:
|
||||
//
|
||||
// go build -ldflags "-X go.ntppool.org/common/version.VERSION=v1.0.0 \
|
||||
// -X go.ntppool.org/common/version.buildTime=2023-01-01T00:00:00Z \
|
||||
// -X go.ntppool.org/common/version.gitVersion=abc123"
|
||||
//
|
||||
// Build time supports both Unix epoch timestamps and RFC3339 format:
|
||||
//
|
||||
// # Unix epoch (simpler, recommended)
|
||||
// go build -ldflags "-X go.ntppool.org/common/version.buildTime=$(date +%s)"
|
||||
//
|
||||
// # RFC3339 format
|
||||
// go build -ldflags "-X go.ntppool.org/common/version.buildTime=$(date -u +%Y-%m-%dT%H:%M:%SZ)"
|
||||
//
|
||||
// Both formats are automatically converted to RFC3339 for consistent output. The buildTime
|
||||
// parameter takes priority over Git commit time. If buildTime is not specified, the package
|
||||
// automatically extracts build information from Go's debug.BuildInfo when available,
|
||||
// providing fallback values for VCS time and revision.
|
||||
package version
|
||||
|
||||
import (
|
||||
@ -5,30 +29,60 @@ import (
|
||||
"log/slog"
|
||||
"runtime"
|
||||
"runtime/debug"
|
||||
"strconv"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
"github.com/spf13/cobra"
|
||||
"golang.org/x/mod/semver"
|
||||
)
|
||||
|
||||
// VERSION has the current software version (set in the build process)
|
||||
var VERSION string
|
||||
var buildTime string
|
||||
var gitVersion string
|
||||
var gitModified bool
|
||||
// VERSION contains the current software version (typically set during the build process via ldflags).
|
||||
// If not set, defaults to "dev-snapshot". The version should follow semantic versioning.
|
||||
var (
|
||||
VERSION string // Semantic version (e.g., "1.0.0" or "v1.0.0")
|
||||
buildTime string // Build timestamp (Unix epoch or RFC3339, normalized to RFC3339)
|
||||
gitVersion string // Git commit hash
|
||||
gitModified bool // Whether the working tree was modified during build
|
||||
)
|
||||
|
||||
// info holds the consolidated version information extracted from build variables and debug.BuildInfo.
|
||||
var info Info
|
||||
|
||||
// parseBuildTime converts a build time string to RFC3339 format.
|
||||
// Supports both Unix epoch timestamps (numeric strings) and RFC3339 format.
|
||||
// Returns the input unchanged if it cannot be parsed as either format.
|
||||
func parseBuildTime(s string) string {
|
||||
if s == "" {
|
||||
return s
|
||||
}
|
||||
|
||||
// Try parsing as Unix epoch timestamp (numeric string)
|
||||
if epoch, err := strconv.ParseInt(s, 10, 64); err == nil {
|
||||
return time.Unix(epoch, 0).UTC().Format(time.RFC3339)
|
||||
}
|
||||
|
||||
// Try parsing as RFC3339 to validate format
|
||||
if _, err := time.Parse(time.RFC3339, s); err == nil {
|
||||
return s // Already in RFC3339 format
|
||||
}
|
||||
|
||||
// Return original string if neither format works (graceful fallback)
|
||||
return s
|
||||
}
|
||||
|
||||
// Info represents structured version and build information.
|
||||
// This struct is used for JSON serialization and programmatic access to build metadata.
|
||||
type Info struct {
|
||||
Version string `json:",omitempty"`
|
||||
GitRev string `json:",omitempty"`
|
||||
GitRevShort string `json:",omitempty"`
|
||||
BuildTime string `json:",omitempty"`
|
||||
Version string `json:",omitempty"` // Semantic version with "v" prefix
|
||||
GitRev string `json:",omitempty"` // Full Git commit hash
|
||||
GitRevShort string `json:",omitempty"` // Shortened Git commit hash (7 characters)
|
||||
BuildTime string `json:",omitempty"` // Build timestamp
|
||||
}
|
||||
|
||||
func init() {
|
||||
|
||||
buildTime = parseBuildTime(buildTime)
|
||||
info.BuildTime = buildTime
|
||||
info.GitRev = gitVersion
|
||||
|
||||
@ -48,9 +102,9 @@ func init() {
|
||||
switch h.Key {
|
||||
case "vcs.time":
|
||||
if len(buildTime) == 0 {
|
||||
buildTime = h.Value
|
||||
buildTime = parseBuildTime(h.Value)
|
||||
info.BuildTime = buildTime
|
||||
}
|
||||
info.BuildTime = h.Value
|
||||
case "vcs.revision":
|
||||
// https://blog.carlmjohnson.net/post/2023/golang-git-hash-how-to/
|
||||
// todo: use BuildInfo.Main.Version if revision is empty
|
||||
@ -78,10 +132,16 @@ func init() {
|
||||
Version()
|
||||
}
|
||||
|
||||
// VersionCmd creates a Cobra command for displaying version information.
|
||||
// The name parameter is used as a prefix in the output (e.g., "myapp v1.0.0").
|
||||
// Returns a configured cobra.Command that can be added to any CLI application.
|
||||
func VersionCmd(name string) *cobra.Command {
|
||||
versionCmd := &cobra.Command{
|
||||
Use: "version",
|
||||
Short: "Print version and build information",
|
||||
Long: `Print detailed version information including semantic version,
|
||||
Git revision, build time, and Go version. Build information is automatically
|
||||
extracted from Go's debug.BuildInfo when available.`,
|
||||
Run: func(cmd *cobra.Command, args []string) {
|
||||
ver := Version()
|
||||
fmt.Printf("%s %s\n", name, ver)
|
||||
@ -90,6 +150,23 @@ func VersionCmd(name string) *cobra.Command {
|
||||
return versionCmd
|
||||
}
|
||||
|
||||
// KongVersionCmd provides a Kong CLI framework compatible version command.
|
||||
// The Name field should be set to the application name for proper output formatting.
|
||||
type KongVersionCmd struct {
|
||||
Name string `kong:"-"` // Application name, excluded from Kong parsing
|
||||
}
|
||||
|
||||
// Run executes the version command for Kong CLI framework.
|
||||
// Prints the application name and version information to stdout.
|
||||
func (cmd *KongVersionCmd) Run() error {
|
||||
fmt.Printf("%s %s\n", cmd.Name, Version())
|
||||
return nil
|
||||
}
|
||||
|
||||
// RegisterMetric registers a Prometheus gauge metric with build information.
|
||||
// If name is provided, it creates a metric named "{name}_build_info", otherwise "build_info".
|
||||
// The metric includes labels for version, build time, Git time, and Git revision.
|
||||
// This is useful for exposing build information in monitoring systems.
|
||||
func RegisterMetric(name string, registry prometheus.Registerer) {
|
||||
if len(name) > 0 {
|
||||
name = strings.ReplaceAll(name, "-", "_")
|
||||
@ -100,13 +177,13 @@ func RegisterMetric(name string, registry prometheus.Registerer) {
|
||||
buildInfo := prometheus.NewGaugeVec(
|
||||
prometheus.GaugeOpts{
|
||||
Name: name,
|
||||
Help: "Build information",
|
||||
Help: "Build information including version, build time, and git revision",
|
||||
},
|
||||
[]string{
|
||||
"version",
|
||||
"buildtime",
|
||||
"gittime",
|
||||
"git",
|
||||
"version", // Combined version/git format (e.g., "v1.0.0/abc123")
|
||||
"buildtime", // Build timestamp from ldflags
|
||||
"gittime", // Git commit timestamp from VCS info
|
||||
"git", // Full Git commit hash
|
||||
},
|
||||
)
|
||||
registry.MustRegister(buildInfo)
|
||||
@ -121,12 +198,20 @@ func RegisterMetric(name string, registry prometheus.Registerer) {
|
||||
).Set(1)
|
||||
}
|
||||
|
||||
// v caches the formatted version string to avoid repeated computation.
|
||||
var v string
|
||||
|
||||
// VersionInfo returns the structured version information.
|
||||
// This provides programmatic access to version details for JSON serialization
|
||||
// or other structured uses.
|
||||
func VersionInfo() Info {
|
||||
return info
|
||||
}
|
||||
|
||||
// Version returns a human-readable version string suitable for display.
|
||||
// The format includes semantic version, Git revision, build time, and Go version.
|
||||
// Example: "v1.0.0/abc123f-M (2023-01-01T00:00:00Z, go1.21.0)"
|
||||
// The "-M" suffix indicates the working tree was modified during build.
|
||||
func Version() string {
|
||||
if len(v) > 0 {
|
||||
return v
|
||||
@ -154,10 +239,23 @@ func Version() string {
|
||||
return v
|
||||
}
|
||||
|
||||
// CheckVersion compares a version against a minimum required version.
|
||||
// Returns true if the version meets or exceeds the minimum requirement.
|
||||
//
|
||||
// Special handling:
|
||||
// - "dev-snapshot" is always considered valid (returns true)
|
||||
// - Git hash suffixes (e.g., "v1.0.0/abc123") are stripped before comparison
|
||||
// - Uses semantic version comparison rules
|
||||
//
|
||||
// Both version and minimumVersion should follow semantic versioning with "v" prefix.
|
||||
func CheckVersion(version, minimumVersion string) bool {
|
||||
if version == "dev-snapshot" {
|
||||
return true
|
||||
}
|
||||
// Strip Git hash suffix if present (e.g., "v1.0.0/abc123" -> "v1.0.0")
|
||||
if idx := strings.Index(version, "/"); idx >= 0 {
|
||||
version = version[0:idx]
|
||||
}
|
||||
if semver.Compare(version, minimumVersion) < 0 {
|
||||
// log.Debug("version too old", "v", cl.Version.Version)
|
||||
return false
|
||||
|
414
version/version_test.go
Normal file
414
version/version_test.go
Normal file
@ -0,0 +1,414 @@
|
||||
package version
|
||||
|
||||
import (
|
||||
"runtime"
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
dto "github.com/prometheus/client_model/go"
|
||||
)
|
||||
|
||||
func TestCheckVersion(t *testing.T) {
|
||||
tests := []struct {
|
||||
In string
|
||||
Min string
|
||||
Expected bool
|
||||
}{
|
||||
// Basic version comparisons
|
||||
{"v3.8.4", "v3.8.5", false},
|
||||
{"v3.9.3", "v3.8.5", true},
|
||||
{"v3.8.5", "v3.8.5", true},
|
||||
// Dev snapshot should always pass
|
||||
{"dev-snapshot", "v3.8.5", true},
|
||||
{"dev-snapshot", "v99.99.99", true},
|
||||
// Versions with Git hashes should be stripped
|
||||
{"v3.8.5/abc123", "v3.8.5", true},
|
||||
{"v3.8.4/abc123", "v3.8.5", false},
|
||||
{"v3.9.0/def456", "v3.8.5", true},
|
||||
// Pre-release versions
|
||||
{"v3.8.5-alpha", "v3.8.5", false},
|
||||
{"v3.8.5", "v3.8.5-alpha", true},
|
||||
{"v3.8.5-beta", "v3.8.5-alpha", true},
|
||||
}
|
||||
|
||||
for _, d := range tests {
|
||||
r := CheckVersion(d.In, d.Min)
|
||||
if r != d.Expected {
|
||||
t.Errorf("CheckVersion(%q, %q) = %t, expected %t", d.In, d.Min, r, d.Expected)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestVersionInfo(t *testing.T) {
|
||||
info := VersionInfo()
|
||||
|
||||
// Check that we get a valid Info struct
|
||||
if info.Version == "" {
|
||||
t.Error("VersionInfo().Version should not be empty")
|
||||
}
|
||||
|
||||
// Version should start with "v" or be "dev-snapshot"
|
||||
if !strings.HasPrefix(info.Version, "v") && info.Version != "dev-snapshot" {
|
||||
t.Errorf("Version should start with 'v' or be 'dev-snapshot', got: %s", info.Version)
|
||||
}
|
||||
|
||||
// GitRevShort should be <= 7 characters if set
|
||||
if info.GitRevShort != "" && len(info.GitRevShort) > 7 {
|
||||
t.Errorf("GitRevShort should be <= 7 characters, got: %s", info.GitRevShort)
|
||||
}
|
||||
|
||||
// GitRevShort should be prefix of GitRev if both are set
|
||||
if info.GitRev != "" && info.GitRevShort != "" {
|
||||
if !strings.HasPrefix(info.GitRev, info.GitRevShort) {
|
||||
t.Errorf("GitRevShort should be prefix of GitRev: %s not prefix of %s",
|
||||
info.GitRevShort, info.GitRev)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func TestVersion(t *testing.T) {
|
||||
version := Version()
|
||||
|
||||
if version == "" {
|
||||
t.Error("Version() should not return empty string")
|
||||
}
|
||||
|
||||
// Should contain Go version
|
||||
if !strings.Contains(version, runtime.Version()) {
|
||||
t.Errorf("Version should contain Go version %s, got: %s", runtime.Version(), version)
|
||||
}
|
||||
|
||||
// Should contain the VERSION variable (or dev-snapshot)
|
||||
info := VersionInfo()
|
||||
if !strings.Contains(version, info.Version) {
|
||||
t.Errorf("Version should contain %s, got: %s", info.Version, version)
|
||||
}
|
||||
|
||||
// Should be in expected format: "version (extras)"
|
||||
if !strings.Contains(version, "(") || !strings.Contains(version, ")") {
|
||||
t.Errorf("Version should be in format 'version (extras)', got: %s", version)
|
||||
}
|
||||
}
|
||||
|
||||
func TestVersionCmd(t *testing.T) {
|
||||
appName := "testapp"
|
||||
cmd := VersionCmd(appName)
|
||||
|
||||
// Test basic command properties
|
||||
if cmd.Use != "version" {
|
||||
t.Errorf("Expected command use to be 'version', got: %s", cmd.Use)
|
||||
}
|
||||
|
||||
if cmd.Short == "" {
|
||||
t.Error("Command should have a short description")
|
||||
}
|
||||
|
||||
if cmd.Long == "" {
|
||||
t.Error("Command should have a long description")
|
||||
}
|
||||
|
||||
if cmd.Run == nil {
|
||||
t.Error("Command should have a Run function")
|
||||
}
|
||||
|
||||
// Test that the command can be executed without error
|
||||
cmd.SetArgs([]string{})
|
||||
err := cmd.Execute()
|
||||
if err != nil {
|
||||
t.Errorf("VersionCmd execution should not return error, got: %s", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestKongVersionCmd(t *testing.T) {
|
||||
cmd := &KongVersionCmd{Name: "testapp"}
|
||||
|
||||
// Test that Run() doesn't return an error
|
||||
err := cmd.Run()
|
||||
if err != nil {
|
||||
t.Errorf("KongVersionCmd.Run() should not return error, got: %s", err)
|
||||
}
|
||||
}
|
||||
|
||||
func TestRegisterMetric(t *testing.T) {
|
||||
// Create a test registry
|
||||
registry := prometheus.NewRegistry()
|
||||
|
||||
// Test registering metric without name
|
||||
RegisterMetric("", registry)
|
||||
|
||||
// Gather metrics
|
||||
metricFamilies, err := registry.Gather()
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to gather metrics: %s", err)
|
||||
}
|
||||
|
||||
// Find the build_info metric
|
||||
var buildInfoFamily *dto.MetricFamily
|
||||
for _, family := range metricFamilies {
|
||||
if family.GetName() == "build_info" {
|
||||
buildInfoFamily = family
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
if buildInfoFamily == nil {
|
||||
t.Fatal("build_info metric not found")
|
||||
}
|
||||
|
||||
if buildInfoFamily.GetHelp() == "" {
|
||||
t.Error("build_info metric should have help text")
|
||||
}
|
||||
|
||||
metrics := buildInfoFamily.GetMetric()
|
||||
if len(metrics) == 0 {
|
||||
t.Fatal("build_info metric should have at least one sample")
|
||||
}
|
||||
|
||||
// Check that the metric has the expected labels
|
||||
metric := metrics[0]
|
||||
labels := metric.GetLabel()
|
||||
|
||||
expectedLabels := []string{"version", "buildtime", "gittime", "git"}
|
||||
labelMap := make(map[string]string)
|
||||
|
||||
for _, label := range labels {
|
||||
labelMap[label.GetName()] = label.GetValue()
|
||||
}
|
||||
|
||||
for _, expectedLabel := range expectedLabels {
|
||||
if _, exists := labelMap[expectedLabel]; !exists {
|
||||
t.Errorf("Expected label %s not found in metric", expectedLabel)
|
||||
}
|
||||
}
|
||||
|
||||
// Check that the metric value is 1
|
||||
if metric.GetGauge().GetValue() != 1 {
|
||||
t.Errorf("Expected build_info metric value to be 1, got %f", metric.GetGauge().GetValue())
|
||||
}
|
||||
}
|
||||
|
||||
func TestRegisterMetricWithName(t *testing.T) {
|
||||
// Create a test registry
|
||||
registry := prometheus.NewRegistry()
|
||||
|
||||
// Test registering metric with custom name
|
||||
appName := "my-test-app"
|
||||
RegisterMetric(appName, registry)
|
||||
|
||||
// Gather metrics
|
||||
metricFamilies, err := registry.Gather()
|
||||
if err != nil {
|
||||
t.Fatalf("Failed to gather metrics: %s", err)
|
||||
}
|
||||
|
||||
// Find the my_test_app_build_info metric
|
||||
expectedName := "my_test_app_build_info"
|
||||
var buildInfoFamily *dto.MetricFamily
|
||||
for _, family := range metricFamilies {
|
||||
if family.GetName() == expectedName {
|
||||
buildInfoFamily = family
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
if buildInfoFamily == nil {
|
||||
t.Fatalf("%s metric not found", expectedName)
|
||||
}
|
||||
}
|
||||
|
||||
func TestVersionConsistency(t *testing.T) {
|
||||
// Call Version() multiple times and ensure it returns the same result
|
||||
v1 := Version()
|
||||
v2 := Version()
|
||||
|
||||
if v1 != v2 {
|
||||
t.Errorf("Version() should return consistent results: %s != %s", v1, v2)
|
||||
}
|
||||
}
|
||||
|
||||
func TestVersionInfoConsistency(t *testing.T) {
|
||||
// Ensure VersionInfo() is consistent with Version()
|
||||
info := VersionInfo()
|
||||
version := Version()
|
||||
|
||||
// Version string should contain the semantic version
|
||||
if !strings.Contains(version, info.Version) {
|
||||
t.Errorf("Version() should contain VersionInfo().Version: %s not in %s",
|
||||
info.Version, version)
|
||||
}
|
||||
|
||||
// If GitRevShort is set, version should contain it
|
||||
if info.GitRevShort != "" {
|
||||
if !strings.Contains(version, info.GitRevShort) {
|
||||
t.Errorf("Version() should contain GitRevShort: %s not in %s",
|
||||
info.GitRevShort, version)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Test edge cases
|
||||
func TestCheckVersionEdgeCases(t *testing.T) {
|
||||
// Test with empty strings
|
||||
if CheckVersion("", "v1.0.0") {
|
||||
t.Error("Empty version should not be >= v1.0.0")
|
||||
}
|
||||
|
||||
// Test with malformed versions (should be handled gracefully)
|
||||
// Note: semver.Compare might panic or return unexpected results for invalid versions
|
||||
// but our function should handle the common cases
|
||||
tests := []struct {
|
||||
version string
|
||||
minimum string
|
||||
desc string
|
||||
}{
|
||||
{"v1.0.0/", "v1.0.0", "version with trailing slash"},
|
||||
{"v1.0.0/abc/def", "v1.0.0", "version with multiple slashes"},
|
||||
}
|
||||
|
||||
for _, test := range tests {
|
||||
// This should not panic
|
||||
result := CheckVersion(test.version, test.minimum)
|
||||
t.Logf("%s: CheckVersion(%q, %q) = %t", test.desc, test.version, test.minimum, result)
|
||||
}
|
||||
}
|
||||
|
||||
// Benchmark version operations
|
||||
func BenchmarkVersion(b *testing.B) {
|
||||
// Reset the cached version to test actual computation
|
||||
v = ""
|
||||
|
||||
b.ResetTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
_ = Version()
|
||||
}
|
||||
}
|
||||
|
||||
func BenchmarkVersionInfo(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
_ = VersionInfo()
|
||||
}
|
||||
}
|
||||
|
||||
func BenchmarkCheckVersion(b *testing.B) {
|
||||
version := "v1.2.3/abc123"
|
||||
minimum := "v1.2.0"
|
||||
|
||||
b.ResetTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
_ = CheckVersion(version, minimum)
|
||||
}
|
||||
}
|
||||
|
||||
func BenchmarkCheckVersionDevSnapshot(b *testing.B) {
|
||||
version := "dev-snapshot"
|
||||
minimum := "v1.2.0"
|
||||
|
||||
b.ResetTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
_ = CheckVersion(version, minimum)
|
||||
}
|
||||
}
|
||||
|
||||
func TestParseBuildTime(t *testing.T) {
|
||||
tests := []struct {
|
||||
name string
|
||||
input string
|
||||
expected string
|
||||
}{
|
||||
{
|
||||
name: "Unix epoch timestamp",
|
||||
input: "1672531200", // 2023-01-01T00:00:00Z
|
||||
expected: "2023-01-01T00:00:00Z",
|
||||
},
|
||||
{
|
||||
name: "Unix epoch zero",
|
||||
input: "0",
|
||||
expected: "1970-01-01T00:00:00Z",
|
||||
},
|
||||
{
|
||||
name: "Valid RFC3339 format",
|
||||
input: "2023-12-25T15:30:45Z",
|
||||
expected: "2023-12-25T15:30:45Z",
|
||||
},
|
||||
{
|
||||
name: "RFC3339 with timezone",
|
||||
input: "2023-12-25T10:30:45-05:00",
|
||||
expected: "2023-12-25T10:30:45-05:00",
|
||||
},
|
||||
{
|
||||
name: "Empty string",
|
||||
input: "",
|
||||
expected: "",
|
||||
},
|
||||
{
|
||||
name: "Invalid format - return unchanged",
|
||||
input: "not-a-date",
|
||||
expected: "not-a-date",
|
||||
},
|
||||
{
|
||||
name: "Invalid timestamp - return unchanged",
|
||||
input: "invalid-timestamp",
|
||||
expected: "invalid-timestamp",
|
||||
},
|
||||
{
|
||||
name: "Partial date - return unchanged",
|
||||
input: "2023-01-01",
|
||||
expected: "2023-01-01",
|
||||
},
|
||||
{
|
||||
name: "Negative epoch - should work",
|
||||
input: "-1",
|
||||
expected: "1969-12-31T23:59:59Z",
|
||||
},
|
||||
{
|
||||
name: "Large epoch timestamp",
|
||||
input: "4102444800", // 2100-01-01T00:00:00Z
|
||||
expected: "2100-01-01T00:00:00Z",
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func(t *testing.T) {
|
||||
result := parseBuildTime(tt.input)
|
||||
if result != tt.expected {
|
||||
t.Errorf("parseBuildTime(%q) = %q, expected %q", tt.input, result, tt.expected)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestParseBuildTimeConsistency(t *testing.T) {
|
||||
// Test that calling parseBuildTime multiple times with the same input returns the same result
|
||||
testInputs := []string{
|
||||
"1672531200",
|
||||
"2023-01-01T00:00:00Z",
|
||||
"invalid-date",
|
||||
"",
|
||||
}
|
||||
|
||||
for _, input := range testInputs {
|
||||
result1 := parseBuildTime(input)
|
||||
result2 := parseBuildTime(input)
|
||||
if result1 != result2 {
|
||||
t.Errorf("parseBuildTime(%q) not consistent: %q != %q", input, result1, result2)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func BenchmarkParseBuildTime(b *testing.B) {
|
||||
inputs := []string{
|
||||
"1672531200", // Unix epoch
|
||||
"2023-01-01T00:00:00Z", // RFC3339
|
||||
"invalid-timestamp", // Invalid
|
||||
"", // Empty
|
||||
}
|
||||
|
||||
for _, input := range inputs {
|
||||
b.Run(input, func(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
_ = parseBuildTime(input)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
@ -1,3 +1,27 @@
|
||||
// Package fastlyxff provides Fastly CDN IP range management for trusted proxy handling.
|
||||
//
|
||||
// This package parses Fastly's public IP ranges JSON file and generates Echo framework
|
||||
// trust options for proper client IP extraction from X-Forwarded-For headers.
|
||||
// It's designed specifically for services deployed behind Fastly's CDN that need
|
||||
// to identify real client IPs for logging, rate limiting, and security purposes.
|
||||
//
|
||||
// Fastly publishes their edge server IP ranges in a JSON format that this package
|
||||
// consumes to automatically configure trusted proxy ranges. This ensures that
|
||||
// X-Forwarded-For headers are only trusted when they originate from legitimate
|
||||
// Fastly edge servers.
|
||||
//
|
||||
// Key features:
|
||||
// - Automatic parsing of Fastly's IP ranges JSON format
|
||||
// - Support for both IPv4 and IPv6 address ranges
|
||||
// - Echo framework integration via TrustOption generation
|
||||
// - CIDR notation parsing and validation
|
||||
//
|
||||
// The JSON file typically contains IP ranges in this format:
|
||||
//
|
||||
// {
|
||||
// "addresses": ["23.235.32.0/20", "43.249.72.0/22", ...],
|
||||
// "ipv6_addresses": ["2a04:4e40::/32", "2a04:4e42::/32", ...]
|
||||
// }
|
||||
package fastlyxff
|
||||
|
||||
import (
|
||||
@ -9,15 +33,29 @@ import (
|
||||
"github.com/labstack/echo/v4"
|
||||
)
|
||||
|
||||
// FastlyXFF represents Fastly's published IP ranges for their CDN edge servers.
|
||||
// This structure matches the JSON format provided by Fastly for their public IP ranges.
|
||||
// It contains separate lists for IPv4 and IPv6 CIDR ranges.
|
||||
type FastlyXFF struct {
|
||||
IPv4 []string `json:"addresses"`
|
||||
IPv6 []string `json:"ipv6_addresses"`
|
||||
IPv4 []string `json:"addresses"` // IPv4 CIDR ranges (e.g., "23.235.32.0/20")
|
||||
IPv6 []string `json:"ipv6_addresses"` // IPv6 CIDR ranges (e.g., "2a04:4e40::/32")
|
||||
}
|
||||
|
||||
// TrustedNets holds parsed network prefixes for efficient IP range checking.
|
||||
// This type is currently unused but reserved for future optimizations
|
||||
// where frequent IP range lookups might benefit from pre-parsed prefixes.
|
||||
type TrustedNets struct {
|
||||
prefixes []netip.Prefix
|
||||
prefixes []netip.Prefix // Parsed network prefixes for efficient lookups
|
||||
}
|
||||
|
||||
// New loads and parses Fastly IP ranges from a JSON file.
|
||||
// The file should contain Fastly's published IP ranges in their standard JSON format.
|
||||
//
|
||||
// Parameters:
|
||||
// - fileName: Path to the Fastly IP ranges JSON file
|
||||
//
|
||||
// Returns the parsed FastlyXFF structure or an error if the file cannot be
|
||||
// read or the JSON format is invalid.
|
||||
func New(fileName string) (*FastlyXFF, error) {
|
||||
b, err := os.ReadFile(fileName)
|
||||
if err != nil {
|
||||
@ -34,6 +72,19 @@ func New(fileName string) (*FastlyXFF, error) {
|
||||
return &d, nil
|
||||
}
|
||||
|
||||
// EchoTrustOption converts Fastly IP ranges into Echo framework trust options.
|
||||
// This method generates trust configurations that tell Echo to accept X-Forwarded-For
|
||||
// headers only from Fastly's edge servers, ensuring accurate client IP extraction.
|
||||
//
|
||||
// The generated trust options should be used with Echo's IP extractor:
|
||||
//
|
||||
// options, err := fastlyRanges.EchoTrustOption()
|
||||
// if err != nil {
|
||||
// return err
|
||||
// }
|
||||
// e.IPExtractor = echo.ExtractIPFromXFFHeader(options...)
|
||||
//
|
||||
// Returns a slice of Echo trust options or an error if any CIDR range cannot be parsed.
|
||||
func (xff *FastlyXFF) EchoTrustOption() ([]echo.TrustOption, error) {
|
||||
ranges := []echo.TrustOption{}
|
||||
|
||||
|
@ -3,14 +3,12 @@ package fastlyxff
|
||||
import "testing"
|
||||
|
||||
func TestFastlyIPRanges(t *testing.T) {
|
||||
|
||||
fastlyxff, err := New("fastly.json")
|
||||
if err != nil {
|
||||
t.Fatalf("could not load test data: %s", err)
|
||||
}
|
||||
|
||||
data, err := fastlyxff.EchoTrustOption()
|
||||
|
||||
if err != nil {
|
||||
t.Fatalf("could not parse test data: %s", err)
|
||||
}
|
||||
@ -19,5 +17,4 @@ func TestFastlyIPRanges(t *testing.T) {
|
||||
t.Logf("only got %d prefixes, expected more", len(data))
|
||||
t.Fail()
|
||||
}
|
||||
|
||||
}
|
||||
|
Loading…
x
Reference in New Issue
Block a user