Skip to content

6543/log

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Observability API

I propose to create an industry standard on observability API in Go. Let outline major problems, first...

Problem #1: Logger

The most popular loggers for Go are:

But there are also a lot of other loggers. All of them have different advantages and disadvantages. And basically there is no industry standard even on the interface of a logger.

Problem #2: Injecting additional observability tooling

When an application evolves it often meets new challenges:

  • The application become to complex and consists of multiple parts, and it becomes difficult to diagnose latencies (do not confuse with CPU profiling, latencies may be caused my network issues for example).
  • In some places logging is not applicable by multiple reasons. For example it may generate too many logs (like millions entries per second).
  • It becomes important to automatically look for regressions.

Thus it is required requires:

And all these tooling (logger, metrics, tracing and error monitoring) has the same pattern: they have structured fields.

Requirements

  1. In contrast to current practices for observability in Go, it is required to apply “dependency injection” and abstract the observability tooling using interfaces (to be able to inject whatever implementation is required).
  2. It should be possible define structured values gradually. Thus a context with fields should allow derivatives. Let’s call such approach “contextual” (see "Contextuality" below).
  3. Usability in a system with multiple services processing a single request.

Contextuality

There are different implementations of structured logging for Go, but the most successful in practice (like zap) allows to define structured fields gradually. For example:

func myFunc0(groupID int, zapLogger *zap.Logger) {
	zapLogger = zapLogger.With(zap.Int("group_id", groupID))
	...
	for _, userID := range group.UserIDs() {
		myFunc1(userID, zapLogger)
	}
	...
}
func myFunc1(userID int, zapLogger *zap.Logger) {
	zapLogger = zapLogger.With(zap.Int("user_id", userID))
	...
	zapLogger.Debug("user password is expired")
	...
}

This will create a log entry with at least two structured values: “group_id” and “user_id”. Or in logrus it will be:

func myFunc0(groupID int, logrusLogger *logrus.Entry) {
	logrusLogger = logrusLogger.WithField("group_id", groupID)
	...
	for _, userID := range group.UserIDs() {
		myFunc1(userID, logrusLogger)
	}
	...
}
func myFunc1(userID int, logrusLogger *logrus.Entry) {
	logrusLogger = logrusLogger.WithField("user_id", userID)
	...
	logrusLogger.Debugf("user password is expired")
	...
}

Thus myFunc1 itself has no responsibility to know about group_id to log it.

Candidates

We are known two candidates to solve the problems:

OpenTelemetry

The problems are:

  • They currently support only metrics and tracing (https://github.com/open-telemetry/opentelemetry-go/blob/257ef7fc150fb5b37fa1b42c83134dd52685861e/README.md#export). And it is unclear how logger API (which is the most important one) will look like. Here is the discussion on this subject: open-telemetry/opentelemetry-go#2010
  • They have too complicated API (difficult to swallow right away by a gopher-newbie). This seems to be caused by attempt to provide be a completely SOLID-compliant solution, which is theoretically good, but in practice may frighten away users from using it. Good enough is better than perfect, and it seems better to sacrifice non-important design features to improve user experience.
  • It looks like OpenTelemetry does not support “contextuality”.
  • OpenTelemetry defines observability tooling as separate tools, instead of unifying them into joint entity “Observability” which:
    • Generalizes the “contextuality” property over all tools.
    • Provides placeholders to inject observability tools (like metrics) when it will be required.

xcontext

The problem is that SOLID is not followed. The extended context is a light version of a God Object.

Solution

Part 1: Logger interface

The properties of the logger:

  • Dependency injection. The logger should be possible to inject (like in dependency injection) and abstract.
  • It should be structured and contextual.
  • It should support logging levels, and they also should be contextual.
  • It should be extendable. It should be possible to attach “hooks“ to modify log entries, or to actually log them (to stdout, file, logstash, whatever).

Part 1.1: Logger interface: structured contextual logging

It should be possible to gradually extend the context by creating derivatives (without modification the original context). There proposed two major approaches:

  • Either the context is a part of the logger itself.
  • Or logger is responsible for logging only, and logging context is a separate entity.

It would be somewhat more SOLID to use the second approach (see "single-responsibility principle"). But in practice it does not change anything except that an user will need to pass-through two entities through the whole call stack instead of one. Thus we propose to use the first approach:

type Fields map[string]interface{}

type Logger interface {
	...
	WithField(fieldName string, fieldValue interface{})
	WithFields(fields Fields)
	WithStruct(interface{})
	...
}

The main downside here is that map[string]interface{} is too expensive in terms of CPU consumption. For example non-sugared zap has a Field structure to avoid interface-boxing and allocating maps. And we can re-use the same approach:

type Field struct {
	Key       string
	Integer   int64
	String    string
	Bytes     []byte
	Type      reflect.Type
	Interface interface{}
}

type Fields []Field

type Logger interface {
	...
	WithValue(key string, value interface{})
	WithFields(fields Fields)
	WithMap(map[string]interface{})
	WithStruct(interface{})
	...
}

[BTW, I had to use the same approach, when I experimented with high-performance map-s in Go]

An usage example:

func myFunc(myLogger Logger) {
	...

	myLogger = myLogger.WithFields(logger.Fields{
		{
			Key: "user_id",
			Integer: userID,
		},
		{
			Key: "user_name",
			String: userName,
		}
	})

	...

	myLogger.WithField("error_msg", err.Error()).Debug("yay!")

	...

	anotherFunc(myLogger.WithStruct(rsaPubKey))
}

Part 1.2: Logger interface: logging levels

There are two main approaches here: to define a method for each logging level or to generalize everything into single method:

type Logger interface {
	...
	Debugf(format string, args ...interface{})
	Infof(format string, args ...interface{})
	Warnf(format string, args ...interface{})
	Errorf(format string, args ...interface{})
	...
}

vs

type Logger interface {
	...
	Log(level Level, format string, args ...interface{})
	...
}

Despite the fact the generalized approach (with Log) provides much less code duplication, it provide worse user experience:

myLogger.Log(logger.LevelDebug, "yay!")

is less readable than:

myLogger.Debug("yay!")

And the logger is a thing which is very rarely modified, but used a lot. Thus it is proposed to define separate methods for each logging level.

And to make logging level contextual it is required to have setter and getter methods:

WithLevel(level Level) Logger
Level() Level

Part 1.3: Logger interface: logging methods

Historically (since "C") a lot of programmers used to format+arguments functions, like:

Printf(format string, args ...interface{})

The problem of this API is that it is oriented for non-structural output, while it is required to encourage users to log data in a structured way. But to allow smooth transition we still need to permit the API people used, to. Thus it is proposed to have two formats:

Debugf(format string, args ...interface{})

and:

Debug(values ...interface{})

If method Debugf is used then basically fmt.Sprintf is used to format the string and it is logged as is with key message.

If method Debug is used then depending on types of input values they are interpreted in a different way:

  • If Field is provided then it is interpreted as Field.
  • If a structure (or a pointer to a structure) is provided, then each struct field is interpreted as a separate value. The key of the value is determined from tag log or from struct field name (if the tag is not set).
  • Everything rest is just converted to string and joined together into message.

If values is empty then log only those fields which were set in this context (without adding anything). Thus these variants are equivalent:

myLogger.Debug("yay!")
myLogger.WithValue("message", "yay!").Debug()

Part 1.4: Logger interface: Flush()

Since logging is highly connected with IO and specifically "output" it is required to provide method "Flush()" to support buffered output (which is usually critical in high performance applications):

type Logger interface {
	...
	Flush()
	...
}

Part 1.5: Logger interface: extendable

It should be possible to add hooks to a Logger which are responsible for modifying/enriching the logged data and performing the actual logging. Therefore it is required to define how a logging entry looks like:

type Entry interface {
	Timestamp time.Time
	Level Level
	Fields Fields
	Caller Caller
}

(see also semantically the same entity in logrus and zap)

Type Caller basically could be completely re-used from zap (it is already optimal):

type Caller struct {
	Defined bool
	PC      uintptr
	File    string
	Line    int
}

In total a hook is:

type Hook interface {
	Process(*Entry)
}

And in turn a Logger is extendable by these hooks:

type Logger interface {
	...
	WithHooks(...Hook) Logger
	Hooks() []Hook
	...
}

To remove all hooks: myLogger = myLogger.WithHooks().

Part 1.6: Logger interface: in total

// Field is a CPU-effective way to add structured values.
type Field struct {
	// Key is the name/key of the value.
	Key string

	// Integer is the value if it is an integer (int8, uint8, int16, uint16, int32, uint32, int64, uint64).
	Integer int64

	// String is the value if it is a string or []byte.
	String string

	// Interface is the value if Integer and/or String cannot be used to store it.
	Interface interface{}

	// Type is a selector if Integer, String or Interface should be used to get the value.
	Type reflect.Type

	// Properties defines Hook-specific options (how to interpret this field).
	Properties FieldProperties
}

// FieldProperty defines a Tool-specific option (how to interpret this field).
type FieldProperty interface{}

// FieldProperties defines Tool-specific options (how to interpret this field).
type FieldProperties []FieldProperty

// Fields is a set of Field-s
type Fields []Field

// Caller is a simplified version of runtime.Frame, which contains only
// data required in a Logger.
type Caller struct {
	// PC is the program counter, see the description in runtime.Frame.
	PC uintptr

	// File is the file name or path, see the description in runtime.Frame.
	File string

	// Line is the code line number, see the description in runtime.Frame.
	Line uint
}

// Entry a single log entry to be logged/written.
type Entry interface {
	// Timestamp defines the time moment when the entry was issued (for example method `Debugf` was called).
	Timestamp time.Time

	// Level is the logging level of the entry.
	Level Level

	// Fields is the set of values to be logged. Field "message" is the default field (and it is used for example to store the final string of `Debugf`).
	Fields Fields

	// Caller is a minimalistic runtime.Frame.
	//
	// If Caller is not set (for example, due to performance reasons) then it should
	// have the zero-value.
	Caller Caller
}

// Hook is a processor for log entries. It may for example modify them or write then into an actual log-storage.
type Hook interface {
	Process(*Entry)

	// Flush gracefully empties any buffers the hook may have.
	Flush()
}

// Logger defines a generic logger
type Logger interface {
	// XXXf creates a log entry with level XXX and field `message` formed as fmt.Sprintf of provided arguments.
	Tracef(format string, args ...interface{})
	Debugf(format string, args ...interface{})
	Infof(format string, args ...interface{})
	Warnf(format string, args ...interface{})
	Errorf(format string, args ...interface{})

	// Panicf creates the log entry and triggers `panic`.
	Panicf(format string, args ...interface{})

	// Fatalf creates the log entry and immediately closes the whole application.
	Fatalf(format string, args ...interface{})

	// XXX create a log entry with level XXX and fields parses from `values`.
	//
	// `values` are parsed in separate:
	// * If a value has type Field then it is parsed as Field.
	// * If a value is a struct or a pointer to a struct then it is parsed as a struct. In this case each field of the struct defines a separate value for the log entry. The key of the value equals to the field name or to tag `log` value if it is defined. To skip the field add tag `log:"-"`
	// * If a value is neither of above then it is appended to field `message` as a string.
	//
	// If no values is provided then the log entry is issued without adding additional fields. Thus these are equivalents:
	//     myLogger.WithField("message", "hello world!").Debug()
	// and
	//     myLogger.Debug("hello", " world!")
	Trace(values ...interface{})
	Debug(values ...interface{})
	Info(values ...interface{})
	Warn(values ...interface{})
	Error(values ...interface{})

	// Panic creates the log entry and triggers `panic`.
	Panic(values ...interface{})

	// Fatal creates the log entry and immediately closes the whole application.
	Fatal(values ...interface{})

	// With* returns a Logger derivative which includes passed values to all issued log entries.
	WithValue(key string, value interface{}) Logger
	WithFields(fields Fields) Logger
	WithMap(map[string]interface{}) Logger
	WithStruct(interface{}) Logger

	// Fields returns the values to be included to issued log entries
	Fields() Fields

	// WithHooks returns a Logger derivative which also includes/appends hooks from the arguments.
	//
	// Special case: to reset hooks use `WithHooks()` (without any arguments).
	WithHooks(...Hook) Logger

	// Hooks returns current Hooks.
	Hooks() []Hook

	// WithLevel returns a Logger derivative which has defined logging level.
	WithLevel(level Level) Logger

	// Level returns current logging level.
	Level() Level

	// Flush forces to flush all buffers of all Hooks.
	Flush()
}

Part 2: Observability API

Part 2.1: Observability API: basics

Currently there are known two approaches:

OpenTelemetry does not provide required properties, while xcontext is designed in as somekind of a "God Object". To combine the best parts of the both approaches it is proposed to create a modular observability object, which does not know anything about specific observability tool, but unifies them together with support of contextual structured data.

type Tool interface {
	WithValue(key string, value interface{}, props ...FieldProperty) Tool
	WithFields(fields Fields) Tool
	WithMap(map[string]interface{}, props ...FieldProperty) Tool
	WithStruct(interface{}, props ...FieldProperty) Tool
	Fields() Fields
}

type Tools []Tools

func (tools Tools) Get(toolType reflect.Type) Tool {
	// linear search is faster than map if there are only few entries
	for _, tool := range tools {
		// reflect.TypeOf has no overhead when inlined, it just returns the type-field of the interface.
		// See: https://github.com/xaionaro-go/benchmarks/tree/master/reflect
		if reflect.TypeOf(tool) == toolType {
			return tool
		}
	}
	return nil
}

type Observer struct { ... }

func (obs *Observer) WithValue(key string, value interface{}, props ...FieldProperty) *Observer { ... }
func (obs *Observer) WithFields(fields ...Fields) *Observer { ... }
func (obs *Observer) WithMap(map[string]interface{}, props ...FieldProperty) *Observer { ... }
func (obs *Observer) WithStruct(interface{}, props ...FieldProperty) *Observer { ... }
func (obs *Observer) Fields() Fields { ... }
func (obs *Observer) WithTools(tools ...Tool) *Observer { ... }
func (obs *Observer) Tools() Tools

Remark: With* methods has optional props ...FieldProperty. It is required to address another problem:

  • Sometimes it is required to add a field to one tool, but to do not add it to other tools. For example if there is a random value then it usually should not be placed into metrics, otherwise it will create a cond-infinite amount of end-metrics (and consume all RAM). Thus it is possible to pass optional properties which could be interpreted by Tool-s to ignore or not.

An example of usage:

func main() {
	...
	obs := obs.NewObserver().WithTools(metrics.New(), logger.New()) // metrics are first, because they should be faster
	...
	obs = obs.WithValue("pid", os.Getpid(), metrics.PropIgnore)
	...
	myFunc(obs)
	...
}

func myFunc(obs *Observer) {
	obs = obs.WithValue("myTag", "myValue")
	defer tracer.FromObs(obs).Start("doingSomeStuff").Finish()
	...
	logger.FromObs(obs).Debug("yay!")
	...
	metrics.FromObs(obs).Count("yay").Add(1)
	...
}

This will resuls into a log entry with fields {"pid":1234, "myTag": "myValue", "message":"yay!"} and it will increment a metric with key yay and tag {"myTag": "myValue"}. And it won't panic on tracer despite the fact it was not initialized, because FromObs is obligated to return a dummy implementation if no tracer is defined in the Observer.

Part 2.2: Observability API: integrating with context

It is possible (but not necessary) to inject Observer into a standard context:

package obscontext

import "context"

type ctxKey string
const ctxKeyObserver = ctxKey("observer")

func WithObs(ctx context.Context, obs *Observer) context.Context {
	return context.WithValue(ctx, ctxKeyObserver, obs)
}
func Obs(ctx context.Context) *Observer {
	observer := ctx.Value(ctxKeyObserver)
	if observer == nil {
		return obs.Default()
	}
	return observer.(*Observer)
}

func WithValue(ctx context.Context, key string, value interface{}) context.Context {
	observer := ctx.Value(ctxKeyObserver)
	if obs == nil {
		return ctx
	}
	return context.WithValue(ctx, obs.WithValue(key, value))
}
func WithFields(ctx context.Context, fields Fields) context.Context { ... }
func WithMap(ctx context.Context, map[string]interface{}) context.Context { ... }
func WithStruct(ctx context.Context, interface{}) context.Context { ... }

It is proposed to implement support of contexts are a separate package, because there is no consensus on is this a right approach or not. It is a well-recognized pattern, but the arguments againts storing observer (usually specifically "Logger") in a context:

  • A context should not contain "behavior".
  • A context abstracts values to interface{} which effectively disables some runtime guarantees. And even if you will somehow avoid panics you still cannot validate at compile time that Observer will be reachable in a specific piece of code.

Counter-arguments:

  • How many is stored in context, and how many is stored in a global singleton is decision for a specific Logger implementation. This approach does not enforce putting "behavior" to context.
  • In this approach we require getters to return the default instance if the context has no Observer. This way it cannot cause runtime errors and guarantee a fallback mechanism to still do logs.

Thus to provide an user a choice use it or not to use we put this code into a separate package.

Part 2.3: Observability API: TraceID

A distributed system requires to diagnose cross-service issues.

The most common way to solve this problem is to have a TraceID which is persistent for an end-user request through all logs from all services.

Thus Observer may also implement:

func (obs *Observer) TraceID() string
func (obs *Observer) WithTraceID(traceID string) *Observer

Another challenges are:

  • A request to a frontend service may leads to multiple requests to other services, which in turn may also multiply amount of requests. Thus a request processing is actually a tree-like process. And sometimes it is required diagnose a specific branch instead of the whole tree.
  • A log entry may be related to multiple TraceID-s. For example, an user may send to requests to a job management service: one to create a job and another to cancel it. There could be a log entry which includes both TraceID-s.

Therefore it is proposed instead of the methods above to enforce methods:

func (obs *Observer) TraceIDs() []string
func (obs *Observer) WithTraceID(traceIDs ...string) *Observer

Each time a new request is generated, a new TraceID is added. For example:

  • An end-user sent a request with TraceID A to service #1 to create a job.
  • Service #1 received the request and sent requests to services #2 and #3. The requests has TraceIDs: A,B and A,C.
  • Service #3 received the request and sent a request to service #4. The TraceID is A,C,D.
  • An end-user sent a request with TraceID E to service #1 to cancel the job.
  • Service #1 sent to service #3 a request with TraceID E,F.
  • Service #3 sent to service #4 a request with TraceID E,F,G.
  • Somewhere in service #4 among other logs there could be logs with TraceID A,C,D,E,F,G because they caused by combination of the first and second end-user requests.

In log storage TraceIDs are supposed to be stored as a set of tags.

Part 2.4: Observability API: in total

package obs

// Field is a CPU-effective way to add structured values.
//
// The same as in Logger (should be type-aliased).
type Field struct {
	// Key is the name/key of the value.
	Key string

	// Integer is the value if it is an integer (int8, uint8, int16, uint16, int32, uint32, int64, uint64).
	Integer int64

	// String is the value if it is a string or []byte.
	String string

	// Interface is the value if Integer and/or String cannot be used to store it.
	Interface interface{}

	// Type is a selector if Integer, String or Interface should be used to get the value.
	Type reflect.Type

	// Properties defines Hook-specific options (how to interpret this field).
	Properties FieldProperties
}

// FieldProperty defines a Tool-specific option (how to interpret this field).
//
// The same as in Logger (should be type-aliased).
type FieldProperty interface{}

// FieldProperties defines Tool-specific options (how to interpret this field).
//
// The same as in Logger (should be type-aliased).
type FieldProperties []FieldProperty

// Tool is an abstract observability tool. It could be a Logger, metrics, tracing or anything else.
type Tool interface {
	// With* returns a Tool derivative which includes passed structured values into issued data.
	WithValue(key string, value interface{}) Tool
	WithField(fields ...Field) Tool
	WithMap(map[string]interface{}) Tool
	WithStruct(interface{}) Tool

	// Fields returns current set of structured values.
	Fields() Fields
}

// Tools is a collection of observability Tool-s.
type Tools []Tool

// Get returns a Tool of a specified type. Returns nil if such Tool is not set.
func (tools Tools) Get(toolType reflect.Type) Tool {
	// linear search is faster than map if there are only few entries
	for _, tool := range tools {
		// reflect.TypeOf has no overhead when inlined, it just returns the type-field of the interface.
		// See: https://github.com/xaionaro-go/benchmarks/tree/master/reflect
		if reflect.TypeOf(tool) == toolType {
			return tool
		}
	}
	return nil
}

// Observer is a collection of observability tools.
type Observer struct { ... }

// With* returns an Observer derivative which includes the passed values into issued data (logs, metrics etc).
func (obs *Observer) WithValue(key string, value interface{}, props ...FieldProperty) *Observer { ... }
func (obs *Observer) WithFields(fields ...Fields) *Observer { ... }
func (obs *Observer) WithMap(map[string]interface{}, props ...FieldProperty) *Observer { ... }
func (obs *Observer) WithStruct(interface{}, props ...FieldProperty) *Observer { ... }
// Fields returns returns the set of fields set in the scope of this Observer.
func (obs *Observer) Fields() Fields { ... }

// WithTools returns an Observer derivative which includes passed Tools to the collection.
// If some Tool is defined twice, then the latest variant is used.
//
// Special case: to reset the collection to the empty set use `WithTools()` (with no arguments).
func (obs *Observer) WithTools(tools ...Tool) *Observer { ... }
// Tools returns the current collection of Tools.
func (obs *Observer) Tools() Tools

// WithTraceID returns an Observer derivative which includes passed traceIDs to the set of TraceIDs.
func (obs *Observer) WithTraceID(traceIDs ...string) *Observer
// TraceIDs returns the current set of TraceID-s.
func (obs *Observer) TraceIDs() []string
package obscontext

import "context"

type ctxKey string
const ctxKeyObserver = ctxKey("observer")

// WithObs returns a context derivative which includes the Observer as a value.
func WithObs(ctx context.Context, obs *Observer) context.Context {
	return context.WithValue(ctx, ctxKeyObserver, obs)
}
// Obs returns the Observer from context values. Returns the default observer if one is not set in the context.
func Obs(ctx context.Context) *Observer {
	observer := ctx.Value(ctxKeyObserver)
	if obs == nil {
		return obs.Default()
	}
	return observer.(*Observer)
}

// With* returns a context derivative with an Observer which includes the passed values.
func WithValue(ctx context.Context, key string, value interface{}) context.Context {
	obs := ctx.Value(ctxKeyObserver)
	if obs == nil {
		return ctx
	}
	return context.WithValue(ctx, obs.WithValue(key, value))
}
func WithFields(ctx context.Context, fields Fields) context.Context { ... }
func WithMap(ctx context.Context, map[string]interface{}) context.Context { ... }
func WithStruct(ctx context.Context, interface{}) context.Context { ... }

Example of initializing with a context:

ctx = obs.WithObs(ctx, obs.NewObserver(
	obsprom.New(prometheusMetrics),
	obszap.New(zapLogger),
	obszipkin.New(zipkinTracer),
	obssentry.New(sentryHandler),
))

Example of usage with a context:

func myFunc(ctx context.Context, [...]) {
	ctx = obs.WithValue(ctx, "user_id", user.ID())
	defer tracer.FromCtx(ctx).StartSpan("fixing user info").Finish()

	[...]

	logger.FromCtx(ctx).Debug(user)

	[...]

	logger.FromCtx(ctx).Debugf("user is assigned into %d groups", len(user.Groups()))
	for _, group := range user.Groups() {
		go func(user *User, group *Group) {
			defer errmon.FromCtx(ctx).Recover()

			ctx = obs.WithValue(ctx, "group_id", group.ID())
			[...]
			anotherFunc(ctx)
		}(user, group)
	}
}

Example of initializing outside of a context:

obs := obs.NewObserver(
	obsprom.New(prometheusMetrics),
	obszap.New(zapLogger),
	obszipkin.New(zipkinTracer),
	obssentry.New(sentryHandler),
))

Example of usage outside of a context:

func myFunc(ctx context.Context, obs *obs.Observer, [...]) {
	obs = obs.WithValue("user_id", user.ID())
	defer tracer.FromObs(obs).StartSpan("fixing user info").Finish()

	[...]

	logger.FromObs(obs).Debug(user)

	[...]

	logger.FromObs(obs).Debugf("user is assigned into %d groups", len(user.Groups()))
	for _, group := range user.Groups() {
		go func(user *User, group *Group) {
			defer errmon.FromObs(obs).Recover()

			obs = obs.WithValue("group_id", group.ID())
			[...]
			anotherFunc(ctx, obs)
		}(user, group)
	}
}

Part 3: Q&A

How to flush and close a Logger atomically, preventing it from further writing?

This is not a generic requirement, but a quite case-specific. It is not easy to find at least one Go project where logger is being closed. And if we will extend the generic interface with all case-specific methods, then it will grow indefinitely. For example there could be extract structured values already set to a Logger, but we also do not add it to the generic interface by the same reason.

If a project will require atomic Flush&Close, then they have multiple ways to solve this:

If they need Close() only in the root function then they can just use it directly:

myLogger := NewMyLogger() // of a concrete type with method `Close`.
defer myLogger.Close()

myFunc(myLogger) // here it is boxed to an interface without method `Close`

Or they can have a close function separately:

myLogger := NewMyLogger() // of a concrete type with method `Close`.
myFunc(myLogger, myLogger.Close) // here it is boxed to an interface without method `Close`

They can apply type-assertion if the logger supports Close:

myLogger.(io.Closer).Close()

Or they can just extend Logger interface:

type Logger interface {
	logger.Logger
	io.Closer
}

and Logger implementation:

type Logger struct {
	*obszap.Logger
}

func (l *Logger) Close() error {
	return l.Logger.Backend.Close()
}

Also if a Logger is being used as a singleton then it is possible to just use sync/atomic to atomically replace the logger with a dummy logger and flush the old one.

Does it makes sense to do a generic Logger if it by design do not cover all possible use cases? May be it is better to use specific Loggers in each specific case?

Generic logger allows different development teams to collaborate. It is similar to interface io.ReadWriteCloser which do not cover all possible use cases for IO objects. But it covers the part which is shared among them which makes such objects interchangable.

Having an own logger in each project is an inexcusable waste of time of talented engineers to re-invent the same wheel over and over. If there will be popular a generic Logger which covers the most of the cases, and the rest could be covered by for example type-assertion then it will provide an opportunity to re-use the same Logger code among the most of the projects.

The approach with functions FromObs and FromCtx should negatively affect performance. What is supposed to do in high-performance applications?

  • It is expected that this kind of performance overhead might be essential only for metrics, because these functions are much less expensive than the log writing or values interface-boxing (for methods like Debugf).
  • If there is a specific high-performance piece of code where such overhead may consume measurable amount of resources, then it is possible to extract a tool once and use it multiple times:
func mySlowFunc(ctx context.Context) {
	[...]
	logger := logger.FromCtx(ctx)
	badUsersMetric := metric.FromCtx(ctx).Count("users_bad")
	for _, user := range billionsOfUsers {
		fastFunc(user, logger, badUsersMetric)
	}
}

It is also possible to pre-type-assertion to make methods a little-bit faster:

func mySlowFunc(ctx context.Context) {
	[...]
	logger := logger.FromCtx(ctx).(*obszap.Logger)
	badUsersMetric := metric.FromCtx(ctx).Count("users_bad").(*obsprom.Counter)
	for _, user := range billionsOfUsers {
		fastFunc(user, logger, badUsersMetric)
	}
}

Similar approach is used in official prometheus client package: they require to have prepared instances of metric families.

About

Define a std **interface** for logging within go

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published